This list is closed, nobody may subscribe to it.
2010 |
Jan
|
Feb
(19) |
Mar
(8) |
Apr
(25) |
May
(16) |
Jun
(77) |
Jul
(131) |
Aug
(76) |
Sep
(30) |
Oct
(7) |
Nov
(3) |
Dec
|
---|---|---|---|---|---|---|---|---|---|---|---|---|
2011 |
Jan
|
Feb
|
Mar
|
Apr
|
May
(2) |
Jun
(2) |
Jul
(16) |
Aug
(3) |
Sep
(1) |
Oct
|
Nov
(7) |
Dec
(7) |
2012 |
Jan
(10) |
Feb
(1) |
Mar
(8) |
Apr
(6) |
May
(1) |
Jun
(3) |
Jul
(1) |
Aug
|
Sep
(1) |
Oct
|
Nov
(8) |
Dec
(2) |
2013 |
Jan
(5) |
Feb
(12) |
Mar
(2) |
Apr
(1) |
May
(1) |
Jun
(1) |
Jul
(22) |
Aug
(50) |
Sep
(31) |
Oct
(64) |
Nov
(83) |
Dec
(28) |
2014 |
Jan
(31) |
Feb
(18) |
Mar
(27) |
Apr
(39) |
May
(45) |
Jun
(15) |
Jul
(6) |
Aug
(27) |
Sep
(6) |
Oct
(67) |
Nov
(70) |
Dec
(1) |
2015 |
Jan
(3) |
Feb
(18) |
Mar
(22) |
Apr
(121) |
May
(42) |
Jun
(17) |
Jul
(8) |
Aug
(11) |
Sep
(26) |
Oct
(15) |
Nov
(66) |
Dec
(38) |
2016 |
Jan
(14) |
Feb
(59) |
Mar
(28) |
Apr
(44) |
May
(21) |
Jun
(12) |
Jul
(9) |
Aug
(11) |
Sep
(4) |
Oct
(2) |
Nov
(1) |
Dec
|
2017 |
Jan
(20) |
Feb
(7) |
Mar
(4) |
Apr
(18) |
May
(7) |
Jun
(3) |
Jul
(13) |
Aug
(2) |
Sep
(4) |
Oct
(9) |
Nov
(2) |
Dec
(5) |
2018 |
Jan
|
Feb
|
Mar
|
Apr
(2) |
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
2019 |
Jan
|
Feb
|
Mar
(1) |
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
From: Jim B. <ba...@ne...> - 2014-11-06 17:13:44
|
Oh no, I added the BIND to get rid of that. I'll redo and update you. > On Nov 6, 2014, at 12:01 PM, Bryan Thompson <br...@sy...> wrote: > > What happens if you replace that last line with: > > ORDER BY ?string_label > > rather than > > ORDER BY STR(?string_label) > > Remember, it is assuming that the ORDER BY is using simple variables. > > Bryan > > On Thu, Nov 6, 2014 at 11:58 AM, Jim Balhoff <ba...@ne...> wrote: > Here is the exact query (with or without DISTINCT) for the linked results: > > PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > SELECT DISTINCT ?term ?string_label > WHERE > { > ?term rdf:type owl:Class . > ?term rdfs:label ?term_label . > BIND (STR(?term_label) AS ?string_label) > } > ORDER BY STR(?string_label) > > > Results (same number of rows either way): > SELECT DISTINCT: > explain: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/with_distinct_explain.html > result: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/with_distinct_result.csv > > SELECT: > explain: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/no_distinct_explain.html > result: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/no_distinct_result.csv > > You can diff the two results files to see the out-of-order blocks. > > I suppose it does look like the DISTINCT query plan has ORDER BY applied before DISTINCT, if I am reading it right. > > Thanks, > Jim > > > > > > On Nov 6, 2014, at 10:10 AM, Bryan Thompson <br...@sy...> wrote: > > > > Jim, > > > > 502 is about support for expressions (other than simple variables in ORDER_BY). > > > > If there is an issue with DISTINCT + ORDER_BY then this would be a new ticket. > > > > Just post the EXPLAIN (attach to the email) for the moment. I want to see how this is being generated. We should then check the specification and make sure that the correct behavior is DISTINCT followed by ORDER BY with any limit applied after the ORDER BY. I can then check the code for how we are handling this. > > > > The relevant logic is in AST2BOpUtility at line 451. You can see that it is already attempting to handle this and that there was a historical ticket for this issue (#563). > > > > > > > > /* > > > > * Note: The DISTINCT operators also enforce the projection. > > > > * > > > > * Note: REDUCED allows, but does not require, either complete or > > > > * partial filtering of duplicates. It is part of what openrdf does > > > > * for a DESCRIBE query. > > > > * > > > > * Note: We do not currently have special operator for REDUCED. One > > > > * could be created using chunk wise DISTINCT. Note that REDUCED may > > > > * not change the order in which the solutions appear (but we are > > > > * evaluating it before ORDER BY so that is Ok.) > > > > * > > > > * TODO If there is an ORDER BY and a DISTINCT then the sort can be > > > > * used to impose the distinct without the overhead of a hash index > > > > * by filtering out the duplicate solutions after the sort. > > > > */ > > > > > > > > // When true, DISTINCT must preserve ORDER BY ordering. > > > > final boolean preserveOrder; > > > > > > > > if (orderBy != null && !orderBy.isEmpty()) { > > > > > > > > /* > > > > * Note: ORDER BY before DISTINCT, so DISTINCT must preserve > > > > * order. > > > > * > > > > * @see https://sourceforge.net/apps/trac/bigdata/ticket/563 > > > > * (ORDER BY + DISTINCT) > > > > */ > > > > > > preserveOrder = true; > > > > > > > > left = addOrderBy(left, queryBase, orderBy, ctx); > > > > > > > > } else { > > > > > > preserveOrder = false; > > > > > > } > > > > > > > > if (projection.isDistinct() || projection.isReduced()) { > > > > > > > > left = addDistinct(left, queryBase, preserveOrder, ctx); > > > > > > > > } > > > > > > > > } else { > > > > > > > > /* > > > > * TODO Under what circumstances can the projection be [null]? > > > > */ > > > > > > if (orderBy != null && !orderBy.isEmpty()) { > > > > > > > > left = addOrderBy(left, queryBase, orderBy, ctx); > > > > > > > > } > > > > > > > > } > > > > > > > > Bryan > > > > > > ---- > > Bryan Thompson > > Chief Scientist & Founder > > SYSTAP, LLC > > 4501 Tower Road > > Greensboro, NC 27410 > > br...@sy... > > http://bigdata.com > > http://mapgraph.io > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > > > > > On Thu, Nov 6, 2014 at 10:03 AM, Jim Balhoff <ba...@ne...> wrote: > > Hi Bryan, > > > > Just to clarify, would you like me to attach the info to ticket 502, or continue posting to the developer list? > > > > Thanks, > > Jim > > > > > > > On Nov 6, 2014, at 8:28 AM, Bryan Thompson <br...@sy...> wrote: > > > > > > The ticket for allowing aggregates in ORDER BY is: > > > > > > - http://trac.bigdata.com/ticket/502 (Allow aggregates in ORDER BY clause) > > > > > > Can you attach the EXPLAIN of the query with and without DISTINCT. The issue may be that the DISTINCT is being applied after the ORDER BY. I seem to remember some issue historically with operations being performed before/after the ORDER BY, but I do not have any distinct recollection of a problematic interaction between DISTINCT and ORDER BY. > > > > > > Bryan > > > > > > ---- > > > Bryan Thompson > > > Chief Scientist & Founder > > > SYSTAP, LLC > > > 4501 Tower Road > > > Greensboro, NC 27410 > > > br...@sy... > > > http://bigdata.com > > > http://mapgraph.io > > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > > > > > > > > > On Wed, Nov 5, 2014 at 6:14 PM, Jim Balhoff <ba...@ne...> wrote: > > > > On Nov 5, 2014, at 5:46 PM, Jeremy J Carroll <jj...@sy...> wrote: > > > > > > > > > > > >> On Nov 5, 2014, at 1:02 PM, Bryan Thompson <br...@sy...> wrote: > > > >> > > > >> There could be an issue with ORDER BY operating on an anonymous and non-projected variable. Try declaring and binding a variable for STR(?label) inside of the query and then using that variable in the ORDER BY clause. > > > > > > > > > > > > Yes I tend to find the results of ORDER BY are more what I expect if I do not include an expression in the ORDER BY but simply variables. I BIND any expression before the ORDER BY. > > > > > > > > I believe there is a trac item for this, but since the workaround is easy, I have never seen it as high priority > > > > > > > > > > As suggested I tried binding a variable as `BIND (STR(?term_label) AS ?string_label)` and using that to sort. Still incorrect ordering. But, I tried removing DISTINCT, and then the ordering is correct. Even going back to the anonymous `ORDER BY STR(?term_label)`, ordering is still correct if I remove DISTINCT. For this specific query DISTINCT is not needed, but I do need it for my application. Is there a reason to not expect DISTINCT to work correctly with ORDER BY? > > > > > > Thanks both of you for all of your help, > > > Jim > > > > > > > > > > > > |
From: Bryan T. <br...@sy...> - 2014-11-06 17:01:12
|
What happens if you replace that last line with: ORDER BY ?string_label rather than ORDER BY STR(?string_label) Remember, it is assuming that the ORDER BY is using simple variables. Bryan On Thu, Nov 6, 2014 at 11:58 AM, Jim Balhoff <ba...@ne...> wrote: > Here is the exact query (with or without DISTINCT) for the linked results: > > PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > SELECT DISTINCT ?term ?string_label > WHERE > { > ?term rdf:type owl:Class . > ?term rdfs:label ?term_label . > BIND (STR(?term_label) AS ?string_label) > } > ORDER BY STR(?string_label) > > > Results (same number of rows either way): > SELECT DISTINCT: > explain: > https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/with_distinct_explain.html > result: > https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/with_distinct_result.csv > > SELECT: > explain: > https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/no_distinct_explain.html > result: > https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/no_distinct_result.csv > > You can diff the two results files to see the out-of-order blocks. > > I suppose it does look like the DISTINCT query plan has ORDER BY applied > before DISTINCT, if I am reading it right. > > Thanks, > Jim > > > > > > On Nov 6, 2014, at 10:10 AM, Bryan Thompson <br...@sy...> wrote: > > > > Jim, > > > > 502 is about support for expressions (other than simple variables in > ORDER_BY). > > > > If there is an issue with DISTINCT + ORDER_BY then this would be a new > ticket. > > > > Just post the EXPLAIN (attach to the email) for the moment. I want to > see how this is being generated. We should then check the specification > and make sure that the correct behavior is DISTINCT followed by ORDER BY > with any limit applied after the ORDER BY. I can then check the code for > how we are handling this. > > > > The relevant logic is in AST2BOpUtility at line 451. You can see that > it is already attempting to handle this and that there was a historical > ticket for this issue (#563). > > > > > > > > /* > > > > * Note: The DISTINCT operators also enforce the projection. > > > > * > > > > * Note: REDUCED allows, but does not require, either > complete or > > > > * partial filtering of duplicates. It is part of what > openrdf does > > > > * for a DESCRIBE query. > > > > * > > > > * Note: We do not currently have special operator for > REDUCED. One > > > > * could be created using chunk wise DISTINCT. Note that > REDUCED may > > > > * not change the order in which the solutions appear (but > we are > > > > * evaluating it before ORDER BY so that is Ok.) > > > > * > > > > * TODO If there is an ORDER BY and a DISTINCT then the sort > can be > > > > * used to impose the distinct without the overhead of a > hash index > > > > * by filtering out the duplicate solutions after the sort. > > > > */ > > > > > > > > // When true, DISTINCT must preserve ORDER BY ordering. > > > > final boolean preserveOrder; > > > > > > > > if (orderBy != null && !orderBy.isEmpty()) { > > > > > > > > /* > > > > * Note: ORDER BY before DISTINCT, so DISTINCT must > preserve > > > > * order. > > > > * > > > > * @see > https://sourceforge.net/apps/trac/bigdata/ticket/563 > > > > * (ORDER BY + DISTINCT) > > > > */ > > > > > > preserveOrder = true; > > > > > > > > left = addOrderBy(left, queryBase, orderBy, ctx); > > > > > > > > } else { > > > > > > preserveOrder = false; > > > > > > } > > > > > > > > if (projection.isDistinct() || projection.isReduced()) { > > > > > > > > left = addDistinct(left, queryBase, preserveOrder, ctx); > > > > > > > > } > > > > > > > > } else { > > > > > > > > /* > > > > * TODO Under what circumstances can the projection be > [null]? > > > > */ > > > > > > if (orderBy != null && !orderBy.isEmpty()) { > > > > > > > > left = addOrderBy(left, queryBase, orderBy, ctx); > > > > > > > > } > > > > > > > > } > > > > > > > > Bryan > > > > > > ---- > > Bryan Thompson > > Chief Scientist & Founder > > SYSTAP, LLC > > 4501 Tower Road > > Greensboro, NC 27410 > > br...@sy... > > http://bigdata.com > > http://mapgraph.io > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > > > > On Thu, Nov 6, 2014 at 10:03 AM, Jim Balhoff <ba...@ne...> > wrote: > > Hi Bryan, > > > > Just to clarify, would you like me to attach the info to ticket 502, or > continue posting to the developer list? > > > > Thanks, > > Jim > > > > > > > On Nov 6, 2014, at 8:28 AM, Bryan Thompson <br...@sy...> wrote: > > > > > > The ticket for allowing aggregates in ORDER BY is: > > > > > > - http://trac.bigdata.com/ticket/502 (Allow aggregates in ORDER BY > clause) > > > > > > Can you attach the EXPLAIN of the query with and without DISTINCT. > The issue may be that the DISTINCT is being applied after the ORDER BY. I > seem to remember some issue historically with operations being performed > before/after the ORDER BY, but I do not have any distinct recollection of a > problematic interaction between DISTINCT and ORDER BY. > > > > > > Bryan > > > > > > ---- > > > Bryan Thompson > > > Chief Scientist & Founder > > > SYSTAP, LLC > > > 4501 Tower Road > > > Greensboro, NC 27410 > > > br...@sy... > > > http://bigdata.com > > > http://mapgraph.io > > > CONFIDENTIALITY NOTICE: This email and its contents and attachments > are for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > > > > > > > > On Wed, Nov 5, 2014 at 6:14 PM, Jim Balhoff <ba...@ne...> > wrote: > > > > On Nov 5, 2014, at 5:46 PM, Jeremy J Carroll <jj...@sy...> wrote: > > > > > > > > > > > >> On Nov 5, 2014, at 1:02 PM, Bryan Thompson <br...@sy...> > wrote: > > > >> > > > >> There could be an issue with ORDER BY operating on an anonymous and > non-projected variable. Try declaring and binding a variable for > STR(?label) inside of the query and then using that variable in the ORDER > BY clause. > > > > > > > > > > > > Yes I tend to find the results of ORDER BY are more what I expect if > I do not include an expression in the ORDER BY but simply variables. I BIND > any expression before the ORDER BY. > > > > > > > > I believe there is a trac item for this, but since the workaround is > easy, I have never seen it as high priority > > > > > > > > > > As suggested I tried binding a variable as `BIND (STR(?term_label) AS > ?string_label)` and using that to sort. Still incorrect ordering. But, I > tried removing DISTINCT, and then the ordering is correct. Even going back > to the anonymous `ORDER BY STR(?term_label)`, ordering is still correct if > I remove DISTINCT. For this specific query DISTINCT is not needed, but I do > need it for my application. Is there a reason to not expect DISTINCT to > work correctly with ORDER BY? > > > > > > Thanks both of you for all of your help, > > > Jim > > > > > > > > > > > > |
From: Jim B. <ba...@ne...> - 2014-11-06 16:58:12
|
Here is the exact query (with or without DISTINCT) for the linked results: PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> PREFIX owl: <http://www.w3.org/2002/07/owl#> SELECT DISTINCT ?term ?string_label WHERE { ?term rdf:type owl:Class . ?term rdfs:label ?term_label . BIND (STR(?term_label) AS ?string_label) } ORDER BY STR(?string_label) Results (same number of rows either way): SELECT DISTINCT: explain: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/with_distinct_explain.html result: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/with_distinct_result.csv SELECT: explain: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/no_distinct_explain.html result: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-6/no_distinct_result.csv You can diff the two results files to see the out-of-order blocks. I suppose it does look like the DISTINCT query plan has ORDER BY applied before DISTINCT, if I am reading it right. Thanks, Jim > On Nov 6, 2014, at 10:10 AM, Bryan Thompson <br...@sy...> wrote: > > Jim, > > 502 is about support for expressions (other than simple variables in ORDER_BY). > > If there is an issue with DISTINCT + ORDER_BY then this would be a new ticket. > > Just post the EXPLAIN (attach to the email) for the moment. I want to see how this is being generated. We should then check the specification and make sure that the correct behavior is DISTINCT followed by ORDER BY with any limit applied after the ORDER BY. I can then check the code for how we are handling this. > > The relevant logic is in AST2BOpUtility at line 451. You can see that it is already attempting to handle this and that there was a historical ticket for this issue (#563). > > > > /* > > * Note: The DISTINCT operators also enforce the projection. > > * > > * Note: REDUCED allows, but does not require, either complete or > > * partial filtering of duplicates. It is part of what openrdf does > > * for a DESCRIBE query. > > * > > * Note: We do not currently have special operator for REDUCED. One > > * could be created using chunk wise DISTINCT. Note that REDUCED may > > * not change the order in which the solutions appear (but we are > > * evaluating it before ORDER BY so that is Ok.) > > * > > * TODO If there is an ORDER BY and a DISTINCT then the sort can be > > * used to impose the distinct without the overhead of a hash index > > * by filtering out the duplicate solutions after the sort. > > */ > > > > // When true, DISTINCT must preserve ORDER BY ordering. > > final boolean preserveOrder; > > > > if (orderBy != null && !orderBy.isEmpty()) { > > > > /* > > * Note: ORDER BY before DISTINCT, so DISTINCT must preserve > > * order. > > * > > * @see https://sourceforge.net/apps/trac/bigdata/ticket/563 > > * (ORDER BY + DISTINCT) > > */ > > > preserveOrder = true; > > > > left = addOrderBy(left, queryBase, orderBy, ctx); > > > > } else { > > > preserveOrder = false; > > > } > > > > if (projection.isDistinct() || projection.isReduced()) { > > > > left = addDistinct(left, queryBase, preserveOrder, ctx); > > > > } > > > > } else { > > > > /* > > * TODO Under what circumstances can the projection be [null]? > > */ > > > if (orderBy != null && !orderBy.isEmpty()) { > > > > left = addOrderBy(left, queryBase, orderBy, ctx); > > > > } > > > > } > > > > Bryan > > > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... > http://bigdata.com > http://mapgraph.io > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > On Thu, Nov 6, 2014 at 10:03 AM, Jim Balhoff <ba...@ne...> wrote: > Hi Bryan, > > Just to clarify, would you like me to attach the info to ticket 502, or continue posting to the developer list? > > Thanks, > Jim > > > > On Nov 6, 2014, at 8:28 AM, Bryan Thompson <br...@sy...> wrote: > > > > The ticket for allowing aggregates in ORDER BY is: > > > > - http://trac.bigdata.com/ticket/502 (Allow aggregates in ORDER BY clause) > > > > Can you attach the EXPLAIN of the query with and without DISTINCT. The issue may be that the DISTINCT is being applied after the ORDER BY. I seem to remember some issue historically with operations being performed before/after the ORDER BY, but I do not have any distinct recollection of a problematic interaction between DISTINCT and ORDER BY. > > > > Bryan > > > > ---- > > Bryan Thompson > > Chief Scientist & Founder > > SYSTAP, LLC > > 4501 Tower Road > > Greensboro, NC 27410 > > br...@sy... > > http://bigdata.com > > http://mapgraph.io > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > > > > > On Wed, Nov 5, 2014 at 6:14 PM, Jim Balhoff <ba...@ne...> wrote: > > > On Nov 5, 2014, at 5:46 PM, Jeremy J Carroll <jj...@sy...> wrote: > > > > > > > > >> On Nov 5, 2014, at 1:02 PM, Bryan Thompson <br...@sy...> wrote: > > >> > > >> There could be an issue with ORDER BY operating on an anonymous and non-projected variable. Try declaring and binding a variable for STR(?label) inside of the query and then using that variable in the ORDER BY clause. > > > > > > > > > Yes I tend to find the results of ORDER BY are more what I expect if I do not include an expression in the ORDER BY but simply variables. I BIND any expression before the ORDER BY. > > > > > > I believe there is a trac item for this, but since the workaround is easy, I have never seen it as high priority > > > > > > > As suggested I tried binding a variable as `BIND (STR(?term_label) AS ?string_label)` and using that to sort. Still incorrect ordering. But, I tried removing DISTINCT, and then the ordering is correct. Even going back to the anonymous `ORDER BY STR(?term_label)`, ordering is still correct if I remove DISTINCT. For this specific query DISTINCT is not needed, but I do need it for my application. Is there a reason to not expect DISTINCT to work correctly with ORDER BY? > > > > Thanks both of you for all of your help, > > Jim > > > > > > |
From: Bryan T. <br...@sy...> - 2014-11-06 15:36:31
|
This is a critical fix release of bigdata(R). All users are encouraged to upgrade immediately. Bigdata is a horizontally-scaled, open-source architecture for indexed data with an emphasis on RDF capable of loading 1B triples in under one hour on a 15 node cluster. Bigdata operates in both a single machine mode (Journal), highly available replication cluster mode (HAJournalServer), and a horizontally sharded cluster mode (BigdataFederation). The Journal provides fast scalable ACID indexed storage for very large data sets, up to 50 billion triples / quads. The HAJournalServer adds replication, online backup, horizontal scaling of query, and high availability. The federation provides fast scalable shard-wise parallel indexed storage using dynamic sharding and shard-wise ACID updates and incremental cluster size growth. Both platforms support fully concurrent readers with snapshot isolation. Distributed processing offers greater throughput but does not reduce query or update latency. Choose the Journal when the anticipated scale and throughput requirements permit. Choose the HAJournalServer for high availability and linear scaling in query throughput. Choose the BigdataFederation when the administrative and machine overhead associated with operating a cluster is an acceptable tradeoff to have essentially unlimited data scaling and throughput. See [1,2,8] for instructions on installing bigdata(R), [4] for the javadoc, and [3,5,6] for news, questions, and the latest developments. For more information about SYSTAP, LLC and bigdata, see [7]. Starting with the 1.0.0 release, we offer a WAR artifact [8] for easy installation of the single machine RDF database. For custom development and cluster installations we recommend checking out the code from SVN using the tag for this release. The code will build automatically under eclipse. You can also build the code using the ant script. The cluster installer requires the use of the ant script. Starting with the 1.3.0 release, we offer a tarball artifact [10] for easy installation of the HA replication cluster. You can download the WAR (standalone) or HA artifacts from: http://sourceforge.net/projects/bigdata/ You can checkout this release from: https://svn.code.sf.net/p/bigdata/code/tags/BIGDATA_RELEASE_1_3_4 Critical or otherwise of note in this minor release: *- #1036 (Journal leaks storage with SPARQL UPDATE and REST API)* New features in 1.3.x: - Java 7 is now required. - High availability [10]. - High availability load balancer. - New RDF/SPARQL workbench. - Blueprints API. - RDF Graph Mining Service (GASService) [12]. - Reification Done Right (RDR) support [11]. - Property Path performance enhancements. - Plus numerous other bug fixes and performance enhancements. Feature summary: - Highly Available Replication Clusters (HAJournalServer [10]) - Single machine data storage to ~50B triples/quads (RWStore); - Clustered data storage is essentially unlimited (BigdataFederation); - Simple embedded and/or webapp deployment (NanoSparqlServer); - Triples, quads, or triples with provenance (SIDs); - Fast RDFS+ inference and truth maintenance; - Fast 100% native SPARQL 1.1 evaluation; - Integrated "analytic" query package; - %100 Java memory manager leverages the JVM native heap (no GC); Road map [3]: - Column-wise indexing; - Runtime Query Optimizer for quads; - Performance optimization for scale-out clusters; and - Simplified deployment, configuration, and administration for scale-out clusters. Change log: Note: Versions with (*) MAY require data migration. For details, see [9]. 1.3.4: - http://trac.bigdata.com/ticket/946 (Empty PROJECTION causes IllegalArgumentException) - http://trac.bigdata.com/ticket/1036 (Journal leaks storage with SPARQL UPDATE and REST API) - http://trac.bigdata.com/ticket/1008 (remote service queries should put parameters in the request body when using POST) 1.3.3: - http://trac.bigdata.com/ticket/980 (Object position of query hint is not a Literal (partial resolution - see #1028 as well)) - http://trac.bigdata.com/ticket/1018 (Add the ability to track and cancel all queries issued through a BigdataSailRemoteRepositoryConnection) - http://trac.bigdata.com/ticket/1021 (Add critical section protection to AbstractJournal.abort() and BigdataSailConnection.rollback()) - http://trac.bigdata.com/ticket/1024 (GregorianCalendar? does weird things before 1582) - http://trac.bigdata.com/ticket/1026 (SPARQL UPDATE with runtime errors causes problems with lexicon indices) - http://trac.bigdata.com/ticket/1028 (very rare NotMaterializedException: XSDBoolean(true)) - http://trac.bigdata.com/ticket/1029 (RWStore commit state not correctly rolled back if abort fails on empty journal) - http://trac.bigdata.com/ticket/1030 (RWStorage stats cleanup) 1.3.2: - http://trac.bigdata.com/ticket/1016 (Jetty/LBS issues when deployed as WAR under tomcat) - http://trac.bigdata.com/ticket/1010 (Upgrade apache http components to 1.3.1 (security)) - http://trac.bigdata.com/ticket/1005 (Invalidate BTree objects if error occurs during eviction) - http://trac.bigdata.com/ticket/1004 (Concurrent binding problem) - http://trac.bigdata.com/ticket/1002 (Concurrency issues in JVMHashJoinUtility caused by MAX_PARALLEL query hint override) - http://trac.bigdata.com/ticket/1000 (Add configuration option to turn off bottom-up evaluation) - http://trac.bigdata.com/ticket/999 (Extend BigdataSailFactory to take arbitrary properties) - http://trac.bigdata.com/ticket/998 (SPARQL Update through BigdataGraph) - http://trac.bigdata.com/ticket/996 (Add custom prefix support for query results) - http://trac.bigdata.com/ticket/995 (Allow general purpose SPARQL queries through BigdataGraph) - http://trac.bigdata.com/ticket/992 (Deadlock between AbstractRunningQuery.cancel(), QueryLog.log(), and ArbitraryLengthPathTask) - http://trac.bigdata.com/ticket/990 (Query hints not recognized in FILTERs) - http://trac.bigdata.com/ticket/989 (Stored query service) - http://trac.bigdata.com/ticket/988 (Bad performance for FILTER EXISTS) - http://trac.bigdata.com/ticket/987 (maven build is broken) - http://trac.bigdata.com/ticket/986 (Improve locality for small allocation slots) - http://trac.bigdata.com/ticket/985 (Deadlock in BigdataTriplePatternMaterializer) - http://trac.bigdata.com/ticket/975 (HA Health Status Page) - http://trac.bigdata.com/ticket/974 (Name2Addr.indexNameScan(prefix) uses scan + filter) - http://trac.bigdata.com/ticket/973 (RWStore.commit() should be more defensive) - http://trac.bigdata.com/ticket/971 (Clarify HTTP Status codes for CREATE NAMESPACE operation) - http://trac.bigdata.com/ticket/968 (no link to wiki from workbench) - http://trac.bigdata.com/ticket/966 (Failed to get namespace under concurrent update) - http://trac.bigdata.com/ticket/965 (Can not run LBS mode with HA1 setup) - http://trac.bigdata.com/ticket/961 (Clone/modify namespace to create a new one) - http://trac.bigdata.com/ticket/960 (Export namespace properties in XML/Java properties text format) - http://trac.bigdata.com/ticket/938 (HA Load Balancer) - http://trac.bigdata.com/ticket/936 (Support larger metabits allocations) - http://trac.bigdata.com/ticket/932 (Bigdata/Rexster integration) - http://trac.bigdata.com/ticket/919 (Formatted Layout for Status pages) - http://trac.bigdata.com/ticket/899 (REST API Query Cancellation) - http://trac.bigdata.com/ticket/885 (Panels do not appear on startup in Firefox) - http://trac.bigdata.com/ticket/884 (Executing a new query should clear the old query results from the console) - http://trac.bigdata.com/ticket/882 (Abbreviate URIs that can be namespaced with one of the defined common namespaces) - http://trac.bigdata.com/ticket/880 (Can't explore an absolute URI with < >) - http://trac.bigdata.com/ticket/878 (Explore page looks weird when empty) - http://trac.bigdata.com/ticket/873 (Allow user to go use browser back & forward buttons to view explore history) - http://trac.bigdata.com/ticket/865 (OutOfMemoryError instead of Timeout for SPARQL Property Paths) - http://trac.bigdata.com/ticket/858 (Change explore URLs to include URI being clicked so user can see what they've clicked on before) - http://trac.bigdata.com/ticket/855 (AssertionError: Child does not have persistent identity) - http://trac.bigdata.com/ticket/850 (Search functionality in workbench) - http://trac.bigdata.com/ticket/847 (Query results panel should recognize well known namespaces for easier reading) - http://trac.bigdata.com/ticket/845 (Display the properties for a namespace) - http://trac.bigdata.com/ticket/843 (Create new tabs for status & performance counters, and add per namespace service/VoID description links) - http://trac.bigdata.com/ticket/837 (Configurator for new namespaces) - http://trac.bigdata.com/ticket/836 (Allow user to create namespace in the workbench) - http://trac.bigdata.com/ticket/830 (Output RDF data from queries in table format) - http://trac.bigdata.com/ticket/829 (Export query results) - http://trac.bigdata.com/ticket/828 (Save selected namespace in browser) - http://trac.bigdata.com/ticket/827 (Explore tab in workbench) - http://trac.bigdata.com/ticket/826 (Create shortcut to execute load/query) - http://trac.bigdata.com/ticket/823 (Disable textarea when a large file is selected) - http://trac.bigdata.com/ticket/820 (Allow non-file:// URLs to be loaded) - http://trac.bigdata.com/ticket/819 (Retrieve default namespace on page load) - http://trac.bigdata.com/ticket/772 (Query timeout only checked at operator start/stop) - http://trac.bigdata.com/ticket/765 (order by expr skips invalid expressions) - http://trac.bigdata.com/ticket/587 (JSP page to configure KBs) - http://trac.bigdata.com/ticket/343 (Stochastic assert in AbstractBTree#writeNodeOrLeaf() in CI) 1.3.1: - http://trac.bigdata.com/ticket/242 (Deadlines do not play well with GROUP_BY, ORDER_BY, etc.) - http://trac.bigdata.com/ticket/256 (Amortize RTO cost) - http://trac.bigdata.com/ticket/257 (Support BOP fragments in the RTO.) - http://trac.bigdata.com/ticket/258 (Integrate RTO into SAIL) - http://trac.bigdata.com/ticket/259 (Dynamically increase RTO sampling limit.) - http://trac.bigdata.com/ticket/526 (Reification done right) - http://trac.bigdata.com/ticket/580 (Problem with the bigdata RDF/XML parser with sids) - http://trac.bigdata.com/ticket/622 (NSS using jetty+windows can lose connections (windows only; jdk 6/7 bug)) - http://trac.bigdata.com/ticket/624 (HA Load Balancer) - http://trac.bigdata.com/ticket/629 (Graph processing API) - http://trac.bigdata.com/ticket/721 (Support HA1 configurations) - http://trac.bigdata.com/ticket/730 (Allow configuration of embedded NSS jetty server using jetty-web.xml) - http://trac.bigdata.com/ticket/759 (multiple filters interfere) - http://trac.bigdata.com/ticket/763 (Stochastic results with Analytic Query Mode) - http://trac.bigdata.com/ticket/774 (Converge on Java 7.) - http://trac.bigdata.com/ticket/779 (Resynchronization of socket level write replication protocol (HA)) - http://trac.bigdata.com/ticket/780 (Incremental or asynchronous purge of HALog files) - http://trac.bigdata.com/ticket/782 (Wrong serialization version) - http://trac.bigdata.com/ticket/784 (Describe Limit/offset don't work as expected) - http://trac.bigdata.com/ticket/787 (Update documentations and samples, they are OUTDATED) - http://trac.bigdata.com/ticket/788 (Name2Addr does not report all root causes if the commit fails.) - http://trac.bigdata.com/ticket/789 (ant task to build sesame fails, docs for setting up bigdata for sesame are ancient) - http://trac.bigdata.com/ticket/790 (should not be pruning any children) - http://trac.bigdata.com/ticket/791 (Clean up query hints) - http://trac.bigdata.com/ticket/793 (Explain reports incorrect value for opCount) - http://trac.bigdata.com/ticket/796 (Filter assigned to sub-query by query generator is dropped from evaluation) - http://trac.bigdata.com/ticket/797 (add sbt setup to getting started wiki) - http://trac.bigdata.com/ticket/798 (Solution order not always preserved) - http://trac.bigdata.com/ticket/799 (mis-optimation of quad pattern vs triple pattern) - http://trac.bigdata.com/ticket/802 (Optimize DatatypeFactory instantiation in DateTimeExtension) - http://trac.bigdata.com/ticket/803 (prefixMatch does not work in full text search) - http://trac.bigdata.com/ticket/804 (update bug deleting quads) - http://trac.bigdata.com/ticket/806 (Incorrect AST generated for OPTIONAL { SELECT }) - http://trac.bigdata.com/ticket/808 (Wildcard search in bigdata for type suggessions) - http://trac.bigdata.com/ticket/810 (Expose GAS API as SPARQL SERVICE) - http://trac.bigdata.com/ticket/815 (RDR query does too much work) - http://trac.bigdata.com/ticket/816 (Wildcard projection ignores variables inside a SERVICE call.) - http://trac.bigdata.com/ticket/817 (Unexplained increase in journal size) - http://trac.bigdata.com/ticket/821 (Reject large files, rather then storing them in a hidden variable) - http://trac.bigdata.com/ticket/831 (UNION with filter issue) - http://trac.bigdata.com/ticket/841 (Using "VALUES" in a query returns lexical error) - http://trac.bigdata.com/ticket/848 (Fix SPARQL Results JSON writer to write the RDR syntax) - http://trac.bigdata.com/ticket/849 (Create writers that support the RDR syntax) - http://trac.bigdata.com/ticket/851 (RDR GAS interface) - http://trac.bigdata.com/ticket/852 (RemoteRepository.cancel() does not consume the HTTP response entity.) - http://trac.bigdata.com/ticket/853 (Follower does not accept POST of idempotent operations (HA)) - http://trac.bigdata.com/ticket/854 (Allow override of maximum length before converting an HTTP GET to an HTTP POST) - http://trac.bigdata.com/ticket/855 (AssertionError: Child does not have persistent identity) - http://trac.bigdata.com/ticket/862 (Create parser for JSON SPARQL Results) - http://trac.bigdata.com/ticket/863 (HA1 commit failure) - http://trac.bigdata.com/ticket/866 (Batch remove API for the SAIL) - http://trac.bigdata.com/ticket/867 (NSS concurrency problem with list namespaces and create namespace) - http://trac.bigdata.com/ticket/869 (HA5 test suite) - http://trac.bigdata.com/ticket/872 (Full text index range count optimization) - http://trac.bigdata.com/ticket/874 (FILTER not applied when there is UNION in the same join group) - http://trac.bigdata.com/ticket/876 (When I upload a file I want to see the filename.) - http://trac.bigdata.com/ticket/877 (RDF Format selector is invisible) - http://trac.bigdata.com/ticket/883 (CANCEL Query fails on non-default kb namespace on HA follower.) - http://trac.bigdata.com/ticket/886 (Provide workaround for bad reverse DNS setups.) - http://trac.bigdata.com/ticket/887 (BIND is leaving a variable unbound) - http://trac.bigdata.com/ticket/892 (HAJournalServer does not die if zookeeper is not running) - http://trac.bigdata.com/ticket/893 (large sparql insert optimization slow?) - http://trac.bigdata.com/ticket/894 (unnecessary synchronization) - http://trac.bigdata.com/ticket/895 (stack overflow in populateStatsMap) - http://trac.bigdata.com/ticket/902 (Update Basic Bigdata Chef Cookbook) - http://trac.bigdata.com/ticket/904 (AssertionError: PropertyPathNode got to ASTJoinOrderByType.optimizeJoinGroup) - http://trac.bigdata.com/ticket/905 (unsound combo query optimization: union + filter) - http://trac.bigdata.com/ticket/906 (DC Prefix Button Appends "</li>") - http://trac.bigdata.com/ticket/907 (Add a quick-start ant task for the BD Server "ant start") - http://trac.bigdata.com/ticket/912 (Provide a configurable IAnalyzerFactory) - http://trac.bigdata.com/ticket/913 (Blueprints API Implementation) - http://trac.bigdata.com/ticket/914 (Settable timeout on SPARQL Query (REST API)) - http://trac.bigdata.com/ticket/915 (DefaultAnalyzerFactory issues) - http://trac.bigdata.com/ticket/920 (Content negotiation orders accept header scores in reverse) - http://trac.bigdata.com/ticket/939 (NSS does not start from command line: bigdata-war/src not found.) - http://trac.bigdata.com/ticket/940 (ProxyServlet in web.xml breaks tomcat WAR (HA LBS) 1.3.0: - http://trac.bigdata.com/ticket/530 (Journal HA) - http://trac.bigdata.com/ticket/621 (Coalesce write cache records and install reads in cache) - http://trac.bigdata.com/ticket/623 (HA TXS) - http://trac.bigdata.com/ticket/639 (Remove triple-buffering in RWStore) - http://trac.bigdata.com/ticket/645 (HA backup) - http://trac.bigdata.com/ticket/646 (River not compatible with newer 1.6.0 and 1.7.0 JVMs) - http://trac.bigdata.com/ticket/648 (Add a custom function to use full text index for filtering.) - http://trac.bigdata.com/ticket/651 (RWS test failure) - http://trac.bigdata.com/ticket/652 (Compress write cache blocks for replication and in HALogs) - http://trac.bigdata.com/ticket/662 (Latency on followers during commit on leader) - http://trac.bigdata.com/ticket/663 (Issue with OPTIONAL blocks) - http://trac.bigdata.com/ticket/664 (RWStore needs post-commit protocol) - http://trac.bigdata.com/ticket/665 (HA3 LOAD non-responsive with node failure) - http://trac.bigdata.com/ticket/666 (Occasional CI deadlock in HALogWriter testConcurrentRWWriterReader) - http://trac.bigdata.com/ticket/670 (Accumulating HALog files cause latency for HA commit) - http://trac.bigdata.com/ticket/671 (Query on follower fails during UPDATE on leader) - http://trac.bigdata.com/ticket/673 (DGC in release time consensus protocol causes native thread leak in HAJournalServer at each commit) - http://trac.bigdata.com/ticket/674 (WCS write cache compaction causes errors in RWS postHACommit()) - http://trac.bigdata.com/ticket/676 (Bad patterns for timeout computations) - http://trac.bigdata.com/ticket/677 (HA deadlock under UPDATE + QUERY) - http://trac.bigdata.com/ticket/678 (DGC Thread and Open File Leaks: sendHALogForWriteSet()) - http://trac.bigdata.com/ticket/679 (HAJournalServer can not restart due to logically empty log file) - http://trac.bigdata.com/ticket/681 (HAJournalServer deadlock: pipelineRemove() and getLeaderId()) - http://trac.bigdata.com/ticket/684 (Optimization with skos altLabel) - http://trac.bigdata.com/ticket/686 (Consensus protocol does not detect clock skew correctly) - http://trac.bigdata.com/ticket/687 (HAJournalServer Cache not populated) - http://trac.bigdata.com/ticket/689 (Missing URL encoding in RemoteRepositoryManager) - http://trac.bigdata.com/ticket/690 (Error when using the alias "a" instead of rdf:type for a multipart insert) - http://trac.bigdata.com/ticket/691 (Failed to re-interrupt thread in HAJournalServer) - http://trac.bigdata.com/ticket/692 (Failed to re-interrupt thread) - http://trac.bigdata.com/ticket/693 (OneOrMorePath SPARQL property path expression ignored) - http://trac.bigdata.com/ticket/694 (Transparently cancel update/query in RemoteRepository) - http://trac.bigdata.com/ticket/695 (HAJournalServer reports "follower" but is in SeekConsensus and is not participating in commits.) - http://trac.bigdata.com/ticket/701 (Problems in BackgroundTupleResult) - http://trac.bigdata.com/ticket/702 (InvocationTargetException on / namespace call) - http://trac.bigdata.com/ticket/704 (ask does not return json) - http://trac.bigdata.com/ticket/705 (Race between QueryEngine.putIfAbsent() and shutdownNow()) - http://trac.bigdata.com/ticket/706 (MultiSourceSequentialCloseableIterator.nextSource() can throw NPE) - http://trac.bigdata.com/ticket/707 (BlockingBuffer.close() does not unblock threads) - http://trac.bigdata.com/ticket/708 (BIND heisenbug - race condition on select query with BIND) - http://trac.bigdata.com/ticket/711 (sparql protocol: mime type application/sparql-query) - http://trac.bigdata.com/ticket/712 (SELECT ?x { OPTIONAL { ?x eg:doesNotExist eg:doesNotExist } } incorrect) - http://trac.bigdata.com/ticket/715 (Interrupt of thread submitting a query for evaluation does not always terminate the AbstractRunningQuery) - http://trac.bigdata.com/ticket/716 (Verify that IRunningQuery instances (and nested queries) are correctly cancelled when interrupted) - http://trac.bigdata.com/ticket/718 (HAJournalServer needs to handle ZK client connection loss) - http://trac.bigdata.com/ticket/720 (HA3 simultaneous service start failure) - http://trac.bigdata.com/ticket/723 (HA asynchronous tasks must be canceled when invariants are changed) - http://trac.bigdata.com/ticket/725 (FILTER EXISTS in subselect) - http://trac.bigdata.com/ticket/726 (Logically empty HALog for committed transaction) - http://trac.bigdata.com/ticket/727 (DELETE/INSERT fails with OPTIONAL non-matching WHERE) - http://trac.bigdata.com/ticket/728 (Refactor to create HAClient) - http://trac.bigdata.com/ticket/729 (ant bundleJar not working) - http://trac.bigdata.com/ticket/731 (CBD and Update leads to 500 status code) - http://trac.bigdata.com/ticket/732 (describe statement limit does not work) - http://trac.bigdata.com/ticket/733 (Range optimizer not optimizing Slice service) - http://trac.bigdata.com/ticket/734 (two property paths interfere) - http://trac.bigdata.com/ticket/736 (MIN() malfunction) - http://trac.bigdata.com/ticket/737 (class cast exception) - http://trac.bigdata.com/ticket/739 (Inconsistent treatment of bind and optional property path) - http://trac.bigdata.com/ticket/741 (ctc-striterators should build as independent top-level project (Apache2)) - http://trac.bigdata.com/ticket/743 (AbstractTripleStore.destroy() does not filter for correct prefix) - http://trac.bigdata.com/ticket/746 (Assertion error) - http://trac.bigdata.com/ticket/747 (BOUND bug) - http://trac.bigdata.com/ticket/748 (incorrect join with subselect renaming vars) - http://trac.bigdata.com/ticket/754 (Failure to setup SERVICE hook and changeLog for Unisolated and Read/Write connections) - http://trac.bigdata.com/ticket/755 (Concurrent QuorumActors can interfere leading to failure to progress) - http://trac.bigdata.com/ticket/756 (order by and group_concat) - http://trac.bigdata.com/ticket/760 (Code review on 2-phase commit protocol) - http://trac.bigdata.com/ticket/764 (RESYNC failure (HA)) - http://trac.bigdata.com/ticket/770 (alpp ordering) - http://trac.bigdata.com/ticket/772 (Query timeout only checked at operator start/stop.) - http://trac.bigdata.com/ticket/776 (Closed as duplicate of #490) - http://trac.bigdata.com/ticket/778 (HA Leader fail results in transient problem with allocations on other services) - http://trac.bigdata.com/ticket/783 (Operator Alerts (HA)) 1.2.4: - http://trac.bigdata.com/ticket/777 (ConcurrentModificationException in ASTComplexOptionalOptimizer) 1.2.3: - http://trac.bigdata.com/ticket/168 (Maven Build) - http://trac.bigdata.com/ticket/196 (Journal leaks memory). - http://trac.bigdata.com/ticket/235 (Occasional deadlock in CI runs in com.bigdata.io.writecache.TestAll) - http://trac.bigdata.com/ticket/312 (CI (mock) quorums deadlock) - http://trac.bigdata.com/ticket/405 (Optimize hash join for subgroups with no incoming bound vars.) - http://trac.bigdata.com/ticket/412 (StaticAnalysis#getDefinitelyBound() ignores exogenous variables.) - http://trac.bigdata.com/ticket/485 (RDFS Plus Profile) - http://trac.bigdata.com/ticket/495 (SPARQL 1.1 Property Paths) - http://trac.bigdata.com/ticket/519 (Negative parser tests) - http://trac.bigdata.com/ticket/531 (SPARQL UPDATE for SOLUTION SETS) - http://trac.bigdata.com/ticket/535 (Optimize JOIN VARS for Sub-Selects) - http://trac.bigdata.com/ticket/555 (Support PSOutputStream/InputStream at IRawStore) - http://trac.bigdata.com/ticket/559 (Use RDFFormat.NQUADS as the format identifier for the NQuads parser) - http://trac.bigdata.com/ticket/570 (MemoryManager Journal does not implement all methods). - http://trac.bigdata.com/ticket/575 (NSS Admin API) - http://trac.bigdata.com/ticket/577 (DESCRIBE with OFFSET/LIMIT needs to use sub-select) - http://trac.bigdata.com/ticket/578 (Concise Bounded Description (CBD)) - http://trac.bigdata.com/ticket/579 (CONSTRUCT should use distinct SPO filter) - http://trac.bigdata.com/ticket/583 (VoID in ServiceDescription) - http://trac.bigdata.com/ticket/586 (RWStore immedateFree() not removing Checkpoint addresses from the historical index cache.) - http://trac.bigdata.com/ticket/590 (nxparser fails with uppercase language tag) - http://trac.bigdata.com/ticket/592 (Optimize RWStore allocator sizes) - http://trac.bigdata.com/ticket/593 (Ugrade to Sesame 2.6.10) - http://trac.bigdata.com/ticket/594 (WAR was deployed using TRIPLES rather than QUADS by default) - http://trac.bigdata.com/ticket/596 (Change web.xml parameter names to be consistent with Jini/River) - http://trac.bigdata.com/ticket/597 (SPARQL UPDATE LISTENER) - http://trac.bigdata.com/ticket/598 (B+Tree branching factor and HTree addressBits are confused in their NodeSerializer implementations) - http://trac.bigdata.com/ticket/599 (BlobIV for blank node : NotMaterializedException) - http://trac.bigdata.com/ticket/600 (BlobIV collision counter hits false limit.) - http://trac.bigdata.com/ticket/601 (Log uncaught exceptions) - http://trac.bigdata.com/ticket/602 (RWStore does not discard logged deletes on reset()) - http://trac.bigdata.com/ticket/607 (History service / index) - http://trac.bigdata.com/ticket/608 (LOG BlockingBuffer not progressing at INFO or lower level) - http://trac.bigdata.com/ticket/609 (bigdata-ganglia is required dependency for Journal) - http://trac.bigdata.com/ticket/611 (The code that processes SPARQL Update has a typo) - http://trac.bigdata.com/ticket/612 (Bigdata scale-up depends on zookeper) - http://trac.bigdata.com/ticket/613 (SPARQL UPDATE response inlines large DELETE or INSERT triple graphs) - http://trac.bigdata.com/ticket/614 (static join optimizer does not get ordering right when multiple tails share vars with ancestry) - http://trac.bigdata.com/ticket/615 (AST2BOpUtility wraps UNION with an unnecessary hash join) - http://trac.bigdata.com/ticket/616 (Row store read/update not isolated on Journal) - http://trac.bigdata.com/ticket/617 (Concurrent KB create fails with "No axioms defined?") - http://trac.bigdata.com/ticket/618 (DirectBufferPool.poolCapacity maximum of 2GB) - http://trac.bigdata.com/ticket/619 (RemoteRepository class should use application/x-www-form-urlencoded for large POST requests) - http://trac.bigdata.com/ticket/620 (UpdateServlet fails to parse MIMEType when doing conneg.) - http://trac.bigdata.com/ticket/626 (Expose performance counters for read-only indices) - http://trac.bigdata.com/ticket/627 (Environment variable override for NSS properties file) - http://trac.bigdata.com/ticket/628 (Create a bigdata-client jar for the NSS REST API) - http://trac.bigdata.com/ticket/631 (ClassCastException in SIDs mode query) - http://trac.bigdata.com/ticket/632 (NotMaterializedException when a SERVICE call needs variables that are provided as query input bindings) - http://trac.bigdata.com/ticket/633 (ClassCastException when binding non-uri values to a variable that occurs in predicate position) - http://trac.bigdata.com/ticket/638 (Change DEFAULT_MIN_RELEASE_AGE to 1ms) - http://trac.bigdata.com/ticket/640 (Conditionally rollback() BigdataSailConnection if dirty) - http://trac.bigdata.com/ticket/642 (Property paths do not work inside of exists/not exists filters) - http://trac.bigdata.com/ticket/643 (Add web.xml parameters to lock down public NSS end points) - http://trac.bigdata.com/ticket/644 (Bigdata2Sesame2BindingSetIterator can fail to notice asynchronous close()) - http://trac.bigdata.com/ticket/650 (Can not POST RDF to a graph using REST API) - http://trac.bigdata.com/ticket/654 (Rare AssertionError in WriteCache.clearAddrMap()) - http://trac.bigdata.com/ticket/655 (SPARQL REGEX operator does not perform case-folding correctly for Unicode data) - http://trac.bigdata.com/ticket/656 (InFactory bug when IN args consist of a single literal) - http://trac.bigdata.com/ticket/647 (SIDs mode creates unnecessary hash join for GRAPH group patterns) - http://trac.bigdata.com/ticket/667 (Provide NanoSparqlServer initialization hook) - http://trac.bigdata.com/ticket/669 (Doubly nested subqueries yield no results with LIMIT) - http://trac.bigdata.com/ticket/675 (Flush indices in parallel during checkpoint to reduce IO latency) - http://trac.bigdata.com/ticket/682 (AtomicRowFilter UnsupportedOperationException) 1.2.2: - http://trac.bigdata.com/ticket/586 (RWStore immedateFree() not removing Checkpoint addresses from the historical index cache.) - http://trac.bigdata.com/ticket/602 (RWStore does not discard logged deletes on reset()) - http://trac.bigdata.com/ticket/603 (Prepare critical maintenance release as branch of 1.2.1) 1.2.1: - http://trac.bigdata.com/ticket/533 (Review materialization for inline IVs) - http://trac.bigdata.com/ticket/539 (NotMaterializedException with REGEX and Vocab) - http://trac.bigdata.com/ticket/540 (SPARQL UPDATE using NSS via index.html) - http://trac.bigdata.com/ticket/541 (MemoryManaged backed Journal mode) - http://trac.bigdata.com/ticket/546 (Index cache for Journal) - http://trac.bigdata.com/ticket/549 (BTree can not be cast to Name2Addr (MemStore recycler)) - http://trac.bigdata.com/ticket/550 (NPE in Leaf.getKey() : root cause was user error) - http://trac.bigdata.com/ticket/558 (SPARQL INSERT not working in same request after INSERT DATA) - http://trac.bigdata.com/ticket/562 (Sub-select in INSERT cause NPE in UpdateExprBuilder) - http://trac.bigdata.com/ticket/563 (DISTINCT ORDER BY) - http://trac.bigdata.com/ticket/567 (Failure to set cached value on IV results in incorrect behavior for complex UPDATE operation) - http://trac.bigdata.com/ticket/568 (DELETE WHERE fails with Java AssertionError) - http://trac.bigdata.com/ticket/569 (LOAD-CREATE-LOAD using virgin journal fails with "Graph exists" exception) - http://trac.bigdata.com/ticket/571 (DELETE/INSERT WHERE handling of blank nodes) - http://trac.bigdata.com/ticket/573 (NullPointerException when attempting to INSERT DATA containing a blank node) 1.2.0: (*) - http://trac.bigdata.com/ticket/92 (Monitoring webapp) - http://trac.bigdata.com/ticket/267 (Support evaluation of 3rd party operators) - http://trac.bigdata.com/ticket/337 (Compact and efficient movement of binding sets between nodes.) - http://trac.bigdata.com/ticket/433 (Cluster leaks threads under read-only index operations: DGC thread leak) - http://trac.bigdata.com/ticket/437 (Thread-local cache combined with unbounded thread pools causes effective memory leak: termCache memory leak & thread-local buffers) - http://trac.bigdata.com/ticket/438 (KeyBeforePartitionException on cluster) - http://trac.bigdata.com/ticket/439 (Class loader problem) - http://trac.bigdata.com/ticket/441 (Ganglia integration) - http://trac.bigdata.com/ticket/443 (Logger for RWStore transaction service and recycler) - http://trac.bigdata.com/ticket/444 (SPARQL query can fail to notice when IRunningQuery.isDone() on cluster) - http://trac.bigdata.com/ticket/445 (RWStore does not track tx release correctly) - http://trac.bigdata.com/ticket/446 (HTTP Repostory broken with bigdata 1.1.0) - http://trac.bigdata.com/ticket/448 (SPARQL 1.1 UPDATE) - http://trac.bigdata.com/ticket/449 (SPARQL 1.1 Federation extension) - http://trac.bigdata.com/ticket/451 (Serialization error in SIDs mode on cluster) - http://trac.bigdata.com/ticket/454 (Global Row Store Read on Cluster uses Tx) - http://trac.bigdata.com/ticket/456 (IExtension implementations do point lookups on lexicon) - http://trac.bigdata.com/ticket/457 ("No such index" on cluster under concurrent query workload) - http://trac.bigdata.com/ticket/458 (Java level deadlock in DS) - http://trac.bigdata.com/ticket/460 (Uncaught interrupt resolving RDF terms) - http://trac.bigdata.com/ticket/461 (KeyAfterPartitionException / KeyBeforePartitionException on cluster) - http://trac.bigdata.com/ticket/463 (NoSuchVocabularyItem with LUBMVocabulary for DerivedNumericsExtension) - http://trac.bigdata.com/ticket/464 (Query statistics do not update correctly on cluster) - http://trac.bigdata.com/ticket/465 (Too many GRS reads on cluster) - http://trac.bigdata.com/ticket/469 (Sail does not flush assertion buffers before query) - http://trac.bigdata.com/ticket/472 (acceptTaskService pool size on cluster) - http://trac.bigdata.com/ticket/475 (Optimize serialization for query messages on cluster) - http://trac.bigdata.com/ticket/476 (Test suite for writeCheckpoint() and recycling for BTree/HTree) - http://trac.bigdata.com/ticket/478 (Cluster does not map input solution(s) across shards) - http://trac.bigdata.com/ticket/480 (Error releasing deferred frees using 1.0.6 against a 1.0.4 journal) - http://trac.bigdata.com/ticket/481 (PhysicalAddressResolutionException against 1.0.6) - http://trac.bigdata.com/ticket/482 (RWStore reset() should be thread-safe for concurrent readers) - http://trac.bigdata.com/ticket/484 (Java API for NanoSparqlServer REST API) - http://trac.bigdata.com/ticket/491 (AbstractTripleStore.destroy() does not clear the locator cache) - http://trac.bigdata.com/ticket/492 (Empty chunk in ThickChunkMessage (cluster)) - http://trac.bigdata.com/ticket/493 (Virtual Graphs) - http://trac.bigdata.com/ticket/496 (Sesame 2.6.3) - http://trac.bigdata.com/ticket/497 (Implement STRBEFORE, STRAFTER, and REPLACE) - http://trac.bigdata.com/ticket/498 (Bring bigdata RDF/XML parser up to openrdf 2.6.3.) - http://trac.bigdata.com/ticket/500 (SPARQL 1.1 Service Description) - http://www.openrdf.org/issues/browse/SES-884 (Aggregation with an solution set as input should produce an empty solution as output) - http://www.openrdf.org/issues/browse/SES-862 (Incorrect error handling for SPARQL aggregation; fix in 2.6.1) - http://www.openrdf.org/issues/browse/SES-873 (Order the same Blank Nodes together in ORDER BY) - http://trac.bigdata.com/ticket/501 (SPARQL 1.1 BINDINGS are ignored) - http://trac.bigdata.com/ticket/503 (Bigdata2Sesame2BindingSetIterator throws QueryEvaluationException were it should throw NoSuchElementException) - http://trac.bigdata.com/ticket/504 (UNION with Empty Group Pattern) - http://trac.bigdata.com/ticket/505 (Exception when using SPARQL sort & statement identifiers) - http://trac.bigdata.com/ticket/506 (Load, closure and query performance in 1.1.x versus 1.0.x) - http://trac.bigdata.com/ticket/508 (LIMIT causes hash join utility to log errors) - http://trac.bigdata.com/ticket/513 (Expose the LexiconConfiguration to Function BOPs) - http://trac.bigdata.com/ticket/515 (Query with two "FILTER NOT EXISTS" expressions returns no results) - http://trac.bigdata.com/ticket/516 (REGEXBOp should cache the Pattern when it is a constant) - http://trac.bigdata.com/ticket/517 (Java 7 Compiler Compatibility) - http://trac.bigdata.com/ticket/518 (Review function bop subclass hierarchy, optimize datatype bop, etc.) - http://trac.bigdata.com/ticket/520 (CONSTRUCT WHERE shortcut) - http://trac.bigdata.com/ticket/521 (Incremental materialization of Tuple and Graph query results) - http://trac.bigdata.com/ticket/525 (Modify the IChangeLog interface to support multiple agents) - http://trac.bigdata.com/ticket/527 (Expose timestamp of LexiconRelation to function bops) - http://trac.bigdata.com/ticket/532 (ClassCastException during hash join (can not be cast to TermId)) - http://trac.bigdata.com/ticket/533 (Review materialization for inline IVs) - http://trac.bigdata.com/ticket/534 (BSBM BI Q5 error using MERGE JOIN) 1.1.0 (*) - http://trac.bigdata.com/ticket/23 (Lexicon joins) - http://trac.bigdata.com/ticket/109 (Store large literals as "blobs") - http://trac.bigdata.com/ticket/181 (Scale-out LUBM "how to" in wiki and build.xml are out of date.) - http://trac.bigdata.com/ticket/203 (Implement an persistence capable hash table to support analytic query) - http://trac.bigdata.com/ticket/209 (AccessPath should visit binding sets rather than elements for high level query.) - http://trac.bigdata.com/ticket/227 (SliceOp appears to be necessary when operator plan should suffice without) - http://trac.bigdata.com/ticket/232 (Bottom-up evaluation semantics). - http://trac.bigdata.com/ticket/246 (Derived xsd numeric data types must be inlined as extension types.) - http://trac.bigdata.com/ticket/254 (Revisit pruning of intermediate variable bindings during query execution) - http://trac.bigdata.com/ticket/261 (Lift conditions out of subqueries.) - http://trac.bigdata.com/ticket/300 (Native ORDER BY) - http://trac.bigdata.com/ticket/324 (Inline predeclared URIs and namespaces in 2-3 bytes) - http://trac.bigdata.com/ticket/330 (NanoSparqlServer does not locate "html" resources when run from jar) - http://trac.bigdata.com/ticket/334 (Support inlining of unicode data in the statement indices.) - http://trac.bigdata.com/ticket/364 (Scalable default graph evaluation) - http://trac.bigdata.com/ticket/368 (Prune variable bindings during query evaluation) - http://trac.bigdata.com/ticket/370 (Direct translation of openrdf AST to bigdata AST) - http://trac.bigdata.com/ticket/373 (Fix StrBOp and other IValueExpressions) - http://trac.bigdata.com/ticket/377 (Optimize OPTIONALs with multiple statement patterns.) - http://trac.bigdata.com/ticket/380 (Native SPARQL evaluation on cluster) - http://trac.bigdata.com/ticket/387 (Cluster does not compute closure) - http://trac.bigdata.com/ticket/395 (HTree hash join performance) - http://trac.bigdata.com/ticket/401 (inline xsd:unsigned datatypes) - http://trac.bigdata.com/ticket/408 (xsd:string cast fails for non-numeric data) - http://trac.bigdata.com/ticket/421 (New query hints model.) - http://trac.bigdata.com/ticket/431 (Use of read-only tx per query defeats cache on cluster) 1.0.3 - http://trac.bigdata.com/ticket/217 (BTreeCounters does not track bytes released) - http://trac.bigdata.com/ticket/269 (Refactor performance counters using accessor interface) - http://trac.bigdata.com/ticket/329 (B+Tree should delete bloom filter when it is disabled.) - http://trac.bigdata.com/ticket/372 (RWStore does not prune the CommitRecordIndex) - http://trac.bigdata.com/ticket/375 (Persistent memory leaks (RWStore/DISK)) - http://trac.bigdata.com/ticket/385 (FastRDFValueCoder2: ArrayIndexOutOfBoundsException) - http://trac.bigdata.com/ticket/391 (Release age advanced on WORM mode journal) - http://trac.bigdata.com/ticket/392 (Add a DELETE by access path method to the NanoSparqlServer) - http://trac.bigdata.com/ticket/393 (Add "context-uri" request parameter to specify the default context for INSERT in the REST API) - http://trac.bigdata.com/ticket/394 (log4j configuration error message in WAR deployment) - http://trac.bigdata.com/ticket/399 (Add a fast range count method to the REST API) - http://trac.bigdata.com/ticket/422 (Support temp triple store wrapped by a BigdataSail) - http://trac.bigdata.com/ticket/424 (NQuads support for NanoSparqlServer) - http://trac.bigdata.com/ticket/425 (Bug fix to DEFAULT_RDF_FORMAT for bulk data loader in scale-out) - http://trac.bigdata.com/ticket/426 (Support either lockfile (procmail) and dotlockfile (liblockfile1) in scale-out) - http://trac.bigdata.com/ticket/427 (BigdataSail#getReadOnlyConnection() race condition with concurrent commit) - http://trac.bigdata.com/ticket/435 (Address is 0L) - http://trac.bigdata.com/ticket/436 (TestMROWTransactions failure in CI) 1.0.2 - http://trac.bigdata.com/ticket/32 (Query time expansion of (foo rdf:type rdfs:Resource) drags in SPORelation for scale-out.) - http://trac.bigdata.com/ticket/181 (Scale-out LUBM "how to" in wiki and build.xml are out of date.) - http://trac.bigdata.com/ticket/356 (Query not terminated by error.) - http://trac.bigdata.com/ticket/359 (NamedGraph pattern fails to bind graph variable if only one binding exists.) - http://trac.bigdata.com/ticket/361 (IRunningQuery not closed promptly.) - http://trac.bigdata.com/ticket/371 (DataLoader fails to load resources available from the classpath.) - http://trac.bigdata.com/ticket/376 (Support for the streaming of bigdata IBindingSets into a sparql query.) - http://trac.bigdata.com/ticket/378 (ClosedByInterruptException during heavy query mix.) - http://trac.bigdata.com/ticket/379 (NotSerializableException for SPOAccessPath.) - http://trac.bigdata.com/ticket/382 (Change dependencies to Apache River 2.2.0) 1.0.1 (*) - http://trac.bigdata.com/ticket/107 (Unicode clean schema names in the sparse row store). - http://trac.bigdata.com/ticket/124 (TermIdEncoder should use more bits for scale-out). - http://trac.bigdata.com/ticket/225 (OSX requires specialized performance counter collection classes). - http://trac.bigdata.com/ticket/348 (BigdataValueFactory.asValue() must return new instance when DummyIV is used). - http://trac.bigdata.com/ticket/349 (TermIdEncoder limits Journal to 2B distinct RDF Values per triple/quad store instance). - http://trac.bigdata.com/ticket/351 (SPO not Serializable exception in SIDS mode (scale-out)). - http://trac.bigdata.com/ticket/352 (ClassCastException when querying with binding-values that are not known to the database). - http://trac.bigdata.com/ticket/353 (UnsupportedOperatorException for some SPARQL queries). - http://trac.bigdata.com/ticket/355 (Query failure when comparing with non materialized value). - http://trac.bigdata.com/ticket/357 (RWStore reports "FixedAllocator returning null address, with freeBits".) - http://trac.bigdata.com/ticket/359 (NamedGraph pattern fails to bind graph variable if only one binding exists.) - http://trac.bigdata.com/ticket/362 (log4j - slf4j bridge.) For more information about bigdata(R), please see the following links: [1] http://wiki.bigdata.com/wiki/index.php/Main_Page [2] http://wiki.bigdata.com/wiki/index.php/GettingStarted [3] http://wiki.bigdata.com/wiki/index.php/Roadmap [4] http://www.bigdata.com/bigdata/docs/api/ [5] http://sourceforge.net/projects/bigdata/ [6] http://www.bigdata.com/blog [7] http://www.systap.com/bigdata.htm [8] http://sourceforge.net/projects/bigdata/files/bigdata/ [9] http://wiki.bigdata.com/wiki/index.php/DataMigration [10] http://wiki.bigdata.com/wiki/index.php/HAJournalServer [11] http://www.bigdata.com/whitepapers/reifSPARQL.pdf [12] http://wiki.bigdata.com/wiki/index.php/RDF_GAS_API About bigdata: Bigdata(R) is a horizontally-scaled, general purpose storage and computing fabric for ordered data (B+Trees), designed to operate on either a single server or a cluster of commodity hardware. Bigdata(R) uses dynamically partitioned key-range shards in order to remove any realistic scaling limits - in principle, bigdata(R) may be deployed on 10s, 100s, or even thousands of machines and new capacity may be added incrementally without requiring the full reload of all data. The bigdata(R) RDF database supports RDFS and OWL Lite reasoning, high-level query (SPARQL), and datum level provenance. |
From: Bryan T. <br...@sy...> - 2014-11-06 15:10:44
|
Jim, 502 is about support for expressions (other than simple variables in ORDER_BY). If there is an issue with DISTINCT + ORDER_BY then this would be a new ticket. Just post the EXPLAIN (attach to the email) for the moment. I want to see how this is being generated. We should then check the specification and make sure that the correct behavior is DISTINCT followed by ORDER BY with any limit applied after the ORDER BY. I can then check the code for how we are handling this. The relevant logic is in AST2BOpUtility at line 451. You can see that it is already attempting to handle this and that there was a historical ticket for this issue (#563). /* * Note: The DISTINCT operators also enforce the projection. * * Note: REDUCED allows, but does not require, either complete or * partial filtering of duplicates. It is part of what openrdf does * for a DESCRIBE query. * * Note: We do not currently have special operator for REDUCED. One * could be created using chunk wise DISTINCT. Note that REDUCED may * not change the order in which the solutions appear (but we are * evaluating it before ORDER BY so that is Ok.) * * TODO If there is an ORDER BY and a DISTINCT then the sort can be * used to impose the distinct without the overhead of a hash index * by filtering out the duplicate solutions after the sort. */ // When true, DISTINCT must preserve ORDER BY ordering. final boolean preserveOrder; if (orderBy != null && !orderBy.isEmpty()) { * /** * * Note: ORDER BY before DISTINCT, so DISTINCT must preserve* * * order.* * * * * * @see https://sourceforge.net/apps/trac/bigdata/ticket/563 <https://sourceforge.net/apps/trac/bigdata/ticket/563>* * * (ORDER BY + DISTINCT)* * */* preserveOrder = true; left = addOrderBy(left, queryBase, orderBy, ctx); } else { preserveOrder = false; } if (projection.isDistinct() || projection.isReduced()) { left = addDistinct(left, queryBase, preserveOrder, ctx); } } else { /* * TODO Under what circumstances can the projection be [null]? */ if (orderBy != null && !orderBy.isEmpty()) { left = addOrderBy(left, queryBase, orderBy, ctx); } } Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Thu, Nov 6, 2014 at 10:03 AM, Jim Balhoff <ba...@ne...> wrote: > Hi Bryan, > > Just to clarify, would you like me to attach the info to ticket 502, or > continue posting to the developer list? > > Thanks, > Jim > > > > On Nov 6, 2014, at 8:28 AM, Bryan Thompson <br...@sy...> wrote: > > > > The ticket for allowing aggregates in ORDER BY is: > > > > - http://trac.bigdata.com/ticket/502 (Allow aggregates in ORDER BY > clause) > > > > Can you attach the EXPLAIN of the query with and without DISTINCT. The > issue may be that the DISTINCT is being applied after the ORDER BY. I seem > to remember some issue historically with operations being performed > before/after the ORDER BY, but I do not have any distinct recollection of a > problematic interaction between DISTINCT and ORDER BY. > > > > Bryan > > > > ---- > > Bryan Thompson > > Chief Scientist & Founder > > SYSTAP, LLC > > 4501 Tower Road > > Greensboro, NC 27410 > > br...@sy... > > http://bigdata.com > > http://mapgraph.io > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > > > > On Wed, Nov 5, 2014 at 6:14 PM, Jim Balhoff <ba...@ne...> wrote: > > > On Nov 5, 2014, at 5:46 PM, Jeremy J Carroll <jj...@sy...> wrote: > > > > > > > > >> On Nov 5, 2014, at 1:02 PM, Bryan Thompson <br...@sy...> wrote: > > >> > > >> There could be an issue with ORDER BY operating on an anonymous and > non-projected variable. Try declaring and binding a variable for > STR(?label) inside of the query and then using that variable in the ORDER > BY clause. > > > > > > > > > Yes I tend to find the results of ORDER BY are more what I expect if I > do not include an expression in the ORDER BY but simply variables. I BIND > any expression before the ORDER BY. > > > > > > I believe there is a trac item for this, but since the workaround is > easy, I have never seen it as high priority > > > > > > > As suggested I tried binding a variable as `BIND (STR(?term_label) AS > ?string_label)` and using that to sort. Still incorrect ordering. But, I > tried removing DISTINCT, and then the ordering is correct. Even going back > to the anonymous `ORDER BY STR(?term_label)`, ordering is still correct if > I remove DISTINCT. For this specific query DISTINCT is not needed, but I do > need it for my application. Is there a reason to not expect DISTINCT to > work correctly with ORDER BY? > > > > Thanks both of you for all of your help, > > Jim > > > > > > |
From: Bryan T. <br...@sy...> - 2014-11-06 13:28:14
|
The ticket for allowing aggregates in ORDER BY is: - http://trac.bigdata.com/ticket/502 (Allow aggregates in ORDER BY clause) Can you attach the EXPLAIN of the query with and without DISTINCT. The issue may be that the DISTINCT is being applied after the ORDER BY. I seem to remember some issue historically with operations being performed before/after the ORDER BY, but I do not have any distinct recollection of a problematic interaction between DISTINCT and ORDER BY. Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Wed, Nov 5, 2014 at 6:14 PM, Jim Balhoff <ba...@ne...> wrote: > > On Nov 5, 2014, at 5:46 PM, Jeremy J Carroll <jj...@sy...> wrote: > > > > > >> On Nov 5, 2014, at 1:02 PM, Bryan Thompson <br...@sy...> wrote: > >> > >> There could be an issue with ORDER BY operating on an anonymous and > non-projected variable. Try declaring and binding a variable for > STR(?label) inside of the query and then using that variable in the ORDER > BY clause. > > > > > > Yes I tend to find the results of ORDER BY are more what I expect if I > do not include an expression in the ORDER BY but simply variables. I BIND > any expression before the ORDER BY. > > > > I believe there is a trac item for this, but since the workaround is > easy, I have never seen it as high priority > > > > As suggested I tried binding a variable as `BIND (STR(?term_label) AS > ?string_label)` and using that to sort. Still incorrect ordering. But, I > tried removing DISTINCT, and then the ordering is correct. Even going back > to the anonymous `ORDER BY STR(?term_label)`, ordering is still correct if > I remove DISTINCT. For this specific query DISTINCT is not needed, but I do > need it for my application. Is there a reason to not expect DISTINCT to > work correctly with ORDER BY? > > Thanks both of you for all of your help, > Jim > > |
From: Jeremy J C. <jj...@sy...> - 2014-11-05 23:14:51
|
> On Nov 5, 2014, at 1:02 PM, Bryan Thompson <br...@sy...> wrote: > > There could be an issue with ORDER BY operating on an anonymous and non-projected variable. Try declaring and binding a variable for STR(?label) inside of the query and then using that variable in the ORDER BY clause. Yes I tend to find the results of ORDER BY are more what I expect if I do not include an expression in the ORDER BY but simply variables. I BIND any expression before the ORDER BY. I believe there is a trac item for this, but since the workaround is easy, I have never seen it as high priority Jeremy |
From: Jim B. <ba...@ne...> - 2014-11-05 23:14:42
|
> On Nov 5, 2014, at 5:46 PM, Jeremy J Carroll <jj...@sy...> wrote: > > >> On Nov 5, 2014, at 1:02 PM, Bryan Thompson <br...@sy...> wrote: >> >> There could be an issue with ORDER BY operating on an anonymous and non-projected variable. Try declaring and binding a variable for STR(?label) inside of the query and then using that variable in the ORDER BY clause. > > > Yes I tend to find the results of ORDER BY are more what I expect if I do not include an expression in the ORDER BY but simply variables. I BIND any expression before the ORDER BY. > > I believe there is a trac item for this, but since the workaround is easy, I have never seen it as high priority > As suggested I tried binding a variable as `BIND (STR(?term_label) AS ?string_label)` and using that to sort. Still incorrect ordering. But, I tried removing DISTINCT, and then the ordering is correct. Even going back to the anonymous `ORDER BY STR(?term_label)`, ordering is still correct if I remove DISTINCT. For this specific query DISTINCT is not needed, but I do need it for my application. Is there a reason to not expect DISTINCT to work correctly with ORDER BY? Thanks both of you for all of your help, Jim |
From: Bryan T. <br...@sy...> - 2014-11-05 22:53:04
|
I think that this is an issue that appeared between sparql 1.1 last call working draft and the sparql 1.1 recommendation. Last moment change to the spec. Bryan On Nov 5, 2014 5:46 PM, "Jeremy J Carroll" <jj...@sy...> wrote: > > > On Nov 5, 2014, at 1:02 PM, Bryan Thompson <br...@sy...> wrote: > > There could be an issue with ORDER BY operating on an anonymous and > non-projected variable. Try declaring and binding a variable for > STR(?label) inside of the query and then using that variable in the ORDER > BY clause. > > > > Yes I tend to find the results of ORDER BY are more what I expect if I do > not include an expression in the ORDER BY but simply variables. I BIND any > expression before the ORDER BY. > > I believe there is a trac item for this, but since the workaround is easy, > I have never seen it as high priority > > Jeremy > > |
From: Bryan T. <br...@sy...> - 2014-11-05 21:53:58
|
All, the 1.3.4 release [1] is now in CI. This is primarily a critical bug fix release. There is a problem in 1.3.3 with SPARQL UPDATE through the REST API preventing storage recycling in the RWStore. This issue has been resolved and is the basis for the 1.3.4 release. We will follow up quickly on this with another release including openrdf 2.7 support and several performance optimizations. The key points for this release are also described at [1]. Thanks, Bryan [1] http://trac.bigdata.com/ticket/1032 |
From: Bryan T. <br...@sy...> - 2014-11-05 21:02:39
|
Does the SPARQL result set show uniform type of RDF Literal? If not, then it is not a sufficient mechanism (de-facto). Again, the correct behavior of STR() in that role would be determined by the SPARQL specification. Here is a snip of the code for STR() which is in StrBOp.java. You can see that a simple literal is just returned and otherwise a new literal is created from just the label. There could be an issue with ORDER BY operating on an anonymous and non-projected variable. Try declaring and binding a variable for STR(?label) inside of the query and then using that variable in the ORDER BY clause. // use to create my simple literals final BigdataValueFactory vf = getValueFactory(); if (val instanceof Literal) { final Literal lit = (Literal) val; if (lit.getDatatype() == null && lit.getLanguage() == null) { // if simple literal return it return iv; } else { // else return new simple literal using Literal.getLabel final BigdataLiteral str = vf.createLiteral(lit.getLabel()); return super.asIV(str, bs); } } else if (val instanceof URI) { // return new simple literal using URI label final BigdataLiteral str = vf.createLiteral(val.stringValue()); return super.asIV(str, bs); } else { throw new SparqlTypeErrorException(); } Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Wed, Nov 5, 2014 at 3:43 PM, Jim Balhoff <ba...@ne...> wrote: > I thought that different types of strings might affect the ordering, so I > also tried this as the last line of the query: > > ORDER BY STR(?term_label) > > This also results in similar incorrect ordering. Would you expect this to > be enough to remove any problems due to different literal types? Based on > the standard, my expectation is that this would. > > Thank you, > Jim > > > > On Nov 5, 2014, at 2:14 PM, Bryan Thompson <br...@sy...> wrote: > > > > Jim, > > > > If you look at the SPARQL output, the labels appear to be present twice > because some of them are: > > > > <literal xml:lang='en'>anterior humeral > ridge</literal> > > > > and some are: > > > > <literal datatype=' > http://www.w3.org/2001/XMLSchema#string'>1st arch mandibular > component</literal> > > > > So they are not the same "type" of literal. > > > > You can probably cast everything to a single type to get around this. > > > > Please check with the standard, but I am not sure that there is a bug > here. > > > > Thanks, > > Bryan > > > > > > ---- > > Bryan Thompson > > Chief Scientist & Founder > > SYSTAP, LLC > > 4501 Tower Road > > Greensboro, NC 27410 > > br...@sy... > > http://bigdata.com > > http://mapgraph.io > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > > > > On Wed, Nov 5, 2014 at 10:36 AM, Jim Balhoff <ba...@ne...> > wrote: > > > On Nov 5, 2014, at 10:20 AM, Bryan Thompson <br...@sy...> wrote: > > > > > > Is there a public endpoint and query that I can use to test this? > > > > I will send you a separate email with this. > > > > > > > > If this is local data, is there a small data set that we can use to > replicate the problem? > > > > I am using the same dataset in a local instance as in the original > ticket: http://purl.obolibrary.org/obo/uberon/releases/2014-10-26/ext.owl > > > > Just the triples in that file. Query: > > > > PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> > > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > > > SELECT DISTINCT ?term ?term_label > > WHERE > > { > > ?term rdf:type owl:Class . > > ?term rdfs:label ?term_label . > > } > > ORDER BY ?term_label > > > > > > > > > In general, the ORDER BY operator should execute once ALL solutions > have been materialized within that operator. It then applies the sort and > the solutions are reported. > > > > > > My questions would be: > > > > > > - What is the EXPLAIN of the query? > > > > I attached a copy of the EXPLAIN output, in HTML format to preserve the > table. To me it looks like the sort is not happening at the end, but > instead earlier, but I don't have much confidence in my understanding of > everything being reported. > > > > > - Does a simple unit test of the MemorySortOp show the same problem? > That is, is this related to the MemorySortOp implementation or the query > engine / query plan generator? > > > > I've only tested SPARQL queries so far. > > > > > - Are there any odd things going on with the unicode setup? Are the > characters "a" and "a" really the same characters. > > > > Not that I know of. I can create a new ticket for this if you would like. > > > > Thanks, > > Jim > > > > > > > ------------------------------------------------------------------------------ > > _______________________________________________ > > Bigdata-developers mailing list > > Big...@li... > > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > > |
From: Jim B. <ba...@ne...> - 2014-11-05 20:43:38
|
I thought that different types of strings might affect the ordering, so I also tried this as the last line of the query: ORDER BY STR(?term_label) This also results in similar incorrect ordering. Would you expect this to be enough to remove any problems due to different literal types? Based on the standard, my expectation is that this would. Thank you, Jim > On Nov 5, 2014, at 2:14 PM, Bryan Thompson <br...@sy...> wrote: > > Jim, > > If you look at the SPARQL output, the labels appear to be present twice because some of them are: > > <literal xml:lang='en'>anterior humeral ridge</literal> > > and some are: > > <literal datatype='http://www.w3.org/2001/XMLSchema#string'>1st arch mandibular component</literal> > > So they are not the same "type" of literal. > > You can probably cast everything to a single type to get around this. > > Please check with the standard, but I am not sure that there is a bug here. > > Thanks, > Bryan > > > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... > http://bigdata.com > http://mapgraph.io > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > On Wed, Nov 5, 2014 at 10:36 AM, Jim Balhoff <ba...@ne...> wrote: > > On Nov 5, 2014, at 10:20 AM, Bryan Thompson <br...@sy...> wrote: > > > > Is there a public endpoint and query that I can use to test this? > > I will send you a separate email with this. > > > > > If this is local data, is there a small data set that we can use to replicate the problem? > > I am using the same dataset in a local instance as in the original ticket: http://purl.obolibrary.org/obo/uberon/releases/2014-10-26/ext.owl > > Just the triples in that file. Query: > > PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > SELECT DISTINCT ?term ?term_label > WHERE > { > ?term rdf:type owl:Class . > ?term rdfs:label ?term_label . > } > ORDER BY ?term_label > > > > > In general, the ORDER BY operator should execute once ALL solutions have been materialized within that operator. It then applies the sort and the solutions are reported. > > > > My questions would be: > > > > - What is the EXPLAIN of the query? > > I attached a copy of the EXPLAIN output, in HTML format to preserve the table. To me it looks like the sort is not happening at the end, but instead earlier, but I don't have much confidence in my understanding of everything being reported. > > > - Does a simple unit test of the MemorySortOp show the same problem? That is, is this related to the MemorySortOp implementation or the query engine / query plan generator? > > I've only tested SPARQL queries so far. > > > - Are there any odd things going on with the unicode setup? Are the characters "a" and "a" really the same characters. > > Not that I know of. I can create a new ticket for this if you would like. > > Thanks, > Jim > > > ------------------------------------------------------------------------------ > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers |
From: Jim B. <ba...@ne...> - 2014-11-05 20:43:35
|
I thought that different types of strings might affect the ordering, so I also tried this as the last line of the query: ORDER BY STR(?term_label) This also results in similar incorrect ordering. Would you expect this to be enough to remove any problems due to different literal types? Based on the standard, my expectation is that this would. Thank you, Jim > On Nov 5, 2014, at 2:14 PM, Bryan Thompson <br...@sy...> wrote: > > Jim, > > If you look at the SPARQL output, the labels appear to be present twice because some of them are: > > <literal xml:lang='en'>anterior humeral ridge</literal> > > and some are: > > <literal datatype='http://www.w3.org/2001/XMLSchema#string'>1st arch mandibular component</literal> > > So they are not the same "type" of literal. > > You can probably cast everything to a single type to get around this. > > Please check with the standard, but I am not sure that there is a bug here. > > Thanks, > Bryan > > > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... > http://bigdata.com > http://mapgraph.io > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > On Wed, Nov 5, 2014 at 10:36 AM, Jim Balhoff <ba...@ne...> wrote: > > On Nov 5, 2014, at 10:20 AM, Bryan Thompson <br...@sy...> wrote: > > > > Is there a public endpoint and query that I can use to test this? > > I will send you a separate email with this. > > > > > If this is local data, is there a small data set that we can use to replicate the problem? > > I am using the same dataset in a local instance as in the original ticket: http://purl.obolibrary.org/obo/uberon/releases/2014-10-26/ext.owl > > Just the triples in that file. Query: > > PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > SELECT DISTINCT ?term ?term_label > WHERE > { > ?term rdf:type owl:Class . > ?term rdfs:label ?term_label . > } > ORDER BY ?term_label > > > > > In general, the ORDER BY operator should execute once ALL solutions have been materialized within that operator. It then applies the sort and the solutions are reported. > > > > My questions would be: > > > > - What is the EXPLAIN of the query? > > I attached a copy of the EXPLAIN output, in HTML format to preserve the table. To me it looks like the sort is not happening at the end, but instead earlier, but I don't have much confidence in my understanding of everything being reported. > > > - Does a simple unit test of the MemorySortOp show the same problem? That is, is this related to the MemorySortOp implementation or the query engine / query plan generator? > > I've only tested SPARQL queries so far. > > > - Are there any odd things going on with the unicode setup? Are the characters "a" and "a" really the same characters. > > Not that I know of. I can create a new ticket for this if you would like. > > Thanks, > Jim > > > ------------------------------------------------------------------------------ > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers |
From: Bryan T. <br...@sy...> - 2014-11-05 19:14:32
|
Jim, If you look at the SPARQL output, the labels appear to be present twice because some of them are: <literal xml:lang='en'>anterior humeral ridge</literal> and some are: <literal datatype='http://www.w3.org/2001/XMLSchema#string'>1st arch mandibular component</literal> So they are not the same "type" of literal. You can probably cast everything to a single type to get around this. Please check with the standard, but I am not sure that there is a bug here. Thanks, Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Wed, Nov 5, 2014 at 10:36 AM, Jim Balhoff <ba...@ne...> wrote: > > On Nov 5, 2014, at 10:20 AM, Bryan Thompson <br...@sy...> wrote: > > > > Is there a public endpoint and query that I can use to test this? > > I will send you a separate email with this. > > > > > If this is local data, is there a small data set that we can use to > replicate the problem? > > I am using the same dataset in a local instance as in the original ticket: > http://purl.obolibrary.org/obo/uberon/releases/2014-10-26/ext.owl > > Just the triples in that file. Query: > > PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > SELECT DISTINCT ?term ?term_label > WHERE > { > ?term rdf:type owl:Class . > ?term rdfs:label ?term_label . > } > ORDER BY ?term_label > > > > > In general, the ORDER BY operator should execute once ALL solutions have > been materialized within that operator. It then applies the sort and the > solutions are reported. > > > > My questions would be: > > > > - What is the EXPLAIN of the query? > > I attached a copy of the EXPLAIN output, in HTML format to preserve the > table. To me it looks like the sort is not happening at the end, but > instead earlier, but I don't have much confidence in my understanding of > everything being reported. > > > - Does a simple unit test of the MemorySortOp show the same problem? > That is, is this related to the MemorySortOp implementation or the query > engine / query plan generator? > > I've only tested SPARQL queries so far. > > > - Are there any odd things going on with the unicode setup? Are the > characters "a" and "a" really the same characters. > > Not that I know of. I can create a new ticket for this if you would like. > > Thanks, > Jim > > |
From: Jim B. <ba...@ne...> - 2014-11-05 15:36:26
|
> On Nov 5, 2014, at 10:20 AM, Bryan Thompson <br...@sy...> wrote: > > Is there a public endpoint and query that I can use to test this? I will send you a separate email with this. > > If this is local data, is there a small data set that we can use to replicate the problem? I am using the same dataset in a local instance as in the original ticket: http://purl.obolibrary.org/obo/uberon/releases/2014-10-26/ext.owl Just the triples in that file. Query: PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> PREFIX owl: <http://www.w3.org/2002/07/owl#> SELECT DISTINCT ?term ?term_label WHERE { ?term rdf:type owl:Class . ?term rdfs:label ?term_label . } ORDER BY ?term_label > In general, the ORDER BY operator should execute once ALL solutions have been materialized within that operator. It then applies the sort and the solutions are reported. > > My questions would be: > > - What is the EXPLAIN of the query? I attached a copy of the EXPLAIN output, in HTML format to preserve the table. To me it looks like the sort is not happening at the end, but instead earlier, but I don't have much confidence in my understanding of everything being reported. > - Does a simple unit test of the MemorySortOp show the same problem? That is, is this related to the MemorySortOp implementation or the query engine / query plan generator? I've only tested SPARQL queries so far. > - Are there any odd things going on with the unicode setup? Are the characters "a" and "a" really the same characters. Not that I know of. I can create a new ticket for this if you would like. Thanks, Jim |
From: Bryan T. <br...@sy...> - 2014-11-05 15:21:03
|
Is there a public endpoint and query that I can use to test this? If this is local data, is there a small data set that we can use to replicate the problem? In general, the ORDER BY operator should execute once ALL solutions have been materialized within that operator. It then applies the sort and the solutions are reported. My questions would be: - What is the EXPLAIN of the query? - Does a simple unit test of the MemorySortOp show the same problem? That is, is this related to the MemorySortOp implementation or the query engine / query plan generator? - Are there any odd things going on with the unicode setup? Are the characters "a" and "a" really the same characters. Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Wed, Nov 5, 2014 at 10:15 AM, Jim Balhoff <ba...@ne...> wrote: > > On Nov 5, 2014, at 6:22 AM, Bryan Thompson <br...@sy...> wrote: > > > > Jim, > > > > I know that Jeremy had made a change to the order by operator a few > months ago. Are you using the current code? > > I am testing with branch BIGDATA_RELEASE_1_3_0, revision 8702, which I > think is the latest. > > > > > Can you online an example of the out of order results? > > Here is the result of the below query: > https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-5/bigdata_order_by.csv > > At the beginning you will see: > anterior humeral ridge > anteroventral process of cleithrum > columnar area > deltoid process > > Then at line 286 the 'a's start again: > YSL > Zymbal's gland > abdomen > abdomen blood vessel > abdomen connective tissue > abdomen element > abdomen musculature > > > Thanks, > Jim > > > > > > > > Bryan > > > > On Tuesday, November 4, 2014, Jim Balhoff <ba...@ne...> wrote: > > Hi Bryan, > > > > I think that the text search may have been a red herring. I tried a > query without search to compare the EXPLAIN results and realized that ORDER > BY is not working for a basic query of all labels. Has anyone been using it > successfully? My query is this: > > > > ************** > > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > > > SELECT DISTINCT ?term ?term_label > > WHERE > > { > > ?term rdf:type owl:Class . > > ?term rdfs:label ?term_label . > > } > > ORDER BY ?term_label > > ************** > > > > I also tried `ORDER BY STR(?term_label)`. The results are different but > still not fully in order. The results have large blocks of correct ordering > but contain misplaced blocks as well. I'm not really sure how to interpret > all of the EXPLAIN output. I attached the explanation in case you have time > to take a look at it. Perhaps I am missing something obvious! > > > > Thank you, > > Jim > > > > > > > > > On Nov 4, 2014, at 7:53 PM, Bryan Thompson <br...@sy...> wrote: > > > > > > Jim, > > > > > > Today, and for probably 1-2 years now, search is translated into a > SERVICE call. Triple patterns are pulled into that service call based on > the shared variable bindings. However, the recommended approach is to > Actually specify the service call directly. This will probably give you > the desired control over the ordering of the results. > > > > > > I am not sure why the ordering was not obeyed. Can you look into the > EXPLAIN of the query and see if you can identify what is going on? I may > then be able to point you towards how to resolve the ticket. > > > > > > Thanks, > > > Bryan > > > > > > > > > > > > -- > > > ---- > > > Bryan Thompson > > > Chief Scientist & Founder > > > SYSTAP, LLC > > > 4501 Tower Road > > > Greensboro, NC 27410 > > > br...@sy... > > > http://bigdata.com > > > http://mapgraph.io > > > CONFIDENTIALITY NOTICE: This email and its contents and attachments > are for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > > > > > > > > > > > -- > > ---- > > Bryan Thompson > > Chief Scientist & Founder > > SYSTAP, LLC > > 4501 Tower Road > > Greensboro, NC 27410 > > br...@sy... > > http://bigdata.com > > http://mapgraph.io > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > > > > |
From: Jim B. <ba...@ne...> - 2014-11-05 15:16:04
|
> On Nov 5, 2014, at 6:22 AM, Bryan Thompson <br...@sy...> wrote: > > Jim, > > I know that Jeremy had made a change to the order by operator a few months ago. Are you using the current code? I am testing with branch BIGDATA_RELEASE_1_3_0, revision 8702, which I think is the latest. > > Can you online an example of the out of order results? Here is the result of the below query: https://dl.dropboxusercontent.com/u/6704325/bigdata/2014-11-5/bigdata_order_by.csv At the beginning you will see: anterior humeral ridge anteroventral process of cleithrum columnar area deltoid process Then at line 286 the 'a's start again: YSL Zymbal's gland abdomen abdomen blood vessel abdomen connective tissue abdomen element abdomen musculature Thanks, Jim > > Bryan > > On Tuesday, November 4, 2014, Jim Balhoff <ba...@ne...> wrote: > Hi Bryan, > > I think that the text search may have been a red herring. I tried a query without search to compare the EXPLAIN results and realized that ORDER BY is not working for a basic query of all labels. Has anyone been using it successfully? My query is this: > > ************** > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > SELECT DISTINCT ?term ?term_label > WHERE > { > ?term rdf:type owl:Class . > ?term rdfs:label ?term_label . > } > ORDER BY ?term_label > ************** > > I also tried `ORDER BY STR(?term_label)`. The results are different but still not fully in order. The results have large blocks of correct ordering but contain misplaced blocks as well. I'm not really sure how to interpret all of the EXPLAIN output. I attached the explanation in case you have time to take a look at it. Perhaps I am missing something obvious! > > Thank you, > Jim > > > > > On Nov 4, 2014, at 7:53 PM, Bryan Thompson <br...@sy...> wrote: > > > > Jim, > > > > Today, and for probably 1-2 years now, search is translated into a SERVICE call. Triple patterns are pulled into that service call based on the shared variable bindings. However, the recommended approach is to Actually specify the service call directly. This will probably give you the desired control over the ordering of the results. > > > > I am not sure why the ordering was not obeyed. Can you look into the EXPLAIN of the query and see if you can identify what is going on? I may then be able to point you towards how to resolve the ticket. > > > > Thanks, > > Bryan > > > > > > > > -- > > ---- > > Bryan Thompson > > Chief Scientist & Founder > > SYSTAP, LLC > > 4501 Tower Road > > Greensboro, NC 27410 > > br...@sy... > > http://bigdata.com > > http://mapgraph.io > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > > > > > > -- > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... > http://bigdata.com > http://mapgraph.io > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > |
From: Bryan T. <br...@sy...> - 2014-11-05 11:22:17
|
Jim, I know that Jeremy had made a change to the order by operator a few months ago. Are you using the current code? Can you online an example of the out of order results? Bryan On Tuesday, November 4, 2014, Jim Balhoff <ba...@ne...> wrote: > Hi Bryan, > > I think that the text search may have been a red herring. I tried a query > without search to compare the EXPLAIN results and realized that ORDER BY is > not working for a basic query of all labels. Has anyone been using it > successfully? My query is this: > > ************** > PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> > PREFIX owl: <http://www.w3.org/2002/07/owl#> > > SELECT DISTINCT ?term ?term_label > WHERE > { > ?term rdf:type owl:Class . > ?term rdfs:label ?term_label . > } > ORDER BY ?term_label > ************** > > I also tried `ORDER BY STR(?term_label)`. The results are different but > still not fully in order. The results have large blocks of correct ordering > but contain misplaced blocks as well. I'm not really sure how to interpret > all of the EXPLAIN output. I attached the explanation in case you have time > to take a look at it. Perhaps I am missing something obvious! > > Thank you, > Jim > > > > > On Nov 4, 2014, at 7:53 PM, Bryan Thompson <br...@sy... > <javascript:;>> wrote: > > > > Jim, > > > > Today, and for probably 1-2 years now, search is translated into a > SERVICE call. Triple patterns are pulled into that service call based on > the shared variable bindings. However, the recommended approach is to > Actually specify the service call directly. This will probably give you > the desired control over the ordering of the results. > > > > I am not sure why the ordering was not obeyed. Can you look into the > EXPLAIN of the query and see if you can identify what is going on? I may > then be able to point you towards how to resolve the ticket. > > > > Thanks, > > Bryan > > > > > > > > -- > > ---- > > Bryan Thompson > > Chief Scientist & Founder > > SYSTAP, LLC > > 4501 Tower Road > > Greensboro, NC 27410 > > br...@sy... <javascript:;> > > http://bigdata.com > > http://mapgraph.io > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > > > -- ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |
From: Jim B. <ba...@ne...> - 2014-11-05 03:15:51
|
Hi Bryan, I think that the text search may have been a red herring. I tried a query without search to compare the EXPLAIN results and realized that ORDER BY is not working for a basic query of all labels. Has anyone been using it successfully? My query is this: ************** PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> PREFIX owl: <http://www.w3.org/2002/07/owl#> SELECT DISTINCT ?term ?term_label WHERE { ?term rdf:type owl:Class . ?term rdfs:label ?term_label . } ORDER BY ?term_label ************** I also tried `ORDER BY STR(?term_label)`. The results are different but still not fully in order. The results have large blocks of correct ordering but contain misplaced blocks as well. I'm not really sure how to interpret all of the EXPLAIN output. I attached the explanation in case you have time to take a look at it. Perhaps I am missing something obvious! Thank you, Jim > On Nov 4, 2014, at 7:53 PM, Bryan Thompson <br...@sy...> wrote: > > Jim, > > Today, and for probably 1-2 years now, search is translated into a SERVICE call. Triple patterns are pulled into that service call based on the shared variable bindings. However, the recommended approach is to Actually specify the service call directly. This will probably give you the desired control over the ordering of the results. > > I am not sure why the ordering was not obeyed. Can you look into the EXPLAIN of the query and see if you can identify what is going on? I may then be able to point you towards how to resolve the ticket. > > Thanks, > Bryan > > > > -- > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... > http://bigdata.com > http://mapgraph.io > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > |
From: Bryan T. <br...@sy...> - 2014-11-05 00:54:06
|
Jim, Today, and for probably 1-2 years now, search is translated into a SERVICE call. Triple patterns are pulled into that service call based on the shared variable bindings. However, the recommended approach is to Actually specify the service call directly. This will probably give you the desired control over the ordering of the results. I am not sure why the ordering was not obeyed. Can you look into the EXPLAIN of the query and see if you can identify what is going on? I may then be able to point you towards how to resolve the ticket. Thanks, Bryan -- ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |
From: Bryan T. <br...@sy...> - 2014-11-04 13:47:47
|
I have gone through the developers list and the code and summarized the issues involved in configuring inference for bigdata at http://wiki.bigdata.com/wiki/index.php/InferenceAndTruthMaintenance Thanks, Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |
From: Bryan T. <br...@sy...> - 2014-11-03 11:42:09
|
Those sample code examples are for the embedded version of bigdata. For HA, use the workbench and the SPARQL and SPARQL update endpoints. Bryan On Monday, November 3, 2014, Ravi Prakash Putchala < rav...@fa...> wrote: > Hi, > > I am new to Bigdata and am trying to setup and use HAJournalServer. I > hope this is the mailing list to seek help regarding the usage of > Bigdata. Else please point me in the right direction. > > I configured 3 servers and installed zookeeper and HAJournalServer by > following the "Basic Deployment" section in the wikipage HAJournalServer > (http://wiki.bigdata.com/wiki/index.php/HAJournalServer). Now I would > like to use this setup to load and query some data just like > bigdata-sails/src/samples/com/bigdata/samples/SampleCode.java does. > I just got stuck here and do not know how to connect to the cluster, > load, query etc. Could you please help by providing some pointers? > > I am using version 1.3.3. Please let me know if I need to provide more > information. > > Thank you. > > Regards, > > Ravi > > > > > ------------------------------------------------------------------------------ > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <javascript:;> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > -- ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |
From: Ravi P. P. <rav...@fa...> - 2014-11-03 07:50:37
|
Hi, I am new to Bigdata and am trying to setup and use HAJournalServer. I hope this is the mailing list to seek help regarding the usage of Bigdata. Else please point me in the right direction. I configured 3 servers and installed zookeeper and HAJournalServer by following the "Basic Deployment" section in the wikipage HAJournalServer (http://wiki.bigdata.com/wiki/index.php/HAJournalServer). Now I would like to use this setup to load and query some data just like bigdata-sails/src/samples/com/bigdata/samples/SampleCode.java does. I just got stuck here and do not know how to connect to the cluster, load, query etc. Could you please help by providing some pointers? I am using version 1.3.3. Please let me know if I need to provide more information. Thank you. Regards, Ravi |
From: Alice E. <ali...@ya...> - 2014-11-01 15:14:35
|
Ok. Thanks a lot I'll ask my manager to contact you soon then :) On Saturday, 1 November 2014 8:12 PM, Bryan Thompson <br...@sy...> wrote: Try using SPARQL. Performance depends greatly on how the platform is configured. We can help you maximize performance for your application under our accelerator program. This program goes beyond our basic developer support program and is designed to give you access to the core development team to help you develop your application and get it ready for the market, including our support for your internal performance tuning of your application. We do not provide these as free services. Proper benchmarking and performance tuning are a complex and time consuming activities. On the other hand, we can provide references to existing customers that have been deeply satisfied by their engagement with Systap around the bigdata platform. If you do your own performance testing, you should realize that misconfigured deployments can result in substantial bais in the test results. Thanks, Bryan On Saturday, November 1, 2014, Alice Everett <ali...@ya...> wrote: Sorry for a chain of mails. If SPARQL* does not make an difference to the performance than its ok I can work with SPARQL too as I have shown in my example. > > > >On Saturday, 1 November 2014 5:01 PM, Alice Everett <ali...@ya...> wrote: > > > >ant start-bigdata is giving me the following output. Thank you again for the help thus far. God Bless You. I think some of the things related to my environment are given below. Can you please share details as to in which environment will bigdata work -- I'll happily change my environment. > > >Buildfile: /home/bigdataAnt/bigdata/build.xml > > >prepare: > [echo] version=bigdata-1.3.2-20141101 > [echo] svn.checkout=true > > >buildinfo: > [echo] > [echo] package com.bigdata; > [echo] public class BuildInfo { > [echo] public static final String buildVersion="1.3.2"; > [echo] public static final String buildVersionOSGI="1.0"; > [echo] public static final String svnRevision="8685"; > [echo] public static final String svnURL="svn://svn.code.sf.net/p/bigdata/code/branches/BIGDATA_RELEASE_1_3_0"; > [echo] public static final String buildTimestamp="2014/11/01 16:12:54 IST"; > [echo] public static final String buildUser=""; > [echo] public static final String buildHost="${env.COMPUTERNAME}"; > [echo] public static final String osArch="amd64"; > [echo] public static final String osName="Linux"; > [echo] public static final String osVersion="3.8.0-44-generic"; > [echo] } > > >compile: > [echo] javac > [echo] destdir="ant-build" > [echo] fork="yes" > [echo] memorymaximumsize="1g" > [echo] debug="yes" > [echo] debuglevel="lines,vars,source" > [echo] verbose="off" > [echo] encoding="Cp1252" > [echo] source="1.7" > [echo] target="1.7" > [javac] Compiling 1 source file to /home/bigdataAnt/bigdata/ant-build/classes > [javac] javac 1.7.0_65 > > >start-bigdata: > [java] > [java] BIGDATA(R) > [java] > [java] Flexible > [java] Reliable > [java] Affordable > [java] Web-Scale Computing for the Enterprise > [java] > [java] Copyright SYSTAP, LLC 2006-2013. All rights reserved. > [java] > [java] -HP-ProBook-4430s > [java] Sat Nov 01 16:12:58 IST 2014 > [java] Linux/3.8.0-44-generic amd64 > [java] Intel(R) Core(TM) i5-3210M CPU @ 2.50GHz Family 6 Model 58 Stepping 9, GenuineIntel #CPU=4 > [java] Oracle Corporation 1.7.0_65 > [java] freeMemory=113623880 > [java] buildVersion=1.3.2 > [java] > [java] Dependency License > [java] ICU http://source.icu-project.org/repos/icu/icu/trunk/license.html > [java] bigdata-ganglia http://www.apache.org/licenses/LICENSE-2.0.html > [java] blueprints-core https://github.com/tinkerpop/blueprints/blob/master/LICENSE.txt > [java] colt http://acs.lbl.gov/software/colt/license.html > [java] commons-codec http://www.apache.org/licenses/LICENSE-2.0.html > [java] commons-fileupload http://www.apache.org/licenses/LICENSE-2.0.html > [java] commons-io http://www.apache.org/licenses/LICENSE-2.0.html > [java] commons-logging http://www.apache.org/licenses/LICENSE-2.0.html > [java] dsiutils http://www.gnu.org/licenses/lgpl-2.1.html > [java] fastutil http://www.apache.org/licenses/LICENSE-2.0.html > [java] flot http://www.opensource.org/licenses/mit-license.php > [java] high-scale-lib http://creativecommons.org/licenses/publicdomain > [java] httpclient http://www.apache.org/licenses/LICENSE-2.0.html > [java] httpclient-cache http://www.apache.org/licenses/LICENSE-2.0.html > [java] httpcore http://www.apache.org/licenses/LICENSE-2.0.html > [java] httpmime http://www.apache.org/licenses/LICENSE-2.0.html > [java] jackson-core http://www.apache.org/licenses/LICENSE-2.0.html > [java] jetty http://www.apache.org/licenses/LICENSE-2.0.html > [java] jquery https://github.com/jquery/jquery/blob/master/MIT-LICENSE.txt > [java] log4j http://www.apache.org/licenses/LICENSE-2.0.html > [java] lucene http://www.apache.org/licenses/LICENSE-2.0.html > [java] nanohttp http://elonen.iki.fi/code/nanohttpd/#license > [java] nxparser http://sw.deri.org/2006/08/nxparser/license.txt > [java] rexster-core https://github.com/tinkerpop/rexster/blob/master/LICENSE.txt > [java] river http://www.apache.org/licenses/LICENSE-2.0.html > [java] servlet-api http://www.apache.org/licenses/LICENSE-2.0.html > [java] sesame http://www.openrdf.org/download.jsp > [java] slf4j http://www.slf4j.org/license.html > [java] zookeeper http://www.apache.org/licenses/LICENSE-2.0.html > [java] > [java] INFO: com.bigdata.util.config.LogUtil: Configure and watch: bigdata-war/src/WEB-INF/classes/log4j.properties > [java] WARN : NanoSparqlServer.java:476: Starting NSS > [java] WARN : ServiceRegistry.java:47: New service class org.openrdf.rio.ntriples.NTriplesParserFactory replaces existing service class com.bigdata.rdf.rio.ntriples.BigdataNTriplesParserFactory > [java] WARN : ServiceRegistry.java:47: New service class org.openrdf.rio.turtle.TurtleParserFactory replaces existing service class com.bigdata.rdf.rio.turtle.BigdataTurtleParserFactory > [java] WARN : ServiceRegistry.java:47: New service class org.openrdf.query.resultio.sparqljson.SPARQLResultsJSONWriterFactory replaces existing service class com.bigdata.rdf.rio.json.BigdataSPARQLResultsJSONWriterFactoryForSelect > [java] WARN : ServiceRegistry.java:47: New service class org.openrdf.rio.turtle.TurtleWriterFactory replaces existing service class com.bigdata.rdf.rio.turtle.BigdataTurtleWriterFactory > [java] serviceURL: http://192.168.145.1:9999 > > > >On Saturday, 1 November 2014 4:56 PM, Alice Everett <ali...@ya...> wrote: > > > >Ok I am using Ubuntu 12.04. So do you suggest I should try in Windows environment. I trying to first make the product workable on my lenovo T430 laptop. > > >Actually we are testing a number of products like Virtuoso, BigData, Jena, etc. And my manager says the company will buy the product which performs best for open source software, as our company would not like to buy support for all the products. Therefore I am asking for a little help with the initial set-up (as otherwise we will not be able to test it). I'll be very grateful if you can help me with this a bit. > > >I'll spoke to my manager to see if we can have a call with you -- but my manager says first I should show some performance on open source, then definitely we'll buy the product support. Hope you understand my stance and help me with this a bit. > > > >On Saturday, 1 November 2014 4:47 PM, Bryan Thompson <br...@sy...> wrote: > > > >Alice, > > >The issue may be the openrdf parsers are not being correctly overridden by the bigdata RDR parsers in your deployment environment. If you want to attempt and diagnose this yourself, it might be a class path ordering issue or a jar metadata odering issue. We have noticed this is some environments, but have not yet reduced it to a root cause. > > >Thanks, >Bryan > >On Saturday, November 1, 2014, Bryan Thompson <br...@sy...> wrote: > >Alice, >> >> >>We do have paid developer support subscriptions for small projects at $500/month. Paid developer support allows us to prioritize your requests. We also have production support subscriptions that provide direct access to the core bigdata team for support of production deployments. >> >> >>The open source support channel is provided as a kindness to the community. It is not an appropriate forum if you have an internal project deadline. Instead, I suggest that you start a developer support subscription. If necessary, we can even do this as a paypal transaction. This will allow us to prioritize your issues along with those of other paying customers. >> >> >>For the moment, it sounds like you have a workaround for this specific issue since you can query the data using the reified triple patterns. >> >> >>If you would like to move forward, I suggest that we also schedule a meeting for next week so we can understand a little more about your use case and applications and help you understand more about the features and offerings for bigdata. >> >> >>Thanks, >>Bryan >> >>On Friday, October 31, 2014, Alice Everett <ali...@ya...> wrote: >> >>That's ok. I need to give a presentation on Monday. So probably you can help on Sunday. >>> >>> >>>Actually, I dont have an issue with the framework I am just not getting how to use it to insert data using RDR mode using CURL. Perhaps, a little example from you can help me with this big time. >>> >>> >>> >>>On Saturday, 1 November 2014 1:35 AM, Bryan Thompson <br...@sy...> wrote: >>> >>> >>> >>>Alice, >>> >>> >>>I am in meetings with a customer today. I could look at this next week. >>> >>> >>>FYI, from the project forum page. If we can not easily recreate the issue then it will not receive any priority under open source support. It is up to you to make the issue as easy to recreate as possible. You can file a ticket and (preferably) create a unit test for the problem. >>> >>> >>>You may use this forum to request help. If you have a bug or a feature request, please log an issue on the tracker [1] and include a unit test which demonstrates the bug. Please follow the instructions [2] when submitting a bug report. >>> >>>If your are interested in services for custom feature development, integration, architecture, or support, please contract the project leads directly. >>>[1] http://trac.bigdata.com/ >>>[2] http://wiki.bigdata.com/wiki/index.php/Submitting_Bugs >>> >>> >>>Thanks, >>>Bryan >>> >>> >>>---- >>>Bryan Thompson >>> >>>Chief Scientist & Founder >>>SYSTAP, LLC >>> >>>4501 Tower Road >>>Greensboro, NC 27410 >>> >>>br...@sy... >>> >>>http://bigdata.com >>> >>>http://mapgraph.io >>> >>>CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. >>> >>>On Fri, Oct 31, 2014 at 3:40 PM, Alice Everett <ali...@ya...> wrote: >>> >>>Dear Bryan, >>>> >>>> >>>>I'll be thankful if you can help me with this a bit. Actually I need to give a small presentation in my company regarding how can frameworks like Bigdata help us. It will be great if I can accompany the presentation with a small demo. >>>> >>>> >>>> >>>> >>>>Cheers, >>>>Alice >>>> >>>> >>>> >>>>On Friday, 31 October 2014 7:55 PM, Alice Everett <ali...@ya...> wrote: >>>> >>>> >>>> >>>>Thanks for the reply Rose but I already tried it..although the loading works perfectly fine yet the database does not contain any data: >>>> >>>> >>>>root:~/bigdataAnt$ curl -X POST http://192.168.145.1:9999/bigdata/namespace/reificationRDR/sparql --data-urlencode 'query=SELECT * {<<?s ?p ?o>> ?p1 ?o1 }' -H 'Accept:application/rdf+xml' >>>><?xml version='1.0' encoding='UTF-8'?> >>>><sparql xmlns='http://www.w3.org/2005/sparql-results#'> >>>><head> >>>><variable name='s'/> >>>><variable name='p'/> >>>><variable name='o'/> >>>><variable name='-sid-1'/> >>>><variable name='p1'/> >>>><variable name='o1'/> >>>></head> >>>><results> >>>></results> >>>></sparql> >>>> >>>> >>>> >>>> >>>>I loaded the following file using in reificationRDR namespace: >>>>@prefix rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> . >>>>@prefix dc: <http://purl.org/dc/elements/1.1/> . >>>>@prefix : <http://example/ns#> . >>>> >>>> >>>>_:c rdf:subject <http://example.org/book/book11> . >>>>_:c rdf:predicate dc:title1 . >>>>_:c rdf:object "a" . >>>>_:c :saidBy "b" . >>>> >>>> >>>> >>>> >>>>But in the output it does not show any result. I dont know where am I going wrong perhaps BigData developers can help with this. I am waiting for their response. >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>>On Friday, 31 October 2014 7:51 PM, Rose Beck <ros...@gm...> wrote: >>>> >>>> >>>> >>>>I tried without tmp.xml and the loading worked perfectly fine with me: >>>> >>>> >>>>curl -X POST --data-binary >>>>'uri=file:///home/bigdataAnt/SmallYagoFacts.ttl' >>>>http://194.668.5.1:9999/bigdata/namespace/reificationRDR/sparql >>>> >>>>On Fri, Oct 31, 2014 at 6:20 PM, Alice Everett <ali...@ya...> wrote: >>>>> Thanks Rose. But I dont think so.. as it works perfectly with google.com >>>>> >>>>> root:~/bigdataAnt$ curl -v google.com >>>>> * About to connect() to google.com port 80 (#0) >>>>> * Trying 74.125.236.68... connected >>>>>> GET / HTTP/1.1 >>>>>> User-Agent: curl/7.22.0 (x86_64-pc-linux-gnu) libcurl/7.22.0 OpenSSL/1.0.1 >>>>>> zlib/1.2.3.4 libidn/1.23 librtmp/2.3 >>>>>> Host: google.com >>>>>> Accept: */* >>>>>> >>>>> < HTTP/1.1 302 Found >>>>> < Cache-Control: private >>>>> < Content-Type: text/html; charset=UTF-8 >>>>> < Location: http://www.google.co.in/?gfe_rd=cr&ei=bYVTVL-gG8jM8gfBzoDgCw >>>>> < Content-Length: 261 >>>>> < Date: Fri, 31 Oct 2014 12:49:49 GMT >>>>> < Server: GFE/2.0 >>>>> < Alternate-Protocol: 80:quic,p=0.01 >>>>> < >>>>> <HTML><HEAD><meta http-equiv="content-type" >>>>> content="text/html;charset=utf-8"> >>>>> <TITLE>302 Moved</TITLE></HEAD><BODY> >>>>> <H1>302 Moved</H1> >>>>> The document has moved >>>>> <A >>>>> HREF="http://www.google.co.in/?gfe_rd=cr&ei=bYVTVL-gG8jM8gfBzoDgCw">here</A>. >>>>> </BODY></HTML> >>>>> * Connection #0 to host google.com left intact >>>>> * Closing connection #0 >>>>> >>>>> >>>>> >>>>> On Friday, 31 October 2014 6:19 PM, Rose Beck <ros...@gm...> wrote: >>>>> >>>>> >>>>> I think its a dns error..can you try doing; >>>>> >>>>> curl -v google.com >>>>> >>>>> >>>>> On Fri, Oct 31, 2014 at 6:02 PM, Bryan Thompson <br...@sy...> wrote: >>>>>> If you use POST with a URL of the resource to be loaded (see the NSS wiki >>>>>> page) then the URL must be accessible by bigdata. If you are using the >>>>>> form >>>>>> of POST that sends the data in the http request body (which is the case >>>>>> here), then it only needs to be visible to the client making the request. >>>>>> >>>>>> Thanks, >>>>>> Bryan >>>>>> >>>>>> ---- >>>>>> Bryan Thompson >>>>>> Chief Scientist & Founder >>>>>> SYSTAP, LLC >>>>>> 4501 Tower Road >>>>>> Greensboro, NC 27410 >>>>>> br...@sy... >>>>>> http://bigdata.com >>>>>> http://mapgraph.io >>>>>> >>>>>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are >>>>>> for >>>>>> the sole use of the intended recipient(s) and are confidential or >>>>>> proprietary to SYSTAP. Any unauthorized review, use, disclosure, >>>>>> dissemination or copying of this email or its contents or attachments is >>>>>> prohibited. If you have received this communication in error, please >>>>>> notify >>>>>> the sender by reply email and permanently delete all copies of the email >>>>>> and >>>>>> its contents and attachments. >>>>>> >>>>>> >>>>>> On Fri, Oct 31, 2014 at 8:30 AM, Alice Everett <ali...@ya...> >>>>>> wrote: >>>>>>> >>>>>>> Thanks Jennifer. But even keeping tmp.xml within the bigdata folder is >>>>>>> not >>>>>>> helping. >>>>>>> >>>>>>> >>>>>>> On Friday, 31 October 2014 5:57 PM, Jennifer >>>>>>> <jen...@re...> wrote: >>>>>>> >>>>>>> >>>>>>> I think she is missing as to where tmp.xml should be kept within her >>>>>>> bigdata/Ant folder as I think bigdata is not able to find tmp.xml. >>>>>>> >>>>>>> Alice I think you should keep tmp.xml within the bigdata folder which you >>>>>>> downloaded. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> From: Alice Everett <ali...@ya...> >>>>>>> Sent: Fri, 31 Oct 2014 17:47:26 >>>>>>> To: Bryan Thompson <br...@sy...> >>>>>>> Cc: "big...@li..." >>>>>>> <big...@li...> >>>>>>> Subject: Re: [Bigdata-developers] How to use RDR with Curl >>>>>>> Ok. Thanks a ton. But still I am a little lost. I used two methods of >>>>>>> inserting as explained below. My namespace's name is reificationRDR. >>>>>>> I'll be very grateful if you can help me with this a bit. >>>>>>> >>>>>>> Insert Method1: >>>>>>> root:~/bigdataAnt$ curl -v -X POST --data-binary >>>>>>> 'uri=file:///home/bigdataAnt/SmallYagoFacts.ttl' @tmp.xml >>>>>>> http://192.168.145.1:9999/bigdata/sparql >>>>>>> output: >>>>>>> * getaddrinfo(3) failed for tmp.xml:80 >>>>>>> * Couldn't resolve host 'tmp.xml' >>>>>>> * Closing connection #0 >>>>>>> curl: (6) Couldn't resolve host 'tmp.xml' >>>>>>> * About to connect() to 192.168.145.1 port 9999 (#0) >>>>>>> * Trying 192.168.145.1... connected >>>>>>> > POST /bigdata/sparql HTTP/1.1 >>>>>>> > User-Agent: curl/7.22.0 (x86_64-pc-linux-gnu) libcurl/7.22.0 >>>>>>> > OpenSSL/1.0.1 zlib/1.2.3.4 libidn/1.23 librtmp/2.3 >>>>>>> > Host: 192.168.145.1:9999 >>>>>>> > Accept: */* >>>>>>> > Content-Length: 52 >>>>>>> > Content-Type: application/x-www-form-urlencoded >>>>>>> > >>>>>>> * upload completely sent off: 52out of 52 bytes >>>>>>> < HTTP/1.1 200 OK >>>>>>> < Content-Type: application/xml; charset=ISO-8859-1 >>>>>>> < Transfer-Encoding: chunked >>>>>>> < Server: Jetty(9.1.4.v20140401) >>>>>>> < >>>>>>> * Connection #0 to host 192.168.145.1 left intact >>>>>>> * Closing connection #0 >>>>>>> >>>>>>> >>>>>>> Insert Method 2: >>>>>>> root:~/bigdataAnt/bigdata$ curl -v -X POST --data-binary >>>>>>> 'uri=file:///home/bigdataAnt/SmallYagoFacts.ttl' >>>>>>> @/home/bigdataAnt/tmp.xml >>>>>>> http://192.168.145.1:9999/bigdata/namespace/reificationRDR/sparql >>>>>>> * getaddrinfo(3) failed for :80 >>>>>>> output >>>>>>> * Couldn't resolve host '' >>>>>>> * Closing connection #0 >>>>>>> curl: (6) Couldn't resolve host '' >>>>>>> * About to connect() to 192.168.145.1 port 9999 (#0) >>>>>>> * Trying 192.168.145.1... connected >>>>>>> > POST /bigdata/namespace/reificationRDR/sparql HTTP/1.1 >>>>>>> > User-Agent: curl/7.22.0 (x86_64-pc-linux-gnu) libcurl/7.22.0 >>>>>>> > OpenSSL/1.0.1 zlib/1.2.3.4 libidn/1.23 librtmp/2.3 >>>>>>> > Host: 192.168.145.1:9999 >>>>>>> > Accept: */* >>>>>>> > Content-Length: 52 >>>>>>> > Content-Type: application/x-www-form-urlencoded >>>>>>> > >>>>>>> * upload completely sent off: 52out of 52 bytes >>>>>>> < HTTP/1.1 500 Server Error >>>>>>> < Content-Type: text/plain >>>>>>> < Transfer-Encoding: chunked >>>>>>> < Server: Jetty(9.1.4.v20140401) >>>>>>> < >>>>>>> uri=[file:/home/bigdataAnt/SmallYagoFacts.ttl], context-uri=[] >>>>>>> java.util.concurrent.ExecutionException: java.lang.RuntimeException: Not >>>>>>> found: namespace=reificationRDR >>>>>>> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >>>>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:188) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.InsertServlet.doPostWithURIs(InsertServlet.java:401) >>>>>>> at >>>>>>> com.bigdata.rdf.sail.webapp.InsertServlet.doPost(InsertServlet.java:117) >>>>>>> at com.bigdata.rdf.sail.webapp.RESTServlet.doPost(RESTServlet.java:267) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.MultiTenancyServlet.doPost(MultiTenancyServlet.java:144) >>>>>>> at javax.servlet.http.HttpServlet.service(HttpServlet.java:707) >>>>>>> at javax.servlet.http.HttpServlet.service(HttpServlet.java:790) >>>>>>> at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:738) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:551) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:568) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:221) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1111) >>>>>>> at >>>>>>> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:478) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:183) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1045) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:199) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:109) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) >>>>>>> at org.eclipse.jetty.server.Server.handle(Server.java:462) >>>>>>> at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:279) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:232) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:534) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:607) >>>>>>> at >>>>>>> >>>>>>> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:536) >>>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>>> Caused by: java.lang.RuntimeException: Not found: >>>>>>> namespace=reificationRDR >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.task.AbstractApiTask.getUnisolatedConnection(AbstractApiTask.java:217) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.InsertServlet$InsertWithURLsTask.call(InsertServlet.java:457) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.InsertServlet$InsertWithURLsTask.call(InsertServlet.java:414) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.task.ApiTaskForIndexManager.call(ApiTaskForIndexManager.java:67) >>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.task.AbstractApiTask.submitApiTask(AbstractApiTask.java:293) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.BigdataServlet.submitApiTask(BigdataServlet.java:220) >>>>>>> ... 26 more >>>>>>> * Connection #0 to host 192.168.145.1 left intact >>>>>>> * Closing connection #0 >>>>>>> >>>>>>> >>>>>>> Query: >>>>>>> curl -X POST >>>>>>> http://192.168.145.1:9999/bigdata/namespace/reificationRDR/sparql >>>>>>> --data-urlencode 'query=SELECT * {<<?s ?p ?o>> ?p1 ?o1 }' -H >>>>>>> 'Accept:application/rdf+xml' >>>>>>> >>>>>>> tmp.xml: >>>>>>> <?xml version="1.0" encoding="UTF-8" standalone="no"?> >>>>>>> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd"> >>>>>>> <properties> >>>>>>> <!-- --> >>>>>>> <!-- NEW KB NAMESPACE (required). --> >>>>>>> <!-- --> >>>>>>> <entry key="com.bigdata.rdf.sail.namespace">reificationRDR</entry> >>>>>>> <!-- --> >>>>>>> <!-- Specify any KB specific properties here to override defaults for the >>>>>>> BigdataSail --> >>>>>>> <!-- AbstractTripleStore, or indices in the namespace of the new KB >>>>>>> instance. --> >>>>>>> <!-- --> >>>>>>> <entry >>>>>>> >>>>>>> key="com.bigdata.rdf.store.AbstractTripleStore.statementIdentifiers">true</entry> >>>>>>> </properties> >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Friday, 31 October 2014 5:30 PM, Bryan Thompson <br...@sy...> >>>>>>> wrote: >>>>>>> >>>>>>> >>>>>>> What is the namespace for the RDR graph? >>>>>>> >>>>>>> The URL you need to be using is >>>>>>> >>>>>>> http://192.168.145.1:9999/bigdata/namespace/MY-GRAPH-NAMESPACE/sparql >>>>>>> >>>>>>> How to address a specific namespace is explicitly covered if you read the >>>>>>> wiki section on the multitenant interface that I linked in my previous >>>>>>> response. >>>>>>> >>>>>>> Thanks, >>>>>>> Bryan >>>>>>> >>>>>>> On Friday, October 31, 2014, Alice Everett <ali...@ya...');" >>>>>>> class="" style="" target=>ali...@ya...> wrote: >>>>>>> >>>>>>> Thanks a lot for the help. >>>>>>> >>>>>>> But I dont know where I am still going wrong: >>>>>>> I inserted data using: curl -v -X POST --data-binary >>>>>>> 'uri=file:///home/reifiedTriples.ttl' @tmp.xml >>>>>>> http://192.168.145.1:9999/bigdata/sparql >>>>>>> And then queried it using: curl -X POST >>>>>>> http://192.168.145.1:9999/bigdata/sparql --data-urlencode @tmp.xml >>>>>>> 'query=SELECT * { <<?s ?p ?o>> ?p ?o }' -H 'Accept:application/rdr' >>>>>>> curl: (6) Couldn't resolve host 'query=SELECT * <<' >>>>>>> Content-Type not recognized as RDF: application/x-www-form-urlencoded >>>>>>> >>>>>>> >>>>>>> On Friday, 31 October 2014 3:55 PM, Bryan Thompson <br...@sy...> >>>>>>> wrote: >>>>>>> >>>>>>> >>>>>>> Alice, >>>>>>> >>>>>>> The workbench choice of the "in use" namespace is recorded in java script >>>>>>> in your browser client. That choice does not effect other workbench >>>>>>> clients >>>>>>> and does not effect the behavior of the various endpoints when using >>>>>>> command >>>>>>> line tools to query or update data in the database. Thus your command >>>>>>> line >>>>>>> requests are being made against a namespace that is not configured for >>>>>>> RDR >>>>>>> support. >>>>>>> >>>>>>> If you want to address a non-default bigdata namespace using curl or >>>>>>> wget, >>>>>>> you must use the appropriate URL for that namespace. This is all >>>>>>> described >>>>>>> on wiki.bigdata.com on the page for the nanoSparqlServer in the section >>>>>>> on >>>>>>> multi-tenancy. >>>>>>> >>>>>>> See >>>>>>> http://wiki.bigdata.com/wiki/index.php/NanoSparqlServer#Multi-Tenancy_API >>>>>>> >>>>>>> Thanks, >>>>>>> Bryan >>>>>>> >>>>>>> On Thursday, October 30, 2014, Alice Everett <ali...@ya...> >>>>>>> wrote: >>>>>>> >>>>>>> I found out an awesome feature in Bigdata called RDR and I am trying to >>>>>>> explore that too. Can you please let me know as to where am I going wrong >>>>>>> while querying RDR data (http://trac.bigdata.com/ticket/815). (My sample >>>>>>> RDF >>>>>>> data, contains reification in its standard form: >>>>>>> http://www.w3.org/2001/sw/DataAccess/rq23/#queryReification) >>>>>>> Loading: >>>>>>> curl -X POST --data-binary 'uri=file:///home/SmallFacts.ttl' >>>>>>> http://192.168.145.1:9999/bigdata/sparql >>>>>>> (Additionally I changed my current namespace within the workbench opened >>>>>>> in my browser to RDR mode). >>>>>>> >>>>>>> After this I fired the following query and got the following error (Can >>>>>>> you please correct me as to where am I going wrong. I'll be very grateful >>>>>>> to >>>>>>> you for the same): >>>>>>> @HP-ProBook-4430s:~/bigdataAnt$ curl -X POST >>>>>>> http://192.168.145.1:9999/bigdata/sparql --header >>>>>>> "X-BIGDATA-MAX-QUERY-MILLIS" --data-urlencode 'query=SELECT * {<<?s ?p >>>>>>> ?o>> >>>>>>> ?p1 ?o1 }' -H 'Accept:application/rdr' >>>>>>> >>>>>>> SELECT * {<<?s ?p ?o>> ?p1 ?o1 } >>>>>>> java.util.concurrent.ExecutionException: >>>>>>> org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: >>>>>>> java.util.concurrent.ExecutionException: java.lang.RuntimeException: >>>>>>> java.util.concurrent.ExecutionException: java.lang.Exception: >>>>>>> >>>>>>> task=ChunkTask{query=eeb24f0d-29b7-49d1-bddf-14869c463e76,bopId=4,partitionId=-1,sinkId=5,altSinkId=null}, >>>>>>> cause=java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: >>>>>>> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >>>>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:188) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:1277) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:503) >>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>>> at >>>>>>> >>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>>>> at >>>>>>> >>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>>> Caused by: org.openrdf.query.QueryEvaluationException: >>>>>>> java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>>>>>> java.lang.Exception: >>>>>>> >>>>>>> task=ChunkTask{query=eeb24f0d-29b7-49d1-bddf-14869c463e76,bopId=4,partitionId=-1,sinkId=5,altSinkId=null}, >>>>>>> cause=java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: >>>>>>> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:188) >>>>>>> at >>>>>>> >>>>>>> org.openrdf.query.impl.TupleQueryResultImpl.hasNext(TupleQueryResultImpl.java:90) >>>>>>> at org.openrdf.query.QueryResultUtil.report(QueryResultUtil.java:52) >>>>>>> at >>>>>>> >>>>>>> org.openrdf.repository.sail.SailTupleQuery.evaluate(SailTupleQuery.java:63) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.BigdataRDFContext$TupleQueryTask.doQuery(BigdataRDFContext.java:1386) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask$SparqlRestApiTask.call(BigdataRDFContext.java:1221) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask$SparqlRestApiTask.call(BigdataRDFContext.java:1171) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.task.ApiTaskForIndexManager.call(ApiTaskForIndexManager.java:67) >>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.task.AbstractApiTask.submitApiTask(AbstractApiTask.java:293) >>>>>>> ... 6 more >>>>>>> Caused by: java.lang.RuntimeException: >>>>>>> java.util.concurrent.ExecutionException: java.lang.RuntimeException: >>>>>>> java.util.concurrent.ExecutionException: java.lang.Exception: >>>>>>> >>>>>>> task=ChunkTask{query=eeb24f0d-29b7-49d1-bddf-14869c463e76,bopId=4,partitionId=-1,sinkId=5,altSinkId=null}, >>>>>>> cause=java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: >>>>>>> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at >>>>>>> >>>>>>> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1523) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator._hasNext(BlockingBuffer.java:1710) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.hasNext(BlockingBuffer.java:1563) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.striterator.AbstractChunkedResolverator._hasNext(AbstractChunkedResolverator.java:365) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.striterator.AbstractChunkedResolverator.hasNext(AbstractChunkedResolverator.java:341) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:134) >>>>>>> ... 15 more >>>>>>> Caused by: java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>>>>>> java.lang.Exception: >>>>>>> >>>>>>> task=ChunkTask{query=eeb24f0d-29b7-49d1-bddf-14869c463e76,bopId=4,partitionId=-1,sinkId=5,altSinkId=null}, >>>>>>> cause=java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: >>>>>>> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >>>>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:188) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1454) >>>>>>> ... 20 more >>>>>>> Caused by: java.lang.RuntimeException: >>>>>>> java.util.concurrent.ExecutionException: java.lang.Exception: >>>>>>> >>>>>>> task=ChunkTask{query=eeb24f0d-29b7-49d1-bddf-14869c463e76,bopId=4,partitionId=-1,sinkId=5,altSinkId=null}, >>>>>>> cause=java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: >>>>>>> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:59) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.RunningQueryCloseableIterator.close(RunningQueryCloseableIterator.java:73) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.RunningQueryCloseableIterator.hasNext(RunningQueryCloseableIterator.java:82) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.striterator.ChunkedWrappedIterator.hasNext(ChunkedWrappedIterator.java:197) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:222) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:197) >>>>>>> >>>>>>> ... 4 more >>>>>>> Caused by: java.util.concurrent.ExecutionException: java.lang.Exception: >>>>>>> >>>>>>> task=ChunkTask{query=eeb24f0d-29b7-49d1-bddf-14869c463e76,bopId=4,partitionId=-1,sinkId=5,altSinkId=null}, >>>>>>> cause=java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: >>>>>>> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at com.bigdata.util.concurrent.Haltable.get(Haltable.java:273) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:1476) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:103) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:46) >>>>>>> ... 9 more >>>>>>> Caused by: java.lang.Exception: >>>>>>> >>>>>>> task=ChunkTask{query=eeb24f0d-29b7-49d1-bddf-14869c463e76,bopId=4,partitionId=-1,sinkId=5,altSinkId=null}, >>>>>>> cause=java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: >>>>>>> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1335) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTaskWrapper.run(ChunkedRunningQuery.java:894) >>>>>>> at >>>>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>>> at com.bigdata.concurrent.FutureTaskMon.run(FutureTaskMon.java:63) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkFutureTask.run(ChunkedRunningQuery.java:789) >>>>>>> ... 3 more >>>>>>> Caused by: java.util.concurrent.ExecutionException: >>>>>>> java.lang.RuntimeException: java.lang.RuntimeException: >>>>>>> java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >>>>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:188) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1315) >>>>>>> ... 8 more >>>>>>> Caused by: java.lang.RuntimeException: java.lang.RuntimeException: >>>>>>> java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at com.bigdata.bop.join.PipelineJoin$JoinTask.call(PipelineJoin.java:643) >>>>>>> at com.bigdata.bop.join.PipelineJoin$JoinTask.call(PipelineJoin.java:343) >>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>>> at com.bigdata.concurrent.FutureTaskMon.run(FutureTaskMon.java:63) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1314) >>>>>>> ... 8 more >>>>>>> Caused by: java.lang.RuntimeException: >>>>>>> java.lang.ArrayIndexOutOfBoundsException: 0 >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.join.PipelineJoin$JoinTask$BindingSetConsumerTask.call(PipelineJoin.java:988) >>>>>>> at >>>>>>> >>>>>>> com.bigdata.bop.join.PipelineJoin$JoinTask.consumeSource(PipelineJoin.java:700) >>>>>>> at com.bigdata.bop.join.PipelineJoin$JoinTask.call(PipelineJoin.java:584) >>>> -- ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://bigdata.com http://mapgraph.io CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |