This list is closed, nobody may subscribe to it.
2010 |
Jan
|
Feb
(19) |
Mar
(8) |
Apr
(25) |
May
(16) |
Jun
(77) |
Jul
(131) |
Aug
(76) |
Sep
(30) |
Oct
(7) |
Nov
(3) |
Dec
|
---|---|---|---|---|---|---|---|---|---|---|---|---|
2011 |
Jan
|
Feb
|
Mar
|
Apr
|
May
(2) |
Jun
(2) |
Jul
(16) |
Aug
(3) |
Sep
(1) |
Oct
|
Nov
(7) |
Dec
(7) |
2012 |
Jan
(10) |
Feb
(1) |
Mar
(8) |
Apr
(6) |
May
(1) |
Jun
(3) |
Jul
(1) |
Aug
|
Sep
(1) |
Oct
|
Nov
(8) |
Dec
(2) |
2013 |
Jan
(5) |
Feb
(12) |
Mar
(2) |
Apr
(1) |
May
(1) |
Jun
(1) |
Jul
(22) |
Aug
(50) |
Sep
(31) |
Oct
(64) |
Nov
(83) |
Dec
(28) |
2014 |
Jan
(31) |
Feb
(18) |
Mar
(27) |
Apr
(39) |
May
(45) |
Jun
(15) |
Jul
(6) |
Aug
(27) |
Sep
(6) |
Oct
(67) |
Nov
(70) |
Dec
(1) |
2015 |
Jan
(3) |
Feb
(18) |
Mar
(22) |
Apr
(121) |
May
(42) |
Jun
(17) |
Jul
(8) |
Aug
(11) |
Sep
(26) |
Oct
(15) |
Nov
(66) |
Dec
(38) |
2016 |
Jan
(14) |
Feb
(59) |
Mar
(28) |
Apr
(44) |
May
(21) |
Jun
(12) |
Jul
(9) |
Aug
(11) |
Sep
(4) |
Oct
(2) |
Nov
(1) |
Dec
|
2017 |
Jan
(20) |
Feb
(7) |
Mar
(4) |
Apr
(18) |
May
(7) |
Jun
(3) |
Jul
(13) |
Aug
(2) |
Sep
(4) |
Oct
(9) |
Nov
(2) |
Dec
(5) |
2018 |
Jan
|
Feb
|
Mar
|
Apr
(2) |
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
2019 |
Jan
|
Feb
|
Mar
(1) |
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
From: Joakim S. <joa...@bl...> - 2015-11-18 15:41:17
|
I have been running Blazegraph 1.5.2 in embedded mode (apache-tomcat-8.0.27) for a few months ( Ill upgrade to 1.5.3 when available on Maven). After moving Tomcat to another volume with a minor update (apache-tomcat-8.0.27) I can no longer run my service. The servlet initializes twice. The first time Blazegraph engine starts, but it seems that at the end of its loading process, it fires an new init() to the servlet (2015-11-18 07:25:37,697) INFO : DataManager.java:62: Default configuration loaded: {serverPathTojnl=/mydata/blazegraph/} (2015-11-18 07:25:37,699) INFO : DataManager.java:78: Blazegraph configuration loaded: {com.bigdata.namespace.kb.spo.com.bigdata.btree.BTree.branchingFactor=1024, com.bigdata.journal.AbstractJournal.bufferMode=DiskRW, com.bigdata.journal.AbstractJournal.file=/mydata/blazegraph/dbpedia2015.jnl, com.bigdata.journal.AbstractJournal.initialExtent=209715200, com.bigdata.rdf.store.AbstractTripleStore.textIndex=true , com.bigdata.btree.BTree.branchingFactor=512, com.bigdata.namespace.kb.lex.com.bigdata.btree.BTree.branchingFactor=400, com.bigdata.rdf.store.AbstractTripleStore.axiomsClass=com.bigdata.rdf.axioms.NoAxioms, com.bigdata.rdf.sail.bufferCapacity=100000, com.bigdata.service.AbstractTransactionService.minReleaseAge=1, com.bigdata.rdf.sail.truthMaintenance=false , com.bigdata.journal.AbstractJournal.maximumExtent=209715200, com.bigdata.rdf.store.AbstractTripleStore.justify=false, com.bigdata.namespace.kb.spo.OSP.com.bigdata.btree.BTree.branchingFactor=512, com.bigdata.rdf.store.AbstractTripleStore.quads=false, com.bigdata.namespace.kb.spo.POS.com.bigdata.btree.BTree.branchingFactor=512, com.bigdata.rdf.rio.RDFParserOptions.datatypeHandling=IGNORE, com.bigdata.btree.writeRetentionQueue.capacity=8000, com.bigdata.rdf.rio.RDFParserOptions.verifyData=false, com.bigdata.rdf.store.DataLoader.commit=Incremental, com.bigdata.rdf.rio.RDFParserOptions.stopAtFirstError=false, com.bigdata.rdf.rio.RDFParserOptions.preserveBNodeIDs=false, com.bigdata.namespace.kb.spo.SPO.com.bigdata.btree.BTree.branchingFactor=512, com.bigdata.rdf.store.AbstractTripleStore.statementIdentifiers=false} (2015-11-18 07:25:37,744) INFO : SystemUtil.java:90: architecture: amd64 (2015-11-18 07:25:37,744) INFO : SystemUtil.java:91: operating system: Linux (2015-11-18 07:25:37,745) INFO : SystemUtil.java:92: operation system version: 3.10.0-229.20.1.el7.x86_64 (2015-11-18 07:25:37,745) INFO : SystemUtil.java:93: #processors: 8 (2015-11-18 07:25:37,745) INFO : SystemUtil.java:94: cpu info: Intel(R) Xeon(R) CPU E5-2670 v2 @ 2.50GHz Family 6 Model 62 Stepping 4, GenuineIntel BlazeGraph(TM) Graph Engine Flexible Reliable Affordable Web-Scale Computing for the Enterprise Copyright SYSTAP, LLC 2006-2015. All rights reserved. ip-172-31-11-213.us-west-2.compute.internal Wed Nov 18 07:25:37 UTC 2015 Linux/3.10.0-229.20.1.el7.x86_64 amd64 Intel(R) Xeon(R) CPU E5-2670 v2 @ 2.50GHz Family 6 Model 62 Stepping 4, GenuineIntel #CPU=8 Oracle Corporation 1.8.0_45 freeMemory=1446459088 buildVersion=1.5.2-SNAPSHOT Dependency License ICU http://source.icu-project.org/repos/icu/icu/trunk/license.html bigdata-ganglia http://www.apache.org/licenses/LICENSE-2.0.html blueprints-core https://github.com/tinkerpop/blueprints/blob/master/LICENSE.txt colt http://acs.lbl.gov/software/colt/license.html commons-codec http://www.apache.org/licenses/LICENSE-2.0.html commons-fileupload http://www.apache.org/licenses/LICENSE-2.0.html commons-io http://www.apache.org/licenses/LICENSE-2.0.html commons-logging http://www.apache.org/licenses/LICENSE-2.0.html dsiutils http://www.gnu.org/licenses/lgpl-2.1.html fastutil http://www.apache.org/licenses/LICENSE-2.0.html flot http://www.opensource.org/licenses/mit-license.php high-scale-lib http://creativecommons.org/licenses/publicdomain httpclient http://www.apache.org/licenses/LICENSE-2.0.html httpclient-cache http://www.apache.org/licenses/LICENSE-2.0.html httpcore http://www.apache.org/licenses/LICENSE-2.0.html httpmime http://www.apache.org/licenses/LICENSE-2.0.html jackson-core http://www.apache.org/licenses/LICENSE-2.0.html jetty http://www.apache.org/licenses/LICENSE-2.0.html jquery https://github.com/jquery/jquery/blob/master/MIT-LICENSE.txt log4j http://www.apache.org/licenses/LICENSE-2.0.html lucene http://www.apache.org/licenses/LICENSE-2.0.html nanohttp http://elonen.iki.fi/code/nanohttpd/#license rexster-core https://github.com/tinkerpop/rexster/blob/master/LICENSE.txt river http://www.apache.org/licenses/LICENSE-2.0.html servlet-api http://www.apache.org/licenses/LICENSE-2.0.html sesame http://www.openrdf.org/download.jsp slf4j http://www.slf4j.org/license.html zookeeper http://www.apache.org/licenses/LICENSE-2.0.html INFO: com.bigdata.util.config.LogUtil: Configure: file:/myserver/opt/apache-tomcat-8.0.27/webapps/dbpedia/WEB-INF/classes/log4j.properties (2015-11-18 07:25:37,754) DEBUG: AbstractDynamicMBean.java:148: preRegister called. Server=com.sun.jmx.mbeanserver.JmxMBeanServer@16b3fc9e, name=log4j:hierarchy=default (2015-11-18 07:25:37,754) DEBUG: HierarchyDynamicMBean.java:260: postRegister is called. (2015-11-18 07:25:38,078) WARN : ServiceProviderHook.java:172: Running. (2015-11-18 07:25:38,087) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.turtle.TurtleParserFactory (2015-11-18 07:25:38,088) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.ntriples.NTriplesParserFactory (2015-11-18 07:25:38,089) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.n3.N3ParserFactory (2015-11-18 07:25:38,089) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.nquads.NQuadsParserFactory (2015-11-18 07:25:38,090) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.binary.BinaryRDFParserFactory (2015-11-18 07:25:38,091) DEBUG: ServiceRegistry.java:61: Registered service class com.bigdata.rdf.rio.ntriples.BigdataNTriplesParserFactory (2015-11-18 07:25:38,091) DEBUG: ServiceRegistry.java:61: Registered service class com.bigdata.rdf.rio.turtle.BigdataTurtleParserFactory (2015-11-18 07:25:38,092) DEBUG: ServiceRegistry.java:61: Registered service class com.bigdata.rdf.rio.json.BigdataSPARQLResultsJSONParserForConstructFactory (2015-11-18 07:25:38,092) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.rdfxml.RDFXMLParserFactory (2015-11-18 07:25:38,093) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.rdfjson.RDFJSONParserFactory (2015-11-18 07:25:38,096) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.trig.TriGParserFactory (2015-11-18 07:25:38,097) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.rio.trix.TriXParserFactory (2015-11-18 07:25:38,104) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.binary.BinaryQueryResultWriterFactory (2015-11-18 07:25:38,104) DEBUG: ServiceRegistry.java:61: Registered service class com.bigdata.rdf.rio.json.BigdataSPARQLResultsJSONWriterFactory (2015-11-18 07:25:38,105) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.text.csv.SPARQLResultsCSVWriterFactory (2015-11-18 07:25:38,105) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.text.tsv.SPARQLResultsTSVWriterFactory (2015-11-18 07:25:38,106) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.sparqlxml.SPARQLResultsXMLWriterFactory (2015-11-18 07:25:38,109) WARN : ServiceRegistry.java:57: New service class org.openrdf.query.resultio.sparqljson.SPARQLResultsJSONWriterFactory replaces existing service class com.bigdata.rdf.rio.json.BigdataSPARQLResultsJSONWriterFactory (2015-11-18 07:25:38,109) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.sparqljson.SPARQLResultsJSONWriterFactory (2015-11-18 07:25:38,114) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.binary.BinaryQueryResultParserFactory (2015-11-18 07:25:38,115) DEBUG: ServiceRegistry.java:61: Registered service class com.bigdata.rdf.rio.json.BigdataSPARQLResultsJSONParserFactory (2015-11-18 07:25:38,115) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.text.csv.SPARQLResultsCSVParserFactory (2015-11-18 07:25:38,116) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.text.tsv.SPARQLResultsTSVParserFactory (2015-11-18 07:25:38,116) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.sparqlxml.SPARQLResultsXMLParserFactory (2015-11-18 07:25:38,119) WARN : ServiceRegistry.java:57: New service class org.openrdf.query.resultio.sparqljson.SPARQLResultsJSONParserFactory replaces existing service class com.bigdata.rdf.rio.json.BigdataSPARQLResultsJSONParserFactory (2015-11-18 07:25:38,119) DEBUG: ServiceRegistry.java:61: Registered service class org.openrdf.query.resultio.sparqljson.SPARQLResultsJSONParserFactory realPath: /myserver/opt/apache-tomcat-8.0.27/webapps/dbpedia/ <== here it has restarted init() log4j:WARN No appenders could be found for logger (org.apache.log4j.jmx.HierarchyDynamicMBean). <== what is this? zookeeper? log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. (2015-11-18 07:25:38,126) DEBUG: PropertyConfigurator.java:770: addAppenderEvent called: logger=root, appender=devDest (2015-11-18 07:25:38,126) DEBUG: PropertyConfigurator.java:770: sending notification. (2015-11-18 07:25:38,127) DEBUG: PropertyConfigurator.java:759: removeAppenderCalled: logger=com.bigdata.relation.rule.eval.RuleLog, appender=ruleLog (2015-11-18 07:25:38,128) DEBUG: PropertyConfigurator.java:770: addAppenderEvent called: logger=com.bigdata.relation.rule.eval.RuleLog, appender=ruleLog (2015-11-18 07:25:38,128) DEBUG: PropertyConfigurator.java:770: sending notification. (2015-11-18 07:25:38,128) INFO : DataManager.java:62: Default configuration loaded: {serverPathTojnl=/mydata/blazegraph/} (2015-11-18 07:25:38,129) INFO : DataManager.java:78: Blazegraph configuration loaded: {com.bigdata.namespace.kb.spo.com.bigdata.btree.BTree.branchingFactor=1024, com.bigdata.journal.AbstractJournal.bufferMode=DiskRW, com.bigdata.journal.AbstractJournal.file=/mydata/blazegraph/dbpedia2015.jnl, com.bigdata.journal.AbstractJournal.initialExtent=209715200, com.bigdata.rdf.store.AbstractTripleStore.textIndex=true , com.bigdata.btree.BTree.branchingFactor=512, com.bigdata.namespace.kb.lex.com.bigdata.btree.BTree.branchingFactor=400, com.bigdata.rdf.store.AbstractTripleStore.axiomsClass=com.bigdata.rdf.axioms.NoAxioms, com.bigdata.rdf.sail.bufferCapacity=100000, com.bigdata.service.AbstractTransactionService.minReleaseAge=1, com.bigdata.rdf.sail.truthMaintenance=false , com.bigdata.journal.AbstractJournal.maximumExtent=209715200, com.bigdata.rdf.store.AbstractTripleStore.justify=false, com.bigdata.namespace.kb.spo.OSP.com.bigdata.btree.BTree.branchingFactor=512, com.bigdata.rdf.store.AbstractTripleStore.quads=false, com.bigdata.namespace.kb.spo.POS.com.bigdata.btree.BTree.branchingFactor=512, com.bigdata.rdf.rio.RDFParserOptions.datatypeHandling=IGNORE, com.bigdata.btree.writeRetentionQueue.capacity=8000, com.bigdata.rdf.rio.RDFParserOptions.verifyData=false, com.bigdata.rdf.store.DataLoader.commit=Incremental, com.bigdata.rdf.rio.RDFParserOptions.stopAtFirstError=false, com.bigdata.rdf.rio.RDFParserOptions.preserveBNodeIDs=false, com.bigdata.namespace.kb.spo.SPO.com.bigdata.btree.BTree.branchingFactor=512, com.bigdata.rdf.store.AbstractTripleStore.statementIdentifiers=false} <= it loads the blazegrpah properties again the line far this is sail = new BigdataSail(blazegraphsProps); which hangs |
From: Michael S. <ms...@me...> - 2015-11-17 20:02:49
|
Hi Quentin, > With regards to the query, is the use of hint:Prior hint:runFirst "true" in a query similar to having DISTINCT in theSELECT clause? I have tested the query with the added pattern and it returned the result set that I expected. actually no: the query hint “runFirst” just tells the optimizer that the preceding construct (“hint:Prior”), namely the SERVICE, should be executed first — this should not change the semantics (respectively outcome) of the query, i.e. in both cases the DISTINCT keyword *should* not be required. Said otherwise: in case we run the SERVICE last there seems to be a problem in the query engine (which is what you reported), while the engine's behavior with SERVICE run first is correct. I’ve created a ticket at https://jira.blazegraph.com/browse/BLZG-1636 <https://jira.blazegraph.com/browse/BLZG-1636> describing the behavior and will have a look at it in the coming days. If you are able to share the data (or could provide a snippet of the data that allows to reproduce the problem) please let me know, this would help in debugging. Best, Michael > > With regards to the runtime statistics, I have added a screenshot of it. > > Thanks again for your help. > > Quentin Reul > Advanced Technology | Global Platform Organization | +1 (917) 891 5490 > Email: Que...@wo... <mailto:Que...@wo...> > Skype: quentin_reul > > From: Michael Schmidt [mailto:ms...@me...] > Sent: Sunday, November 15, 2015 3:42 AM > To: Reul, Quentin > Cc: Bryan Thompson; big...@li... > Subject: Re: [Bigdata-developers] Duplicate in SELECT queries when using SERVICE > > Dear Quentin, > > I’ve quickly set up you initial scenario with two triples in each namespace and was not able to reproduce the behavior, i.e. I always get one result there. > > Looking at your query, it is much more complex. In fact, ?profileURI might be bound to the same URI multiple times when evaluating the SERVICE, which could lead to duplicates (@Bryan: it might be that we’re missing out a distinct projection here? Though this would not explain why results vary from time to time.). Could you try running the query *without* the SERVICE call and confirm that the number of results is stable? Also, how many results do you get then? And do you know whether there a skos:prefLabel for each of them? (just trying to nail down what’s going on) > > Also, right at the end of the query plan in the EXPLAIN is a table showing runtime statistics (in particular, how many results were flowing through the operators). It would be quite useful to have that one too (both for the original query and the query without SERVICE). Could you share that as well, preferably as a screenshot. > > Regarding the option to run the SERVICE first: as you mentioned already, it does not help just inverting the order — the optimizer makes its own decision, independently from the order in which you write things. What you need there is a query hint that forces the optimizer to run the SERVICE keyword first (see triple pattern in the last line, to be placed right after the SERVICE): > > PREFIX dcterms: <http://purl.org/dc/terms/ <http://purl.org/dc/terms/>> > PREFIX skos: <http://www.w3.org/2004/02/skos/core# <http://www.w3.org/2004/02/skos/core>> > > SELECT ?doc ?concept ?label > WHERE { > ?doc dcterms:subject ?concept . > SERVICE <http://localhost:9999/bigdata/namespace/skos/sparql <http://localhost:9999/bigdata/namespace/skos/sparql>> { > ?concept skos:prefLabel ?label . > } > hint:Prior hint:runFirst "true" . > } > > However, using that hint for your query would mean extracting all "?profileStatusURI skos:prefLabel ?profileStatusLabel” patterns from the remote endpoint first (without any restriction), so from a performance perspective this might not be the best option. But I’d be interested in the result of this query, i.e. whether you still get duplicates. > > Best, > Michael > > On 13 Nov 2015, at 23:24, Reul, Quentin <que...@wo... <mailto:que...@wo...>> wrote: > > Hi Bryan, > > I have checked that there were not any duplicate triples in either namespaces. > > For the query in the attached file, I should retrieve 21 records in my result set (which I do when DISTINCT is used). However, I retrieve 79 records in my result set when DISTINCT is not used. I have tried to use the SERVICE part of the query as the first part of the query, but I obtained the same results. > > Kind regards, > > Quentin Reul > > From: Bryan Thompson [mailto:br...@sy... <mailto:br...@sy...>] > Sent: Friday, November 13, 2015 4:01 PM > To: Reul, Quentin > Cc: big...@li... <mailto:big...@li...> > Subject: Re: [Bigdata-developers] Duplicate in SELECT queries when using SERVICE > > Quentin, > > Can you please look at the EXPLAIN of the query (if you are using the workbench, this is a checkbox under the advanced options, otherwise you can just add &explain to the query - see [1]). I am curious whether the triple pattern is running before or after the SERVICE call. > > Can you please confirm that you are observing that behavior on endpoints having just those two triples each? If there are many triples in the first endpoint, then one possibility is that the SERVICE call is being invoked more than one because multiple "chunks" of solutions are presented to it. In this case, the join could well have duplicates (same ?concept in different chunks leading to more than one solution with the same bindings from the B endpoint) and a DISTINCT would be required. > > Another option would be to run the SERVICE first. This could be accomplished with a query hint. See [2]. It is also possible to force certain operations to run exactly once. For example, but pushing something into a subquery and using the runOnce query hint. > > Another thing that you can try is to enable the solutions logger. This logger provides details on the inputs and outputs of each operator. You will be able to see the solutions produced by the triple pattern and those produced by the SERVICE call for each invocation of those operators. This will help you to understand whether multiple SERVICE invocations are occurring. > > ## > # Solutions trace (tab delimited file). Uncomment the next line to enable. > #log4j.logger.com.bigdata.bop.engine.SolutionsLog=INFO,solutionsLog > log4j.additivity.com.bigdata.bop.engine.SolutionsLog=false > log4j.appender.solutionsLog=org.apache.log4j.ConsoleAppender > #log4j.appender.solutionsLog=org.apache.log4j.FileAppender > log4j.appender.solutionsLog.Threshold=ALL > #log4j.appender.solutionsLog.File=solutions.csv > #log4j.appender.solutionsLog.Append=true > # I find that it is nicer to have this unbuffered since you can see what > # is going on and to make sure that I have complete rule evaluation logs > # on shutdown. > #log4j.appender.solutionsLog.BufferedIO=false > log4j.appender.solutionsLog.layout=org.apache.log4j.PatternLayout > log4j.appender.solutionsLog.layout.ConversionPattern=SOLUTION:\t%m > > If your investigations do not suggest an obvious solution, then it might be best if you create a ticket from this query and attach the EXPLAIN (which is an html page) to that ticket. Please add both myself and Michael Schmidt to the ticket as watchers so we will see any updates on the ticket. > > Thanks, > Bryan > > [1] https://wiki.blazegraph.com/wiki/index.php/Explain#NSS_Explain_Mode <https://wiki.blazegraph.com/wiki/index.php/Explain#NSS_Explain_Mode> > [2] https://wiki.blazegraph.com/wiki/index.php/QueryHints <https://wiki.blazegraph.com/wiki/index.php/QueryHints> > > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... <mailto:br...@sy...> > http://blazegraph.com <http://blazegraph.com/> > http://blog.blazegraph.com <http://blog.blazegraph.com/> > > Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > On Fri, Nov 13, 2015 at 11:35 AM, Reul, Quentin <que...@wo... <mailto:que...@wo...>> wrote: > Hi all, > > I'm encountering a weird behaviour when running SPARQL SELECT queries including SERVICE in the WHERE clause. More specifically, I seem to retrieve duplicate records in the result set when DISTINCT is not used. We are using RemoteRepositoryManager to access the BlazeGraph 1.5.2 instance and both namespaces are defined on the same machine. > > Let us imagine that I have the following triples in namespace A: > <doc1> dcterms:subject <conceptB> > <doc1> dcterms:title "Title of Document"^^xsd:string > > and some triples in namespace B: > <conceptB> rdf:type skos:Concept > <conceptB> skos:prefLabel "concept label"@en > > If I run the following SELECT query > SELECT ?doc ?concept ?label > WHERE { > ?doc dcterms:subject ?concept . > SERVICE <http:localhost:9999/bigdata/namespace/B/sparql> { > ?concept skos:prefLabel ?label . > } > } > then I would the following result set: > || ?doc || ?concept || ?label || > | <doc1> | <conceptB> | "concept label" | > | <doc1> | <conceptB> | "concept label" | > | <doc1> | <conceptB> | "concept label" | > > Interestingly, the number of duplicated records can change from processing to processing. Is this something that other people have encountered? > > Kind regards, > > Quentin Reul > > > ------------------------------------------------------------------------------ > > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> > > <sparql_query_explain.txt>------------------------------------------------------------------------------ > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> > > <BG - Runtime statistics on SERVICE.jpg> |
From: Reul, Q. <que...@wo...> - 2015-11-16 19:19:59
|
Hi Michael, Thanks for the detailed explanation. With regards to the query, is the use of hint:Prior hint:runFirst "true" in a query similar to having DISTINCT in the SELECT clause? I have tested the query with the added pattern and it returned the result set that I expected. With regards to the runtime statistics, I have added a screenshot of it. Thanks again for your help. Quentin Reul Advanced Technology | Global Platform Organization | +1 (917) 891 5490 Email: Que...@wo...<mailto:Que...@wo...> Skype: quentin_reul From: Michael Schmidt [mailto:ms...@me...] Sent: Sunday, November 15, 2015 3:42 AM To: Reul, Quentin Cc: Bryan Thompson; big...@li... Subject: Re: [Bigdata-developers] Duplicate in SELECT queries when using SERVICE Dear Quentin, I’ve quickly set up you initial scenario with two triples in each namespace and was not able to reproduce the behavior, i.e. I always get one result there. Looking at your query, it is much more complex. In fact, ?profileURI might be bound to the same URI multiple times when evaluating the SERVICE, which could lead to duplicates (@Bryan: it might be that we’re missing out a distinct projection here? Though this would not explain why results vary from time to time.). Could you try running the query *without* the SERVICE call and confirm that the number of results is stable? Also, how many results do you get then? And do you know whether there a skos:prefLabel for each of them? (just trying to nail down what’s going on) Also, right at the end of the query plan in the EXPLAIN is a table showing runtime statistics (in particular, how many results were flowing through the operators). It would be quite useful to have that one too (both for the original query and the query without SERVICE). Could you share that as well, preferably as a screenshot. Regarding the option to run the SERVICE first: as you mentioned already, it does not help just inverting the order — the optimizer makes its own decision, independently from the order in which you write things. What you need there is a query hint that forces the optimizer to run the SERVICE keyword first (see triple pattern in the last line, to be placed right after the SERVICE): PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX skos: <http://www.w3.org/2004/02/skos/core#<http://www.w3.org/2004/02/skos/core>> SELECT ?doc ?concept ?label WHERE { ?doc dcterms:subject ?concept . SERVICE <http://localhost:9999/bigdata/namespace/skos/sparql> { ?concept skos:prefLabel ?label . } hint:Prior hint:runFirst "true" . } However, using that hint for your query would mean extracting all "?profileStatusURI skos:prefLabel ?profileStatusLabel” patterns from the remote endpoint first (without any restriction), so from a performance perspective this might not be the best option. But I’d be interested in the result of this query, i.e. whether you still get duplicates. Best, Michael On 13 Nov 2015, at 23:24, Reul, Quentin <que...@wo...<mailto:que...@wo...>> wrote: Hi Bryan, I have checked that there were not any duplicate triples in either namespaces. For the query in the attached file, I should retrieve 21 records in my result set (which I do when DISTINCT is used). However, I retrieve 79 records in my result set when DISTINCT is not used. I have tried to use the SERVICE part of the query as the first part of the query, but I obtained the same results. Kind regards, Quentin Reul From: Bryan Thompson [mailto:br...@sy...] Sent: Friday, November 13, 2015 4:01 PM To: Reul, Quentin Cc: big...@li...<mailto:big...@li...> Subject: Re: [Bigdata-developers] Duplicate in SELECT queries when using SERVICE Quentin, Can you please look at the EXPLAIN of the query (if you are using the workbench, this is a checkbox under the advanced options, otherwise you can just add &explain to the query - see [1]). I am curious whether the triple pattern is running before or after the SERVICE call. Can you please confirm that you are observing that behavior on endpoints having just those two triples each? If there are many triples in the first endpoint, then one possibility is that the SERVICE call is being invoked more than one because multiple "chunks" of solutions are presented to it. In this case, the join could well have duplicates (same ?concept in different chunks leading to more than one solution with the same bindings from the B endpoint) and a DISTINCT would be required. Another option would be to run the SERVICE first. This could be accomplished with a query hint. See [2]. It is also possible to force certain operations to run exactly once. For example, but pushing something into a subquery and using the runOnce query hint. Another thing that you can try is to enable the solutions logger. This logger provides details on the inputs and outputs of each operator. You will be able to see the solutions produced by the triple pattern and those produced by the SERVICE call for each invocation of those operators. This will help you to understand whether multiple SERVICE invocations are occurring. ## # Solutions trace (tab delimited file). Uncomment the next line to enable. #log4j.logger.com.bigdata.bop.engine.SolutionsLog=INFO,solutionsLog log4j.additivity.com.bigdata.bop.engine.SolutionsLog=false log4j.appender.solutionsLog=org.apache.log4j.ConsoleAppender #log4j.appender.solutionsLog=org.apache.log4j.FileAppender log4j.appender.solutionsLog.Threshold=ALL #log4j.appender.solutionsLog.File=solutions.csv #log4j.appender.solutionsLog.Append=true # I find that it is nicer to have this unbuffered since you can see what # is going on and to make sure that I have complete rule evaluation logs # on shutdown. #log4j.appender.solutionsLog.BufferedIO=false log4j.appender.solutionsLog.layout=org.apache.log4j.PatternLayout log4j.appender.solutionsLog.layout.ConversionPattern=SOLUTION:\t%m If your investigations do not suggest an obvious solution, then it might be best if you create a ticket from this query and attach the EXPLAIN (which is an html page) to that ticket. Please add both myself and Michael Schmidt to the ticket as watchers so we will see any updates on the ticket. Thanks, Bryan [1] https://wiki.blazegraph.com/wiki/index.php/Explain#NSS_Explain_Mode [2] https://wiki.blazegraph.com/wiki/index.php/QueryHints ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy...<mailto:br...@sy...> http://blazegraph.com<http://blazegraph.com/> http://blog.blazegraph.com<http://blog.blazegraph.com/> Blazegraph™<http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Fri, Nov 13, 2015 at 11:35 AM, Reul, Quentin <que...@wo...<mailto:que...@wo...>> wrote: Hi all, I'm encountering a weird behaviour when running SPARQL SELECT queries including SERVICE in the WHERE clause. More specifically, I seem to retrieve duplicate records in the result set when DISTINCT is not used. We are using RemoteRepositoryManager to access the BlazeGraph 1.5.2 instance and both namespaces are defined on the same machine. Let us imagine that I have the following triples in namespace A: <doc1> dcterms:subject <conceptB> <doc1> dcterms:title "Title of Document"^^xsd:string and some triples in namespace B: <conceptB> rdf:type skos:Concept <conceptB> skos:prefLabel "concept label"@en If I run the following SELECT query SELECT ?doc ?concept ?label WHERE { ?doc dcterms:subject ?concept . SERVICE <http:localhost:9999/bigdata/namespace/B/sparql> { ?concept skos:prefLabel ?label . } } then I would the following result set: || ?doc || ?concept || ?label || | <doc1> | <conceptB> | "concept label" | | <doc1> | <conceptB> | "concept label" | | <doc1> | <conceptB> | "concept label" | Interestingly, the number of duplicated records can change from processing to processing. Is this something that other people have encountered? Kind regards, Quentin Reul ------------------------------------------------------------------------------ _______________________________________________ Bigdata-developers mailing list Big...@li...<mailto:Big...@li...> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <sparql_query_explain.txt>------------------------------------------------------------------------------ _______________________________________________ Bigdata-developers mailing list Big...@li...<mailto:Big...@li...> https://lists.sourceforge.net/lists/listinfo/bigdata-developers |
From: Brad B. <be...@sy...> - 2015-11-16 14:31:12
|
Felix, Great -- yes, please do post your results to the mailing list. Regarding using the Vocabulary, yes, you must create a new Journal with the property for the vocabulary class as you have in your example. When you start Blazegraph, you must also make sure that the vocabulary class is in the classpath, i.e. something like java -cp /path/to/vocab -jar bigdata-1.5.2-bundled.jar. We would recommend that you update to the 1.5.3 bundled jar, if possible. The 2.0 release at the end of the year will have a number of loading enhancements as well. com.bigdata.rdf.store.AbstractTripleStore.vocabularyClass=MyCustomVocab Good luck and let us know how it goes. Thanks, --Brad On Mon, Nov 16, 2015 at 3:36 AM, Felix Conrads < mai...@st...> wrote: > Hey, > > First off thanks alot! It took a while to come back to it, as some other > work showed up which needs to be done first. > Nevertheless i have one more question: > > I took the fastload properties and generated a Vocabulary exactly as your > JIRA Ticket [2]. How do i say blazegraph it should use my vocab? Is it as > simple as to put in the properties following line: > com.bigdata.rdf.store.AbstractTripleStore.vocabularyClass=MyCustomVocab > (Generated File is MyCustomVocab.java in the directory with > bigdata-1.5.2-bundled.jar) > > > I'll test Blazegraph with a DBPedia Dataset and a SWDF Dataset with 250 > Queries (ASK, SELECT, CONSTRUCT, DESCRIBE) generated with FEASIBLE [1] and > at least DBPedia with Updates from DBpedia Live [2] (round about 250 > inserts and deletes). The test should show how the triplestores act when > they to not just only have 1 SPARQL User but several SPARQL Users and one > UPDATE user. For this we'll test blazegraph and other triplestores with > 1,2,4,8,16 SPARQL Users and 0 UPDATE Users, and 1,2,4,8,16 SPARQL Users and > 1 Update User. > The whole Benchmark will be tested with IGUANA [3]. > > If you're interest in the results. I can send them to you :) > > [1] http://aksw.org/Projects/FEASIBLE.html > [2] http://live.dbpedia.org/ > [3] https://github.com/AKSW/IGUANA > > Thanks alot > Felix > > > Am 13.10.2015 um 16:28 schrieb Brad Bebee: > > Felix, > > You will definitely want to adjust the properties a bit. Checkout [1] and > then generate a Vocabulary per the JIRA ticket [2]. That should improve > your performance substantially. I've also copied the blazegraph developers > list to see if anyone else has recommendations. > > Thanks, --Brad > > [1] > https://sourceforge.net/p/bigdata/git/ci/master/tree/bigdata-sails/src/samples/com/bigdata/samples/fastload.properties > > > [2] https://jira.blazegraph.com/browse/BLZG-1509 > > > > On Tue, Oct 13, 2015 at 10:19 AM, <mai...@st...> > wrote: > >> Hey >> >> thanks for the fast answer. >> The journal file looks like this: >> >> com.bigdata.journal.AbstractJournal.bufferMode=DiskRW >> com.bigdata.journal.AbstractJournal.file=bigdata.jnl >> >> Thanks >> Felix >> >> >> Zitat von Brad Bebee <be...@sy...>: >> >> Felix, >>> >>> Thank you. Can you send us some details of your journal configuration? >>> It may be beneficial to use a Vocabulary to inline the DBPedia values >>> for >>> improved load performance. >>> >>> Thanks, --Brad >>> >>> On Tue, Oct 13, 2015 at 10:10 AM, Blazegraph Web Site Contact < >>> bla...@bl...> wrote: >>> >>> From: Felix Conrads<mai...@st...> >>>> Subject: Benchmarking Blazegraph >>>> >>>> Message Body: >>>> >>>> Name : Felix Conrads >>>> Email : mai...@st... >>>> Message : Hello, >>>> >>>> I'm Felix Conrads from AKSW Research Group at University Leipzig. >>>> Currently i evaluate a Benchmark execution framework with several >>>> triplestores including Blazegrah. While blazegraph works fine it needs >>>> too >>>> much time to upload some of our datasets into it. >>>> These are the dbpedia datasets with ~217.000.000 triples and >>>> ~434.000.000 >>>> triples. >>>> I tried to upload the first one with the method described here: >>>> https://wiki.blazegraph.com/wiki/index.php/Bulk_Data_Load >>>> If you could tell me if there is any possiblity to upload them faster >>>> than >>>> >15 hours (i aborted it there) i would be very thankful . >>>> >>>> Thanks in advance and best regards >>>> Felix Conrads >>>> >>>> -- >>>> This e-mail was sent from a contact form on ( >>>> <http://www.blazegraph.com>http://www.blazegraph.com) >>>> >>>> -- >>>> You received this message because you are subscribed to the Google >>>> Groups >>>> "contact-us" group. >>>> To unsubscribe from this group and stop receiving emails from it, send >>>> an >>>> email to con...@sy.... >>>> To post to this group, send email to <con...@sy...> >>>> con...@sy.... >>>> To view this discussion on the web visit >>>> >>>> https://groups.google.com/a/systap.com/d/msgid/contact-us/8ece997cd271234401ab075d427430ca%40www.blazegraph.com >>>> . >>>> >>>> >>> >>> >>> -- >>> _______________ >>> Brad Bebee >>> CEO, Managing Partner >>> SYSTAP, LLC >>> e: be...@sy... >>> m: 202.642.7961 >>> f: 571.367.5000 >>> w: www.blazegraph.com >>> >>> Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance >>> graph >>> database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. >>> Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new >>> technology >>> to use GPUs to accelerate data-parallel graph analytics. >>> >>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are >>> for the sole use of the intended recipient(s) and are confidential or >>> proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, >>> dissemination or copying of this email or its contents or attachments is >>> prohibited. If you have received this communication in error, please >>> notify >>> the sender by reply email and permanently delete all copies of the email >>> and its contents and attachments. >>> >> >> >> >> -- >> You received this message because you are subscribed to the Google Groups >> "blazegraph" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to bla...@sy.... >> To post to this group, send email to <bla...@sy...> >> bla...@sy.... >> To view this discussion on the web visit >> <https://groups.google.com/a/systap.com/d/msgid/blazegraph/20151013161930.Horde.W_LJNqOM0w-ESzbfcFB7bw2%40mail.uni-leipzig.de> >> https://groups.google.com/a/systap.com/d/msgid/blazegraph/20151013161930.Horde.W_LJNqOM0w-ESzbfcFB7bw2%40mail.uni-leipzig.de >> . >> > > > > -- > _______________ > Brad Bebee > CEO, Managing Partner > SYSTAP, LLC > e: be...@sy... > m: 202.642.7961 > f: 571.367.5000 > w: <http://www.blazegraph.com>www.blazegraph.com > > Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance > graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints > APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new > technology to use GPUs to accelerate data-parallel graph analytics. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > > ------------------------------ > [image: Avast logo] <https://www.avast.com/antivirus> > > Diese E-Mail wurde von Avast Antivirus-Software auf Viren geprüft. > www.avast.com <https://www.avast.com/antivirus> > > -- _______________ Brad Bebee CEO, Managing Partner SYSTAP, LLC e: be...@sy... m: 202.642.7961 f: 571.367.5000 w: www.blazegraph.com Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new technology to use GPUs to accelerate data-parallel graph analytics. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |
From: Felix C. <mai...@st...> - 2015-11-16 08:56:13
|
Hey, First off thanks alot! It took a while to come back to it, as some other work showed up which needs to be done first. Nevertheless i have one more question: I took the fastload properties and generated a Vocabulary exactly as your JIRA Ticket [2]. How do i say blazegraph it should use my vocab? Is it as simple as to put in the properties following line: com.bigdata.rdf.store.AbstractTripleStore.vocabularyClass=MyCustomVocab (Generated File is MyCustomVocab.java in the directory with bigdata-1.5.2-bundled.jar) I'll test Blazegraph with a DBPedia Dataset and a SWDF Dataset with 250 Queries (ASK, SELECT, CONSTRUCT, DESCRIBE) generated with FEASIBLE [1] and at least DBPedia with Updates from DBpedia Live [2] (round about 250 inserts and deletes). The test should show how the triplestores act when they to not just only have 1 SPARQL User but several SPARQL Users and one UPDATE user. For this we'll test blazegraph and other triplestores with 1,2,4,8,16 SPARQL Users and 0 UPDATE Users, and 1,2,4,8,16 SPARQL Users and 1 Update User. The whole Benchmark will be tested with IGUANA [3]. If you're interest in the results. I can send them to you :) [1] http://aksw.org/Projects/FEASIBLE.html [2] http://live.dbpedia.org/ [3] https://github.com/AKSW/IGUANA Thanks alot Felix Am 13.10.2015 um 16:28 schrieb Brad Bebee: > Felix, > > You will definitely want to adjust the properties a bit. Checkout [1] > and then generate a Vocabulary per the JIRA ticket [2]. That should > improve your performance substantially. I've also copied the > blazegraph developers list to see if anyone else has recommendations. > > Thanks, --Brad > > [1] https://sourceforge.net/p/bigdata/git/ci/master/tree/bigdata-sails/src/samples/com/bigdata/samples/fastload.properties > > [2] https://jira.blazegraph.com/browse/BLZG-1509 > > > > On Tue, Oct 13, 2015 at 10:19 AM, <mai...@st... > <mailto:mai...@st...>> wrote: > > Hey > > thanks for the fast answer. > The journal file looks like this: > > com.bigdata.journal.AbstractJournal.bufferMode=DiskRW > com.bigdata.journal.AbstractJournal.file=bigdata.jnl > > Thanks > Felix > > > Zitat von Brad Bebee <be...@sy... <mailto:be...@sy...>>: > > Felix, > > Thank you. Can you send us some details of your journal > configuration? > It may be beneficial to use a Vocabulary to inline the > DBPedia values for > improved load performance. > > Thanks, --Brad > > On Tue, Oct 13, 2015 at 10:10 AM, Blazegraph Web Site Contact < > bla...@bl... <mailto:bla...@bl...>> > wrote: > > From: Felix Conrads<mai...@st... > <mailto:mai...@st...>> > Subject: Benchmarking Blazegraph > > Message Body: > > Name : Felix Conrads > Email : mai...@st... > <mailto:mai...@st...> > Message : Hello, > > I'm Felix Conrads from AKSW Research Group at University > Leipzig. > Currently i evaluate a Benchmark execution framework with > several > triplestores including Blazegrah. While blazegraph works > fine it needs too > much time to upload some of our datasets into it. > These are the dbpedia datasets with ~217.000.000 triples > and ~434.000.000 > triples. > I tried to upload the first one with the method described > here: > https://wiki.blazegraph.com/wiki/index.php/Bulk_Data_Load > If you could tell me if there is any possiblity to upload > them faster than > >15 hours (i aborted it there) i would be very thankful . > > Thanks in advance and best regards > Felix Conrads > > -- > This e-mail was sent from a contact form on > (http://www.blazegraph.com) > > -- > You received this message because you are subscribed to > the Google Groups > "contact-us" group. > To unsubscribe from this group and stop receiving emails > from it, send an > email to con...@sy... > <mailto:contact-us%2Bu...@sy...>. > To post to this group, send email to con...@sy... > <mailto:con...@sy...>. > To view this discussion on the web visit > https://groups.google.com/a/systap.com/d/msgid/contact-us/8ece997cd271234401ab075d427430ca%40www.blazegraph.com > . > > > > > -- > _______________ > Brad Bebee > CEO, Managing Partner > SYSTAP, LLC > e: be...@sy... <mailto:be...@sy...> > m: 202.642.7961 <tel:202.642.7961> > f: 571.367.5000 <tel:571.367.5000> > w: www.blazegraph.com <http://www.blazegraph.com> > > Blazegraph™ <http://www.blazegraph.com> is our ultra > high-performance graph > database that supports both RDF/SPARQL and > Tinkerpop/Blueprints APIs. > Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive > new technology > to use GPUs to accelerate data-parallel graph analytics. > > CONFIDENTIALITY NOTICE: This email and its contents and > attachments are > for the sole use of the intended recipient(s) and are > confidential or > proprietary to SYSTAP, LLC. Any unauthorized review, use, > disclosure, > dissemination or copying of this email or its contents or > attachments is > prohibited. If you have received this communication in error, > please notify > the sender by reply email and permanently delete all copies of > the email > and its contents and attachments. > > > > > -- > You received this message because you are subscribed to the Google > Groups "blazegraph" group. > To unsubscribe from this group and stop receiving emails from it, > send an email to bla...@sy... > <mailto:blazegraph%2Bu...@sy...>. > To post to this group, send email to bla...@sy... > <mailto:bla...@sy...>. > To view this discussion on the web visit > https://groups.google.com/a/systap.com/d/msgid/blazegraph/20151013161930.Horde.W_LJNqOM0w-ESzbfcFB7bw2%40mail.uni-leipzig.de. > > > > > -- > _______________ > Brad Bebee > CEO, Managing Partner > SYSTAP, LLC > e: be...@sy... <mailto:be...@sy...> > m: 202.642.7961 > f: 571.367.5000 > w: www.blazegraph.com <http://www.blazegraph.com> > > Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance > graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints > APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive > new technology to use GPUs to accelerate data-parallel graph analytics. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments > are for the sole use of the intended recipient(s) and are confidential > or proprietary to SYSTAP, LLC. Any unauthorized review, use, > disclosure, dissemination or copying of this email or its contents or > attachments is prohibited. If you have received this communication in > error, please notify the sender by reply email and permanently delete > all copies of the email and its contents and attachments. --- Diese E-Mail wurde von Avast Antivirus-Software auf Viren geprüft. https://www.avast.com/antivirus |
From: Michael S. <ms...@me...> - 2015-11-15 10:05:36
|
Dear Quentin, I’ve quickly set up you initial scenario with two triples in each namespace and was not able to reproduce the behavior, i.e. I always get one result there. Looking at your query, it is much more complex. In fact, ?profileURI might be bound to the same URI multiple times when evaluating the SERVICE, which could lead to duplicates (@Bryan: it might be that we’re missing out a distinct projection here? Though this would not explain why results vary from time to time.). Could you try running the query *without* the SERVICE call and confirm that the number of results is stable? Also, how many results do you get then? And do you know whether there a skos:prefLabel for each of them? (just trying to nail down what’s going on) Also, right at the end of the query plan in the EXPLAIN is a table showing runtime statistics (in particular, how many results were flowing through the operators). It would be quite useful to have that one too (both for the original query and the query without SERVICE). Could you share that as well, preferably as a screenshot. Regarding the option to run the SERVICE first: as you mentioned already, it does not help just inverting the order — the optimizer makes its own decision, independently from the order in which you write things. What you need there is a query hint that forces the optimizer to run the SERVICE keyword first (see triple pattern in the last line, to be placed right after the SERVICE): PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX skos: <http://www.w3.org/2004/02/skos/core#> SELECT ?doc ?concept ?label WHERE { ?doc dcterms:subject ?concept . SERVICE <http://localhost:9999/bigdata/namespace/skos/sparql> { ?concept skos:prefLabel ?label . } hint:Prior hint:runFirst "true" . } However, using that hint for your query would mean extracting all "?profileStatusURI skos:prefLabel ?profileStatusLabel” patterns from the remote endpoint first (without any restriction), so from a performance perspective this might not be the best option. But I’d be interested in the result of this query, i.e. whether you still get duplicates. Best, Michael > On 13 Nov 2015, at 23:24, Reul, Quentin <que...@wo...> wrote: > > Hi Bryan, > > I have checked that there were not any duplicate triples in either namespaces. > > For the query in the attached file, I should retrieve 21 records in my result set (which I do when DISTINCT is used). However, I retrieve 79 records in my result set when DISTINCT is not used. I have tried to use the SERVICE part of the query as the first part of the query, but I obtained the same results. > > Kind regards, > > Quentin Reul > > From: Bryan Thompson [mailto:br...@sy... <mailto:br...@sy...>] > Sent: Friday, November 13, 2015 4:01 PM > To: Reul, Quentin > Cc: big...@li... <mailto:big...@li...> > Subject: Re: [Bigdata-developers] Duplicate in SELECT queries when using SERVICE > > Quentin, > > Can you please look at the EXPLAIN of the query (if you are using the workbench, this is a checkbox under the advanced options, otherwise you can just add &explain to the query - see [1]). I am curious whether the triple pattern is running before or after the SERVICE call. > > Can you please confirm that you are observing that behavior on endpoints having just those two triples each? If there are many triples in the first endpoint, then one possibility is that the SERVICE call is being invoked more than one because multiple "chunks" of solutions are presented to it. In this case, the join could well have duplicates (same ?concept in different chunks leading to more than one solution with the same bindings from the B endpoint) and a DISTINCT would be required. > > Another option would be to run the SERVICE first. This could be accomplished with a query hint. See [2]. It is also possible to force certain operations to run exactly once. For example, but pushing something into a subquery and using the runOnce query hint. > > Another thing that you can try is to enable the solutions logger. This logger provides details on the inputs and outputs of each operator. You will be able to see the solutions produced by the triple pattern and those produced by the SERVICE call for each invocation of those operators. This will help you to understand whether multiple SERVICE invocations are occurring. > > ## > # Solutions trace (tab delimited file). Uncomment the next line to enable. > #log4j.logger.com.bigdata.bop.engine.SolutionsLog=INFO,solutionsLog > log4j.additivity.com.bigdata.bop.engine.SolutionsLog=false > log4j.appender.solutionsLog=org.apache.log4j.ConsoleAppender > #log4j.appender.solutionsLog=org.apache.log4j.FileAppender > log4j.appender.solutionsLog.Threshold=ALL > #log4j.appender.solutionsLog.File=solutions.csv > #log4j.appender.solutionsLog.Append=true > # I find that it is nicer to have this unbuffered since you can see what > # is going on and to make sure that I have complete rule evaluation logs > # on shutdown. > #log4j.appender.solutionsLog.BufferedIO=false > log4j.appender.solutionsLog.layout=org.apache.log4j.PatternLayout > log4j.appender.solutionsLog.layout.ConversionPattern=SOLUTION:\t%m > > If your investigations do not suggest an obvious solution, then it might be best if you create a ticket from this query and attach the EXPLAIN (which is an html page) to that ticket. Please add both myself and Michael Schmidt to the ticket as watchers so we will see any updates on the ticket. > > Thanks, > Bryan > > [1] https://wiki.blazegraph.com/wiki/index.php/Explain#NSS_Explain_Mode <https://wiki.blazegraph.com/wiki/index.php/Explain#NSS_Explain_Mode> > [2] https://wiki.blazegraph.com/wiki/index.php/QueryHints <https://wiki.blazegraph.com/wiki/index.php/QueryHints> > > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... <mailto:br...@sy...> > http://blazegraph.com <http://blazegraph.com/> > http://blog.blazegraph.com <http://blog.blazegraph.com/> > > Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > On Fri, Nov 13, 2015 at 11:35 AM, Reul, Quentin <que...@wo... <mailto:que...@wo...>> wrote: > Hi all, > > I'm encountering a weird behaviour when running SPARQL SELECT queries including SERVICE in the WHERE clause. More specifically, I seem to retrieve duplicate records in the result set when DISTINCT is not used. We are using RemoteRepositoryManager to access the BlazeGraph 1.5.2 instance and both namespaces are defined on the same machine. > > Let us imagine that I have the following triples in namespace A: > <doc1> dcterms:subject <conceptB> > <doc1> dcterms:title "Title of Document"^^xsd:string > > and some triples in namespace B: > <conceptB> rdf:type skos:Concept > <conceptB> skos:prefLabel "concept label"@en > > If I run the following SELECT query > SELECT ?doc ?concept ?label > WHERE { > ?doc dcterms:subject ?concept . > SERVICE <http:localhost:9999/bigdata/namespace/B/sparql> { > ?concept skos:prefLabel ?label . > } > } > then I would the following result set: > || ?doc || ?concept || ?label || > | <doc1> | <conceptB> | "concept label" | > | <doc1> | <conceptB> | "concept label" | > | <doc1> | <conceptB> | "concept label" | > > Interestingly, the number of duplicated records can change from processing to processing. Is this something that other people have encountered? > > Kind regards, > > Quentin Reul > > > ------------------------------------------------------------------------------ > > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> > > <sparql_query_explain.txt>------------------------------------------------------------------------------ > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> |
From: Bryan T. <br...@sy...> - 2015-11-14 02:03:16
|
Just indicates a full GC pause during IO. ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://blazegraph.com http://blog.blazegraph.com Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Fri, Nov 13, 2015 at 8:13 PM, Joakim Soderberg < joa...@bl...> wrote: > Is this warning something I should take serious and act upon? > > WARN : AbstractBTree.java:3716: wrote: name=kb.spo.OCSP, 10 records > (#nodes=1, #leaves=9) in 5283ms : addrRoot=-207112492768283427 > > > > ------------------------------------------------------------------------------ > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > |
From: Joakim S. <joa...@bl...> - 2015-11-14 01:13:18
|
Is this warning something I should take serious and act upon? WARN : AbstractBTree.java:3716: wrote: name=kb.spo.OCSP, 10 records (#nodes=1, #leaves=9) in 5283ms : addrRoot=-207112492768283427 |
From: Bryan T. <br...@sy...> - 2015-11-13 22:31:32
|
The query engine will make its own decision about when different parts of the query run. If you want something to run first, you need to indicate this explicitly with either a query hint or by disabling the query optimizer. You can not have duplicate triples in the database. If you are in quads mode, then there could be identical triples in different named graphs. But this is a separate question from how evaluation is occurring. I would be surprised if DISTINCT were necessary with just the two triples in each endpoint. But with many triples in the two endpoints, I think that it is necessary since the SERVICE is not otherwise guaranteed to be invoked exactly once and the same ?concept could be submitted into different invocations of the service, in which case duplicates would occur. To avoid duplicates, force the SERVICE call to run exactly once (which might slow down the time to the first solution for query if there are a lot of solutions for the first triple pattern), or use DISTINCT. Thanks, Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://blazegraph.com http://blog.blazegraph.com Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Fri, Nov 13, 2015 at 5:24 PM, Reul, Quentin < que...@wo...> wrote: > Hi Bryan, > > > > I have checked that there were not any duplicate triples in either > namespaces. > > > > For the query in the attached file, I should retrieve 21 records in my > result set (which I do when DISTINCT is used). However, I retrieve 79 > records in my result set when DISTINCT is not used. I have tried to use > the SERVICE part of the query as the first part of the query, but I > obtained the same results. > > > > Kind regards, > > > > *Quentin Reul* > > > > *From:* Bryan Thompson [mailto:br...@sy...] > *Sent:* Friday, November 13, 2015 4:01 PM > *To:* Reul, Quentin > *Cc:* big...@li... > *Subject:* Re: [Bigdata-developers] Duplicate in SELECT queries when > using SERVICE > > > > Quentin, > > > > Can you please look at the EXPLAIN of the query (if you are using the > workbench, this is a checkbox under the advanced options, otherwise you can > just add &explain to the query - see [1]). I am curious whether the triple > pattern is running before or after the SERVICE call. > > > > Can you please confirm that you are observing that behavior on endpoints > having just those two triples each? If there are many triples in the first > endpoint, then one possibility is that the SERVICE call is being invoked > more than one because multiple "chunks" of solutions are presented to it. > In this case, the join could well have duplicates (same ?concept in > different chunks leading to more than one solution with the same bindings > from the B endpoint) and a DISTINCT would be required. > > > > Another option would be to run the SERVICE first. This could be > accomplished with a query hint. See [2]. It is also possible to force > certain operations to run exactly once. For example, but pushing something > into a subquery and using the runOnce query hint. > > > > Another thing that you can try is to enable the solutions logger. This > logger provides details on the inputs and outputs of each operator. You > will be able to see the solutions produced by the triple pattern and those > produced by the SERVICE call for each invocation of those operators. This > will help you to understand whether multiple SERVICE invocations are > occurring. > > > > ## > > # Solutions trace (tab delimited file). Uncomment the next line to enable. > > #log4j.logger.com.bigdata.bop.engine.SolutionsLog=INFO,solutionsLog > > log4j.additivity.com.bigdata.bop.engine.SolutionsLog=false > > log4j.appender.solutionsLog=org.apache.log4j.ConsoleAppender > > #log4j.appender.solutionsLog=org.apache.log4j.FileAppender > > log4j.appender.solutionsLog.Threshold=ALL > > #log4j.appender.solutionsLog.File=solutions.csv > > #log4j.appender.solutionsLog.Append=true > > # I find that it is nicer to have this unbuffered since you can see what > > # is going on and to make sure that I have complete rule evaluation logs > > # on shutdown. > > #log4j.appender.solutionsLog.BufferedIO=false > > log4j.appender.solutionsLog.layout=org.apache.log4j.PatternLayout > > log4j.appender.solutionsLog.layout.ConversionPattern=SOLUTION:\t%m > > > > If your investigations do not suggest an obvious solution, then it might > be best if you create a ticket from this query and attach the EXPLAIN > (which is an html page) to that ticket. Please add both myself and Michael > Schmidt to the ticket as watchers so we will see any updates on the ticket. > > > > Thanks, > > Bryan > > > > [1] https://wiki.blazegraph.com/wiki/index.php/Explain#NSS_Explain_Mode > > [2] https://wiki.blazegraph.com/wiki/index.php/QueryHints > > > ---- > Bryan Thompson > > Chief Scientist & Founder > SYSTAP, LLC > > 4501 Tower Road > Greensboro, NC 27410 > > br...@sy... > > http://blazegraph.com > > http://blog.blazegraph.com > > > > Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance > graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints > APIs. Blazegraph is now available with GPU acceleration using > our disruptive technology to accelerate data-parallel graph analytics and > graph query. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > > On Fri, Nov 13, 2015 at 11:35 AM, Reul, Quentin < > que...@wo...> wrote: > > Hi all, > > > > I'm encountering a weird behaviour when running SPARQL SELECT queries > including SERVICE in the WHERE clause. More specifically, I seem to > retrieve duplicate records in the result set when DISTINCT is not used. > We are using RemoteRepositoryManager to access the BlazeGraph 1.5.2 > instance and both namespaces are defined on the same machine. > > > > Let us imagine that I have the following triples in namespace A: > > <doc1> dcterms:subject <conceptB> > > <doc1> dcterms:title "Title of Document"^^xsd:string > > > > and some triples in namespace B: > > <conceptB> rdf:type skos:Concept > > <conceptB> skos:prefLabel "concept label"@en > > > > If I run the following SELECT query > > SELECT ?doc ?concept ?label > > WHERE { > > ?doc dcterms:subject ?concept . > > SERVICE <http:localhost:9999/bigdata/namespace/B/sparql> { > > ?concept skos:prefLabel ?label . > > } > > } > > then I would the following result set: > > || ?doc || ?concept || ?label || > > | <doc1> | <conceptB> | "concept label" | > > | <doc1> | <conceptB> | "concept label" | > > | <doc1> | <conceptB> | "concept label" | > > > > Interestingly, the number of duplicated records can change from processing > to processing. Is this something that other people have encountered? > > > > Kind regards, > > > > *Quentin Reul * > > > > > > ------------------------------------------------------------------------------ > > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > > > |
From: Reul, Q. <que...@wo...> - 2015-11-13 22:24:19
|
Query SPARQL PREFIX pci-core: <http://onto.wolterskluwer.com/pci/core/> PREFIX pci-cpe: <http://onto.wolterskluwer.com/pci/cpe/> PREFIX skos: <http://www.w3.org/2004/02/skos/core#> PREFIX dcterms: <http://purl.org/dc/terms/> SELECT ?profileURI ?profileID ?profileLabel ?profileType ?profileNotation ?profileDate ?profileStatusURI ?profileStatusLabel WHERE { ?profileURI a pci-cpe:Profile . ?profileURI dcterms:identifier ?profileID . ?profileURI skos:prefLabel ?profileLabel . ?profileURI dc:type ?profileType . ?profileURI skos:notation ?profileNotation . ?profileURI dcterms:date ?profileDate . ?profileURI pci-core:hasStatus ?profileStatusURI . SERVICE <http://10.0.1.69:9999/bigdata/namespace/schema/sparql> { ?profileStatusURI skos:prefLabel ?profileStatusLabel . } } Parse Tree QueryContainer PrefixDecl (prefix=pci-core) IRI (http://onto.wolterskluwer.com/pci/core/) PrefixDecl (prefix=pci-cpe) IRI (http://onto.wolterskluwer.com/pci/cpe/) PrefixDecl (prefix=skos) IRI (http://www.w3.org/2004/02/skos/core#) PrefixDecl (prefix=dcterms) IRI (http://purl.org/dc/terms/) SelectQuery Select ProjectionElem Var (profileURI) ProjectionElem Var (profileID) ProjectionElem Var (profileLabel) ProjectionElem Var (profileType) ProjectionElem Var (profileNotation) ProjectionElem Var (profileDate) ProjectionElem Var (profileStatusURI) ProjectionElem Var (profileStatusLabel) WhereClause GraphPatternGroup BasicGraphPattern TriplesSameSubjectPath Var (profileURI) PropertyListPath PathAlternative PathSequence PathElt IRI (http://www.w3.org/1999/02/22-rdf-syntax-ns#type) ObjectList IRI (http://onto.wolterskluwer.com/pci/cpe/Profile) TriplesSameSubjectPath Var (profileURI) PropertyListPath PathAlternative PathSequence PathElt IRI (http://purl.org/dc/terms/identifier) ObjectList Var (profileID) TriplesSameSubjectPath Var (profileURI) PropertyListPath PathAlternative PathSequence PathElt IRI (http://www.w3.org/2004/02/skos/core#prefLabel) ObjectList Var (profileLabel) TriplesSameSubjectPath Var (profileURI) PropertyListPath PathAlternative PathSequence PathElt IRI (http://purl.org/dc/elements/1.1/type) ObjectList Var (profileType) TriplesSameSubjectPath Var (profileURI) PropertyListPath PathAlternative PathSequence PathElt IRI (http://www.w3.org/2004/02/skos/core#notation) ObjectList Var (profileNotation) TriplesSameSubjectPath Var (profileURI) PropertyListPath PathAlternative PathSequence PathElt IRI (http://purl.org/dc/terms/date) ObjectList Var (profileDate) TriplesSameSubjectPath Var (profileURI) PropertyListPath PathAlternative PathSequence PathElt IRI (http://onto.wolterskluwer.com/pci/core/hasStatus) ObjectList Var (profileStatusURI) ServiceGraphPattern IRI (http://10.0.1.69:9999/bigdata/namespace/schema/sparql) GraphPatternGroup BasicGraphPattern TriplesSameSubjectPath Var (profileStatusURI) PropertyListPath PathAlternative PathSequence PathElt IRI (http://www.w3.org/2004/02/skos/core#prefLabel) ObjectList Var (profileStatusLabel) Original AST PREFIX pci-core: <http://onto.wolterskluwer.com/pci/core/> PREFIX pci-cpe: <http://onto.wolterskluwer.com/pci/cpe/> PREFIX skos: <http://www.w3.org/2004/02/skos/core#> PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> PREFIX sesame: <http://www.openrdf.org/schema/sesame#> PREFIX owl: <http://www.w3.org/2002/07/owl#> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX fn: <http://www.w3.org/2005/xpath-functions#> PREFIX foaf: <http://xmlns.com/foaf/0.1/> PREFIX dc: <http://purl.org/dc/elements/1.1/> PREFIX hint: <http://www.bigdata.com/queryHints#> PREFIX bd: <http://www.bigdata.com/rdf#> PREFIX bds: <http://www.bigdata.com/rdf/search#> QueryType: SELECT includeInferred=true SELECT VarNode(profileURI) VarNode(profileID) VarNode(profileLabel) VarNode(profileType) VarNode(profileNotation) VarNode(profileDate) VarNode(profileStatusURI) VarNode(profileStatusLabel) JoinGroupNode { StatementPatternNode(VarNode(profileURI), ConstantNode(Vocab(14)[http://www.w3.org/1999/02/22-rdf-syntax-ns#type]), ConstantNode(TermId(28003U)[http://onto.wolterskluwer.com/pci/cpe/Profile])) [scope=DEFAULT_CONTEXTS] StatementPatternNode(VarNode(profileURI), ConstantNode(TermId(28006U)[http://purl.org/dc/terms/identifier]), VarNode(profileID)) [scope=DEFAULT_CONTEXTS] StatementPatternNode(VarNode(profileURI), ConstantNode(Vocab(-90)[http://www.w3.org/2004/02/skos/core#prefLabel]), VarNode(profileLabel)) [scope=DEFAULT_CONTEXTS] StatementPatternNode(VarNode(profileURI), ConstantNode(Vocab(-66)[http://purl.org/dc/elements/1.1/type]), VarNode(profileType)) [scope=DEFAULT_CONTEXTS] StatementPatternNode(VarNode(profileURI), ConstantNode(Vocab(-92)[http://www.w3.org/2004/02/skos/core#notation]), VarNode(profileNotation)) [scope=DEFAULT_CONTEXTS] StatementPatternNode(VarNode(profileURI), ConstantNode(TermId(28005U)[http://purl.org/dc/terms/date]), VarNode(profileDate)) [scope=DEFAULT_CONTEXTS] StatementPatternNode(VarNode(profileURI), ConstantNode(TermId(28001U)[http://onto.wolterskluwer.com/pci/core/hasStatus]), VarNode(profileStatusURI)) [scope=DEFAULT_CONTEXTS] SERVICE <ConstantNode(TermId(0U)[http://10.0.1.69:9999/bigdata/namespace/schema/sparql])> { JoinGroupNode { StatementPatternNode(VarNode(profileStatusURI), ConstantNode(Vocab(-90)[http://www.w3.org/2004/02/skos/core#prefLabel]), VarNode(profileStatusLabel)) [scope=DEFAULT_CONTEXTS] } } } Static Analysis Statistics object category elapsed [ms] elapsed [us] numCalls ParseTime Parser 1 1035 1 Optimizers Optimizers (all) 5 5443 1 ASTQueryHintOptimizer Optimizer 0 184 1 ASTFilterNormalizationOptimizer Optimizer 0 2 1 ASTStaticBindingsOptimizer Optimizer 0 44 1 ASTSearchInSearchOptimizer Optimizer 0 3 1 ASTPropertyPathOptimizer Optimizer 0 2 1 ASTSetValueExpressionsOptimizer Optimizer 0 168 1 ASTFlattenUnionsOptimizer Optimizer 0 2 1 ASTUnionFiltersOptimizer Optimizer 0 2 1 ASTEmptyGroupOptimizer Optimizer 0 3 1 ASTWildcardProjectionOptimizer Optimizer 0 99 1 ASTSearchOptimizer Optimizer 0 10 1 ASTFulltextSearchOptimizer Optimizer 0 7 1 AskOptimizer Optimizer 0 1 1 ASTDescribeOptimizer Optimizer 0 1 1 ASTConstructOptimizer Optimizer 0 1 1 ASTExistsOptimizer Optimizer 0 2 1 ASTGraphGroupOptimizer Optimizer 0 30 1 ASTLiftPreFiltersOptimizer Optimizer 0 0 1 ASTBottomUpOptimizer Optimizer 0 278 1 ASTSimpleOptionalOptimizer Optimizer 0 5 1 ASTFlattenJoinGroupsOptimizer Optimizer 0 2 1 ASTALPServiceOptimizer Optimizer 0 9 1 ASTServiceNodeOptimizer Optimizer 0 4 1 ASTJoinGroupOrderOptimizer Optimizer 0 225 2 ASTRunFirstRunLastOptimizer Optimizer 0 3 1 ASTRangeOptimizer Optimizer 0 3 1 ASTRangeCountOptimizer Optimizer 4 4128 1 ASTCardinalityOptimizer Optimizer 0 9 1 ASTFastRangeCountOptimizer Optimizer 0 16 1 ASTSimpleGroupByAndCountOptimizer Optimizer 0 2 1 ASTDistinctTermScanOptimizer Optimizer 0 3 1 ASTStaticJoinOptimizer Optimizer 0 65 1 ASTAttachJoinFiltersOptimizer Optimizer 0 27 1 ASTSparql11SubqueryOptimizer Optimizer 0 1 1 ASTNamedSubqueryOptimizer Optimizer 0 0 1 ASTSubGroupJoinVarOptimizer Optimizer 0 1 1 RangeCount Range Count 0 49 7 Optimized AST PREFIX pci-core: <http://onto.wolterskluwer.com/pci/core/> PREFIX pci-cpe: <http://onto.wolterskluwer.com/pci/cpe/> PREFIX skos: <http://www.w3.org/2004/02/skos/core#> PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> PREFIX sesame: <http://www.openrdf.org/schema/sesame#> PREFIX owl: <http://www.w3.org/2002/07/owl#> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX fn: <http://www.w3.org/2005/xpath-functions#> PREFIX foaf: <http://xmlns.com/foaf/0.1/> PREFIX dc: <http://purl.org/dc/elements/1.1/> PREFIX hint: <http://www.bigdata.com/queryHints#> PREFIX bd: <http://www.bigdata.com/rdf#> PREFIX bds: <http://www.bigdata.com/rdf/search#> QueryType: SELECT includeInferred=true SELECT ( VarNode(profileURI) AS VarNode(profileURI) ) ( VarNode(profileID) AS VarNode(profileID) ) ( VarNode(profileLabel) AS VarNode(profileLabel) ) ( VarNode(profileType) AS VarNode(profileType) ) ( VarNode(profileNotation) AS VarNode(profileNotation) ) ( VarNode(profileDate) AS VarNode(profileDate) ) ( VarNode(profileStatusURI) AS VarNode(profileStatusURI) ) ( VarNode(profileStatusLabel) AS VarNode(profileStatusLabel) ) JoinGroupNode { StatementPatternNode(VarNode(profileURI), ConstantNode(TermId(28001U)[http://onto.wolterskluwer.com/pci/core/hasStatus]), VarNode(profileStatusURI)) [scope=DEFAULT_CONTEXTS] AST2BOpBase.estimatedCardinality=2897 AST2BOpBase.originalIndex=POS StatementPatternNode(VarNode(profileURI), ConstantNode(Vocab(14)[http://www.w3.org/1999/02/22-rdf-syntax-ns#type]), ConstantNode(TermId(28003U)[http://onto.wolterskluwer.com/pci/cpe/Profile])) [scope=DEFAULT_CONTEXTS] AST2BOpBase.estimatedCardinality=21 AST2BOpBase.originalIndex=POS StatementPatternNode(VarNode(profileURI), ConstantNode(Vocab(-66)[http://purl.org/dc/elements/1.1/type]), VarNode(profileType)) [scope=DEFAULT_CONTEXTS] AST2BOpBase.estimatedCardinality=21 AST2BOpBase.originalIndex=POS StatementPatternNode(VarNode(profileURI), ConstantNode(TermId(28005U)[http://purl.org/dc/terms/date]), VarNode(profileDate)) [scope=DEFAULT_CONTEXTS] AST2BOpBase.estimatedCardinality=21 AST2BOpBase.originalIndex=POS StatementPatternNode(VarNode(profileURI), ConstantNode(Vocab(-92)[http://www.w3.org/2004/02/skos/core#notation]), VarNode(profileNotation)) [scope=DEFAULT_CONTEXTS] AST2BOpBase.estimatedCardinality=2308 AST2BOpBase.originalIndex=POS StatementPatternNode(VarNode(profileURI), ConstantNode(TermId(28006U)[http://purl.org/dc/terms/identifier]), VarNode(profileID)) [scope=DEFAULT_CONTEXTS] AST2BOpBase.estimatedCardinality=2897 AST2BOpBase.originalIndex=POS StatementPatternNode(VarNode(profileURI), ConstantNode(Vocab(-90)[http://www.w3.org/2004/02/skos/core#prefLabel]), VarNode(profileLabel)) [scope=DEFAULT_CONTEXTS] AST2BOpBase.estimatedCardinality=2897 AST2BOpBase.originalIndex=POS SERVICE <ConstantNode(TermId(0U)[http://10.0.1.69:9999/bigdata/namespace/schema/sparql])> { JoinGroupNode { StatementPatternNode(VarNode(profileStatusURI), ConstantNode(Vocab(-90)[http://www.w3.org/2004/02/skos/core#prefLabel]), VarNode(profileStatusLabel)) [scope=DEFAULT_CONTEXTS] } } } with static (exogeneous) bindings defined as follows: { { } } Query Plan com.bigdata.bop.solutions.ProjectionOp[17](ServiceCallJoin[15])[ BOp.bopId=17, BOp.evaluationContext=CONTROLLER, PipelineOp.sharedState=true, JoinAnnotations.select=[profileURI, profileID, profileLabel, profileType, profileNotation, profileDate, profileStatusURI, profileStatusLabel], QueryEngine.queryId=7ce14b6f-21da-4c7b-9217-544174212639] com.bigdata.bop.controller.ServiceCallJoin[15](ChunkedMaterializationOp[16])[ BOp.bopId=15, BOp.evaluationContext=CONTROLLER, PipelineOp.pipelined=false, PipelineOp.sharedState=true, ServiceCallJoin.serviceNode= SERVICE <ConstantNode(TermId(0U)[http://10.0.1.69:9999/bigdata/namespace/schema/sparql])> { JoinGroupNode { StatementPatternNode(VarNode(profileStatusURI), ConstantNode(Vocab(-90)[http://www.w3.org/2004/02/skos/core#prefLabel]), VarNode(profileStatusLabel)) [scope=DEFAULT_CONTEXTS] } }, ServiceCallJoin.namespace=profile-manager-data, ServiceCallJoin.timestamp=1447452827601, HashJoinAnnotations.joinVars=[profileStatusURI], JoinAnnotations.constraints=null] com.bigdata.bop.rdf.join.ChunkedMaterializationOp[16](PipelineJoin[14])[ ChunkedMaterializationOp.vars=[profileStatusURI], IPredicate.relationName=[profile-manager-data.lex], IPredicate.timestamp=1447452827601, ChunkedMaterializationOp.materializeAll=false, PipelineOp.sharedState=true, PipelineOp.reorderSolutions=true, PipelineOp.maxParallel=5, BOp.bopId=16] com.bigdata.bop.join.PipelineJoin[14](PipelineJoin[12])[ BOp.bopId=14, JoinAnnotations.constraints=null, AST2BOpBase.simpleJoin=true, BOp.evaluationContext=ANY, AccessPathJoinAnnotations.predicate=com.bigdata.rdf.spo.SPOPredicate[13](profileURI=null, Vocab(-90)[http://www.w3.org/2004/02/skos/core#prefLabel], profileLabel=null)[ IPredicate.relationName=[profile-manager-data.spo], IPredicate.timestamp=1447452827601, BOp.bopId=13, AST2BOpBase.estimatedCardinality=2897, AST2BOpBase.originalIndex=POS, IPredicate.flags=[KEYS,VALS,READONLY,PARALLEL]]] com.bigdata.bop.join.PipelineJoin[12](PipelineJoin[10])[ BOp.bopId=12, JoinAnnotations.constraints=null, AST2BOpBase.simpleJoin=true, BOp.evaluationContext=ANY, AccessPathJoinAnnotations.predicate=com.bigdata.rdf.spo.SPOPredicate[11](profileURI=null, TermId(28006U)[http://purl.org/dc/terms/identifier], profileID=null)[ IPredicate.relationName=[profile-manager-data.spo], IPredicate.timestamp=1447452827601, BOp.bopId=11, AST2BOpBase.estimatedCardinality=2897, AST2BOpBase.originalIndex=POS, IPredicate.flags=[KEYS,VALS,READONLY,PARALLEL]]] com.bigdata.bop.join.PipelineJoin[10](PipelineJoin[8])[ BOp.bopId=10, JoinAnnotations.constraints=null, AST2BOpBase.simpleJoin=true, BOp.evaluationContext=ANY, AccessPathJoinAnnotations.predicate=com.bigdata.rdf.spo.SPOPredicate[9](profileURI=null, Vocab(-92)[http://www.w3.org/2004/02/skos/core#notation], profileNotation=null)[ IPredicate.relationName=[profile-manager-data.spo], IPredicate.timestamp=1447452827601, BOp.bopId=9, AST2BOpBase.estimatedCardinality=2308, AST2BOpBase.originalIndex=POS, IPredicate.flags=[KEYS,VALS,READONLY,PARALLEL]]] com.bigdata.bop.join.PipelineJoin[8](PipelineJoin[6])[ BOp.bopId=8, JoinAnnotations.constraints=null, AST2BOpBase.simpleJoin=true, BOp.evaluationContext=ANY, AccessPathJoinAnnotations.predicate=com.bigdata.rdf.spo.SPOPredicate[7](profileURI=null, TermId(28005U)[http://purl.org/dc/terms/date], profileDate=null)[ IPredicate.relationName=[profile-manager-data.spo], IPredicate.timestamp=1447452827601, BOp.bopId=7, AST2BOpBase.estimatedCardinality=21, AST2BOpBase.originalIndex=POS, IPredicate.flags=[KEYS,VALS,READONLY,PARALLEL]]] com.bigdata.bop.join.PipelineJoin[6](PipelineJoin[4])[ BOp.bopId=6, JoinAnnotations.constraints=null, AST2BOpBase.simpleJoin=true, BOp.evaluationContext=ANY, AccessPathJoinAnnotations.predicate=com.bigdata.rdf.spo.SPOPredicate[5](profileURI=null, Vocab(-66)[http://purl.org/dc/elements/1.1/type], profileType=null)[ IPredicate.relationName=[profile-manager-data.spo], IPredicate.timestamp=1447452827601, BOp.bopId=5, AST2BOpBase.estimatedCardinality=21, AST2BOpBase.originalIndex=POS, IPredicate.flags=[KEYS,VALS,READONLY,PARALLEL]]] com.bigdata.bop.join.PipelineJoin[4](PipelineJoin[2])[ BOp.bopId=4, JoinAnnotations.constraints=null, AST2BOpBase.simpleJoin=true, BOp.evaluationContext=ANY, AccessPathJoinAnnotations.predicate=com.bigdata.rdf.spo.SPOPredicate[3](profileURI=null, Vocab(14)[http://www.w3.org/1999/02/22-rdf-syntax-ns#type], TermId(28003U)[http://onto.wolterskluwer.com/pci/cpe/Profile])[ IPredicate.relationName=[profile-manager-data.spo], IPredicate.timestamp=1447452827601, BOp.bopId=3, AST2BOpBase.estimatedCardinality=21, AST2BOpBase.originalIndex=POS, IPredicate.flags=[KEYS,VALS,READONLY,PARALLEL]]] com.bigdata.bop.join.PipelineJoin[2]()[ BOp.bopId=2, JoinAnnotations.constraints=null, AST2BOpBase.simpleJoin=true, BOp.evaluationContext=ANY, AccessPathJoinAnnotations.predicate=com.bigdata.rdf.spo.SPOPredicate[1](profileURI=null, TermId(28001U)[http://onto.wolterskluwer.com/pci/core/hasStatus], profileStatusURI=null)[ IPredicate.relationName=[profile-manager-data.spo], IPredicate.timestamp=1447452827601, BOp.bopId=1, AST2BOpBase.estimatedCardinality=2897, AST2BOpBase.originalIndex=POS, IPredicate.flags=[KEYS,VALS,READONLY,PARALLEL]]] |
From: Bryan T. <br...@sy...> - 2015-11-13 22:17:24
|
Only a single process may have the Journal open at a time. An file overlap exception is expected if multiple processes are attempting to open the journal. The journal is thread safe, but must be open for only a single process. The thread safety requires an awareness of both readers and writers with concurrent access to the journal. Thanks, Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://blazegraph.com http://blog.blazegraph.com Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Fri, Nov 13, 2015 at 5:10 PM, Joakim Soderberg < joa...@bl...> wrote: > The OS is CentOS Linux release 7.1.1503 (Core) > I am batch loading triples and it is possible that there are several > processes running > > > On Nov 13, 2015, at 11:41 AM, Bryan Thompson <br...@sy...> wrote: > > What OS? Are you trying to open the journal in two separate processes > (this is the most common cause). > > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... > http://blazegraph.com > http://blog.blazegraph.com > > Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance > graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints > APIs. Blazegraph is now available with GPU acceleration using our disruptive > technology to accelerate data-parallel graph analytics and graph query. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > On Fri, Nov 13, 2015 at 2:26 PM, Joakim Soderberg < > joa...@bl...> wrote: > >> Hi, >> What did most likely go wrong if I get the following error: >> >> *exception* >> >> javax.servlet.ServletException: Servlet.init() for servlet com.blippar.servlet.SparqlServlet threw exception >> org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) >> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) >> org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) >> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) >> org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) >> org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) >> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) >> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >> org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) >> java.lang.Thread.run(Thread.java:745) >> >> *root cause* >> >> java.lang.RuntimeException: FileLock Overlap >> com.bigdata.journal.FileMetadata.reopenChannel(FileMetadata.java:1245) >> com.bigdata.journal.FileMetadata.access$000(FileMetadata.java:58) >> com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1163) >> com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1153) >> com.bigdata.journal.FileMetadata.<init>(FileMetadata.java:946) >> com.bigdata.journal.FileMetadata.createInstance(FileMetadata.java:1470) >> com.bigdata.journal.AbstractJournal.<init>(AbstractJournal.java:1156) >> com.bigdata.journal.Journal.<init>(Journal.java:275) >> com.bigdata.journal.Journal.<init>(Journal.java:268) >> com.bigdata.rdf.sail.BigdataSail.createLTS(BigdataSail.java:710) >> com.bigdata.rdf.sail.BigdataSail.<init>(BigdataSail.java:689) >> com.blippar.servlet.DataManager.initialize(DataManager.java:74) >> com.blippar.servlet.SparqlServlet.init(SparqlServlet.java:42) >> javax.servlet.GenericServlet.init(GenericServlet.java:158) >> org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) >> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) >> org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) >> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) >> org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) >> org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) >> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) >> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >> org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) >> java.lang.Thread.run(Thread.java:745) >> >> *root cause* >> >> java.nio.channels.OverlappingFileLockException >> sun.nio.ch.SharedFileLockTable.checkList(FileLockTable.java:255) >> sun.nio.ch.SharedFileLockTable.add(FileLockTable.java:152) >> sun.nio.ch.FileChannelImpl.tryLock(FileChannelImpl.java:1075) >> com.bigdata.journal.FileMetadata.reopenChannel(FileMetadata.java:1210) >> com.bigdata.journal.FileMetadata.access$000(FileMetadata.java:58) >> com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1163) >> com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1153) >> com.bigdata.journal.FileMetadata.<init>(FileMetadata.java:946) >> com.bigdata.journal.FileMetadata.createInstance(FileMetadata.java:1470) >> com.bigdata.journal.AbstractJournal.<init>(AbstractJournal.java:1156) >> com.bigdata.journal.Journal.<init>(Journal.java:275) >> com.bigdata.journal.Journal.<init>(Journal.java:268) >> com.bigdata.rdf.sail.BigdataSail.createLTS(BigdataSail.java:710) >> com.bigdata.rdf.sail.BigdataSail.<init>(BigdataSail.java:689) >> com.blippar.servlet.DataManager.initialize(DataManager.java:74) >> com.blippar.servlet.SparqlServlet.init(SparqlServlet.java:42) >> javax.servlet.GenericServlet.init(GenericServlet.java:158) >> org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) >> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) >> org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) >> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) >> org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) >> org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) >> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) >> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >> org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) >> java.lang.Thread.run(Thread.java:745) >> >> >> >> >> ------------------------------------------------------------------------------ >> >> _______________________________________________ >> Bigdata-developers mailing list >> Big...@li... >> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >> >> > > |
From: Joakim S. <joa...@bl...> - 2015-11-13 22:10:11
|
The OS is CentOS Linux release 7.1.1503 (Core) I am batch loading triples and it is possible that there are several processes running > On Nov 13, 2015, at 11:41 AM, Bryan Thompson <br...@sy...> wrote: > > What OS? Are you trying to open the journal in two separate processes (this is the most common cause). > > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... <mailto:br...@sy...> > http://blazegraph.com <http://blazegraph.com/> > http://blog.blazegraph.com <http://blog.blazegraph.com/> > > Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > > On Fri, Nov 13, 2015 at 2:26 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: > Hi, > What did most likely go wrong if I get the following error: > > exception > > javax.servlet.ServletException: Servlet.init() for servlet com.blippar.servlet.SparqlServlet threw exception > org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) > org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) > org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) > org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) > org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) > org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) > java.lang.Thread.run(Thread.java:745) > root cause > > java.lang.RuntimeException: FileLock Overlap > com.bigdata.journal.FileMetadata.reopenChannel(FileMetadata.java:1245) > com.bigdata.journal.FileMetadata.access$000(FileMetadata.java:58) > com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1163) > com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1153) > com.bigdata.journal.FileMetadata.<init>(FileMetadata.java:946) > com.bigdata.journal.FileMetadata.createInstance(FileMetadata.java:1470) > com.bigdata.journal.AbstractJournal.<init>(AbstractJournal.java:1156) > com.bigdata.journal.Journal.<init>(Journal.java:275) > com.bigdata.journal.Journal.<init>(Journal.java:268) > com.bigdata.rdf.sail.BigdataSail.createLTS(BigdataSail.java:710) > com.bigdata.rdf.sail.BigdataSail.<init>(BigdataSail.java:689) > com.blippar.servlet.DataManager.initialize(DataManager.java:74) > com.blippar.servlet.SparqlServlet.init(SparqlServlet.java:42) > javax.servlet.GenericServlet.init(GenericServlet.java:158) > org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) > org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) > org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) > org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) > org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) > org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) > java.lang.Thread.run(Thread.java:745) > root cause > > java.nio.channels.OverlappingFileLockException > sun.nio.ch.SharedFileLockTable.checkList(FileLockTable.java:255) > sun.nio.ch.SharedFileLockTable.add(FileLockTable.java:152) > sun.nio.ch.FileChannelImpl.tryLock(FileChannelImpl.java:1075) > com.bigdata.journal.FileMetadata.reopenChannel(FileMetadata.java:1210) > com.bigdata.journal.FileMetadata.access$000(FileMetadata.java:58) > com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1163) > com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1153) > com.bigdata.journal.FileMetadata.<init>(FileMetadata.java:946) > com.bigdata.journal.FileMetadata.createInstance(FileMetadata.java:1470) > com.bigdata.journal.AbstractJournal.<init>(AbstractJournal.java:1156) > com.bigdata.journal.Journal.<init>(Journal.java:275) > com.bigdata.journal.Journal.<init>(Journal.java:268) > com.bigdata.rdf.sail.BigdataSail.createLTS(BigdataSail.java:710) > com.bigdata.rdf.sail.BigdataSail.<init>(BigdataSail.java:689) > com.blippar.servlet.DataManager.initialize(DataManager.java:74) > com.blippar.servlet.SparqlServlet.init(SparqlServlet.java:42) > javax.servlet.GenericServlet.init(GenericServlet.java:158) > org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) > org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) > org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) > org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) > org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) > org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) > java.lang.Thread.run(Thread.java:745) > > > ------------------------------------------------------------------------------ > > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> > > |
From: Bryan T. <br...@sy...> - 2015-11-13 22:01:36
|
Quentin, Can you please look at the EXPLAIN of the query (if you are using the workbench, this is a checkbox under the advanced options, otherwise you can just add &explain to the query - see [1]). I am curious whether the triple pattern is running before or after the SERVICE call. Can you please confirm that you are observing that behavior on endpoints having just those two triples each? If there are many triples in the first endpoint, then one possibility is that the SERVICE call is being invoked more than one because multiple "chunks" of solutions are presented to it. In this case, the join could well have duplicates (same ?concept in different chunks leading to more than one solution with the same bindings from the B endpoint) and a DISTINCT would be required. Another option would be to run the SERVICE first. This could be accomplished with a query hint. See [2]. It is also possible to force certain operations to run exactly once. For example, but pushing something into a subquery and using the runOnce query hint. Another thing that you can try is to enable the solutions logger. This logger provides details on the inputs and outputs of each operator. You will be able to see the solutions produced by the triple pattern and those produced by the SERVICE call for each invocation of those operators. This will help you to understand whether multiple SERVICE invocations are occurring. ## # Solutions trace (tab delimited file). Uncomment the next line to enable. #log4j.logger.com.bigdata.bop.engine.SolutionsLog=INFO,solutionsLog log4j.additivity.com.bigdata.bop.engine.SolutionsLog=false log4j.appender.solutionsLog=org.apache.log4j.ConsoleAppender #log4j.appender.solutionsLog=org.apache.log4j.FileAppender log4j.appender.solutionsLog.Threshold=ALL #log4j.appender.solutionsLog.File=solutions.csv #log4j.appender.solutionsLog.Append=true # I find that it is nicer to have this unbuffered since you can see what # is going on and to make sure that I have complete rule evaluation logs # on shutdown. #log4j.appender.solutionsLog.BufferedIO=false log4j.appender.solutionsLog.layout=org.apache.log4j.PatternLayout log4j.appender.solutionsLog.layout.ConversionPattern=SOLUTION:\t%m If your investigations do not suggest an obvious solution, then it might be best if you create a ticket from this query and attach the EXPLAIN (which is an html page) to that ticket. Please add both myself and Michael Schmidt to the ticket as watchers so we will see any updates on the ticket. Thanks, Bryan [1] https://wiki.blazegraph.com/wiki/index.php/Explain#NSS_Explain_Mode [2] https://wiki.blazegraph.com/wiki/index.php/QueryHints ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://blazegraph.com http://blog.blazegraph.com Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Fri, Nov 13, 2015 at 11:35 AM, Reul, Quentin < que...@wo...> wrote: > Hi all, > > > > I'm encountering a weird behaviour when running SPARQL SELECT queries > including SERVICE in the WHERE clause. More specifically, I seem to > retrieve duplicate records in the result set when DISTINCT is not used. > We are using RemoteRepositoryManager to access the BlazeGraph 1.5.2 > instance and both namespaces are defined on the same machine. > > > > Let us imagine that I have the following triples in namespace A: > > <doc1> dcterms:subject <conceptB> > > <doc1> dcterms:title "Title of Document"^^xsd:string > > > > and some triples in namespace B: > > <conceptB> rdf:type skos:Concept > > <conceptB> skos:prefLabel "concept label"@en > > > > If I run the following SELECT query > > SELECT ?doc ?concept ?label > > WHERE { > > ?doc dcterms:subject ?concept . > > SERVICE <http:localhost:9999/bigdata/namespace/B/sparql> { > > ?concept skos:prefLabel ?label . > > } > > } > > then I would the following result set: > > || ?doc || ?concept || ?label || > > | <doc1> | <conceptB> | "concept label" | > > | <doc1> | <conceptB> | "concept label" | > > | <doc1> | <conceptB> | "concept label" | > > > > Interestingly, the number of duplicated records can change from processing > to processing. Is this something that other people have encountered? > > > > Kind regards, > > > > *Quentin Reul * > > > > > ------------------------------------------------------------------------------ > > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > > |
From: Bryan T. <br...@sy...> - 2015-11-13 19:41:19
|
What OS? Are you trying to open the journal in two separate processes (this is the most common cause). ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://blazegraph.com http://blog.blazegraph.com Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Fri, Nov 13, 2015 at 2:26 PM, Joakim Soderberg < joa...@bl...> wrote: > Hi, > What did most likely go wrong if I get the following error: > > *exception* > > javax.servlet.ServletException: Servlet.init() for servlet com.blippar.servlet.SparqlServlet threw exception > org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) > org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) > org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) > org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) > org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) > org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) > java.lang.Thread.run(Thread.java:745) > > *root cause* > > java.lang.RuntimeException: FileLock Overlap > com.bigdata.journal.FileMetadata.reopenChannel(FileMetadata.java:1245) > com.bigdata.journal.FileMetadata.access$000(FileMetadata.java:58) > com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1163) > com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1153) > com.bigdata.journal.FileMetadata.<init>(FileMetadata.java:946) > com.bigdata.journal.FileMetadata.createInstance(FileMetadata.java:1470) > com.bigdata.journal.AbstractJournal.<init>(AbstractJournal.java:1156) > com.bigdata.journal.Journal.<init>(Journal.java:275) > com.bigdata.journal.Journal.<init>(Journal.java:268) > com.bigdata.rdf.sail.BigdataSail.createLTS(BigdataSail.java:710) > com.bigdata.rdf.sail.BigdataSail.<init>(BigdataSail.java:689) > com.blippar.servlet.DataManager.initialize(DataManager.java:74) > com.blippar.servlet.SparqlServlet.init(SparqlServlet.java:42) > javax.servlet.GenericServlet.init(GenericServlet.java:158) > org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) > org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) > org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) > org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) > org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) > org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) > java.lang.Thread.run(Thread.java:745) > > *root cause* > > java.nio.channels.OverlappingFileLockException > sun.nio.ch.SharedFileLockTable.checkList(FileLockTable.java:255) > sun.nio.ch.SharedFileLockTable.add(FileLockTable.java:152) > sun.nio.ch.FileChannelImpl.tryLock(FileChannelImpl.java:1075) > com.bigdata.journal.FileMetadata.reopenChannel(FileMetadata.java:1210) > com.bigdata.journal.FileMetadata.access$000(FileMetadata.java:58) > com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1163) > com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1153) > com.bigdata.journal.FileMetadata.<init>(FileMetadata.java:946) > com.bigdata.journal.FileMetadata.createInstance(FileMetadata.java:1470) > com.bigdata.journal.AbstractJournal.<init>(AbstractJournal.java:1156) > com.bigdata.journal.Journal.<init>(Journal.java:275) > com.bigdata.journal.Journal.<init>(Journal.java:268) > com.bigdata.rdf.sail.BigdataSail.createLTS(BigdataSail.java:710) > com.bigdata.rdf.sail.BigdataSail.<init>(BigdataSail.java:689) > com.blippar.servlet.DataManager.initialize(DataManager.java:74) > com.blippar.servlet.SparqlServlet.init(SparqlServlet.java:42) > javax.servlet.GenericServlet.init(GenericServlet.java:158) > org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) > org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) > org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) > org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) > org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) > org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) > org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) > java.lang.Thread.run(Thread.java:745) > > > > > ------------------------------------------------------------------------------ > > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > > |
From: Joakim S. <joa...@bl...> - 2015-11-13 19:26:21
|
Hi, What did most likely go wrong if I get the following error: exception javax.servlet.ServletException: Servlet.init() for servlet com.blippar.servlet.SparqlServlet threw exception org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) java.lang.Thread.run(Thread.java:745) root cause java.lang.RuntimeException: FileLock Overlap com.bigdata.journal.FileMetadata.reopenChannel(FileMetadata.java:1245) com.bigdata.journal.FileMetadata.access$000(FileMetadata.java:58) com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1163) com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1153) com.bigdata.journal.FileMetadata.<init>(FileMetadata.java:946) com.bigdata.journal.FileMetadata.createInstance(FileMetadata.java:1470) com.bigdata.journal.AbstractJournal.<init>(AbstractJournal.java:1156) com.bigdata.journal.Journal.<init>(Journal.java:275) com.bigdata.journal.Journal.<init>(Journal.java:268) com.bigdata.rdf.sail.BigdataSail.createLTS(BigdataSail.java:710) com.bigdata.rdf.sail.BigdataSail.<init>(BigdataSail.java:689) com.blippar.servlet.DataManager.initialize(DataManager.java:74) com.blippar.servlet.SparqlServlet.init(SparqlServlet.java:42) javax.servlet.GenericServlet.init(GenericServlet.java:158) org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) java.lang.Thread.run(Thread.java:745) root cause java.nio.channels.OverlappingFileLockException sun.nio.ch.SharedFileLockTable.checkList(FileLockTable.java:255) sun.nio.ch.SharedFileLockTable.add(FileLockTable.java:152) sun.nio.ch.FileChannelImpl.tryLock(FileChannelImpl.java:1075) com.bigdata.journal.FileMetadata.reopenChannel(FileMetadata.java:1210) com.bigdata.journal.FileMetadata.access$000(FileMetadata.java:58) com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1163) com.bigdata.journal.FileMetadata$1.reopenChannel(FileMetadata.java:1153) com.bigdata.journal.FileMetadata.<init>(FileMetadata.java:946) com.bigdata.journal.FileMetadata.createInstance(FileMetadata.java:1470) com.bigdata.journal.AbstractJournal.<init>(AbstractJournal.java:1156) com.bigdata.journal.Journal.<init>(Journal.java:275) com.bigdata.journal.Journal.<init>(Journal.java:268) com.bigdata.rdf.sail.BigdataSail.createLTS(BigdataSail.java:710) com.bigdata.rdf.sail.BigdataSail.<init>(BigdataSail.java:689) com.blippar.servlet.DataManager.initialize(DataManager.java:74) com.blippar.servlet.SparqlServlet.init(SparqlServlet.java:42) javax.servlet.GenericServlet.init(GenericServlet.java:158) org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:617) org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1091) org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1527) org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1484) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) java.lang.Thread.run(Thread.java:745) |
From: Reul, Q. <que...@wo...> - 2015-11-13 19:09:14
|
Hi all, I'm encountering a weird behaviour when running SPARQL SELECT queries including SERVICE in the WHERE clause. More specifically, I seem to retrieve duplicate records in the result set when DISTINCT is not used. We are using RemoteRepositoryManager to access the BlazeGraph 1.5.2 instance and both namespaces are defined on the same machine. Let us imagine that I have the following triples in namespace A: <doc1> dcterms:subject <conceptB> <doc1> dcterms:title "Title of Document"^^xsd:string and some triples in namespace B: <conceptB> rdf:type skos:Concept <conceptB> skos:prefLabel "concept label"@en If I run the following SELECT query SELECT ?doc ?concept ?label WHERE { ?doc dcterms:subject ?concept . SERVICE <http:localhost:9999/bigdata/namespace/B/sparql> { ?concept skos:prefLabel ?label . } } then I would the following result set: || ?doc || ?concept || ?label || | <doc1> | <conceptB> | "concept label" | | <doc1> | <conceptB> | "concept label" | | <doc1> | <conceptB> | "concept label" | Interestingly, the number of duplicated records can change from processing to processing. Is this something that other people have encountered? Kind regards, Quentin Reul |
From: Brad B. <be...@sy...> - 2015-11-10 17:08:41
|
Great -- thanks. I've commented on the ticket with the workaround. Thanks, --Brad On Tue, Nov 10, 2015 at 11:57 AM, Joakim Soderberg < joa...@bl...> wrote: > Thanks Brad, > I am operating in embedded mode, so I will use SPARQLResultsJSONWriter() until > it’s fixed. > > /J > > > On Nov 10, 2015, at 8:53 AM, Brad Bebee <be...@sy...> wrote: > > Joakim, > > It looks like this is a bug in the ExportKB [1]. As a work-around, you > can export the KBs via the REST API. > > [1] https://jira.blazegraph.com/browse/BLZG-1603 > > Thanks, --Brad > > On Mon, Nov 9, 2015 at 3:08 PM, Joakim Soderberg < > joa...@bl...> wrote: > >> Hi >> Has anyone tried to export a named graph using ExportKB? >> >> After digging on the web I came up with this: >> >> String namespace = the sub graph that I want to export >> >> tripleStore = (AbstractTripleStore) bd.getQueryEngine().getIndexManager().getResourceLocator().locate( >> namespace, ITx.UNISOLATED); >> >> export = new ExportKB( tripleStore, outFile , RDFFormat.NTRIPLES, false); >> >> But I can’t get it to work >> >> >> ------------------------------------------------------------------------------ >> Presto, an open source distributed SQL query engine for big data, >> initially >> developed by Facebook, enables you to easily query your data on Hadoop in >> a >> more interactive manner. Teradata is also now providing full enterprise >> support for Presto. Download a free open source copy now. >> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >> _______________________________________________ >> Bigdata-developers mailing list >> Big...@li... >> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >> >> > > > -- > _______________ > Brad Bebee > CEO, Managing Partner > SYSTAP, LLC > e: be...@sy... > m: 202.642.7961 > f: 571.367.5000 > w: www.blazegraph.com > > Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance > graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints > APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new > technology to use GPUs to accelerate data-parallel graph analytics. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > > -- _______________ Brad Bebee CEO, Managing Partner SYSTAP, LLC e: be...@sy... m: 202.642.7961 f: 571.367.5000 w: www.blazegraph.com Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new technology to use GPUs to accelerate data-parallel graph analytics. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |
From: Joakim S. <joa...@bl...> - 2015-11-10 16:58:06
|
Thanks Brad, I am operating in embedded mode, so I will use SPARQLResultsJSONWriter() until it’s fixed. /J > On Nov 10, 2015, at 8:53 AM, Brad Bebee <be...@sy...> wrote: > > Joakim, > > It looks like this is a bug in the ExportKB [1]. As a work-around, you can export the KBs via the REST API. > > [1] https://jira.blazegraph.com/browse/BLZG-1603 <https://jira.blazegraph.com/browse/BLZG-1603> > > Thanks, --Brad > > On Mon, Nov 9, 2015 at 3:08 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: > Hi > Has anyone tried to export a named graph using ExportKB? > > After digging on the web I came up with this: > > String namespace = the sub graph that I want to export > > tripleStore = (AbstractTripleStore) bd.getQueryEngine().getIndexManager().getResourceLocator().locate( namespace, ITx.UNISOLATED); > > export = new ExportKB( tripleStore, outFile , RDFFormat.NTRIPLES, false); > > But I can’t get it to work > > ------------------------------------------------------------------------------ > Presto, an open source distributed SQL query engine for big data, initially > developed by Facebook, enables you to easily query your data on Hadoop in a > more interactive manner. Teradata is also now providing full enterprise > support for Presto. Download a free open source copy now. > http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140> > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> > > > > > -- > _______________ > Brad Bebee > CEO, Managing Partner > SYSTAP, LLC > e: be...@sy... <mailto:be...@sy...> > m: 202.642.7961 > f: 571.367.5000 > w: www.blazegraph.com <http://www.blazegraph.com/> > > Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new technology to use GPUs to accelerate data-parallel graph analytics. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > |
From: Brad B. <be...@sy...> - 2015-11-10 16:53:29
|
Joakim, It looks like this is a bug in the ExportKB [1]. As a work-around, you can export the KBs via the REST API. [1] https://jira.blazegraph.com/browse/BLZG-1603 Thanks, --Brad On Mon, Nov 9, 2015 at 3:08 PM, Joakim Soderberg < joa...@bl...> wrote: > Hi > Has anyone tried to export a named graph using ExportKB? > > After digging on the web I came up with this: > > String namespace = the sub graph that I want to export > > tripleStore = (AbstractTripleStore) bd.getQueryEngine().getIndexManager().getResourceLocator().locate( > namespace, ITx.UNISOLATED); > > export = new ExportKB( tripleStore, outFile , RDFFormat.NTRIPLES, false); > > But I can’t get it to work > > > ------------------------------------------------------------------------------ > Presto, an open source distributed SQL query engine for big data, initially > developed by Facebook, enables you to easily query your data on Hadoop in a > more interactive manner. Teradata is also now providing full enterprise > support for Presto. Download a free open source copy now. > http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > > -- _______________ Brad Bebee CEO, Managing Partner SYSTAP, LLC e: be...@sy... m: 202.642.7961 f: 571.367.5000 w: www.blazegraph.com Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new technology to use GPUs to accelerate data-parallel graph analytics. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |
From: Brad B. <be...@sy...> - 2015-11-10 14:16:25
|
Alex, Good news. Let us know if you hit any other items. Thanks, Brad _______________ Brad Bebee CEO, Managing Partner SYSTAP, LLC e: be...@sy... m: 202.642.7961 f: 571.367.5000 w: www.systap.com Blazegraph™ is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. MapGraph™ is our disruptive new technology to use GPUs to accelerate data-parallel graph analytics. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Nov 10, 2015 8:28 AM, "Alex Muir" <ale...@gm...> wrote: > Thanks Brad,, > > That did work out great.. Thanks > > > > Regards > Alex > www.tilogeo.com > > On Tue, Nov 10, 2015 at 4:03 AM, Brad Bebee <be...@sy...> wrote: > >> Alex, >> >> What should be working is something like: >> >> curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf >> http://62.59.40.122:9999/bigdata/sparql?context-uri=http://abc.com/id/graph/xyz <http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz> >> >> curl -X POST http://62.59.40.122:9999/bigdata/sparql >> --data-urlencode 'query=construct where {?s ?p ?o}' -H 'Accept: application/rdf+xml' --data-urlencode 'named-graph-uri=http://abc.com/id/graph/xyz <http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz>' >> >> >> Let us know how if that works on your end. >> >> Thanks, --Brad >> >> >> >> >> >> On Mon, Nov 9, 2015 at 5:39 PM, Alex Muir <ale...@gm...> wrote: >> >>> I've tried various means of encoding the components but I always get the >>> same service description results or notifications about encoding and >>> expected data. Can't seem to get the right combination. >>> >>> Hope someone out there can give an example of curl that imports rdf and >>> exports the same ref using context... >>> >>> Is there any difference between these? It seems different names for the >>> same concept in the rest api. Is that correct? >>> >>> named-graph-uri= >>> The Context (aka Named Graph) c= >>> context-uri= >>> >>> Thanks >>> >>> >>> >>> >>> >>> Regards >>> Alex >>> www.tilogeo.com >>> >>> On Mon, Nov 9, 2015 at 9:44 PM, Martynas Jusevičius < >>> mar...@gr...> wrote: >>> >>>> *Note that.. >>>> >>>> On Mon, Nov 9, 2015 at 10:39 PM, Martynas Jusevičius < >>>> mar...@gr...> wrote: >>>> >>>>> Not sure what the command line does, better if you send the full >>>>> request URI. >>>>> >>>>> Not that you only have to encode the components, such as querystring >>>>> params/values, not the whole URI. >>>>> >>>>> On Mon, Nov 9, 2015 at 10:37 PM, Alex Muir <ale...@gm...> >>>>> wrote: >>>>> >>>>>> oops.. actually sorry that was not true... had a bug in that.. :) >>>>>> >>>>>> >>>>>> Regards >>>>>> Alex >>>>>> www.tilogeo.com >>>>>> >>>>>> On Mon, Nov 9, 2015 at 9:34 PM, Alex Muir <ale...@gm...> >>>>>> wrote: >>>>>> >>>>>>> Hi martynas, >>>>>>> >>>>>>> Sorry sent that last one by accident.. >>>>>>> >>>>>>> I get the same result with the following, encoding the url. >>>>>>> >>>>>>> URLENCODE=$(cat $1?named-graph-uri=$2 | xxd -plain | tr -d '\n' | >>>>>>> sed 's/\(..\)/%\1/g') >>>>>>> curl -H "Accept: application/rdf+xml" $URLENCODED -o $3/$4.rdf >>>>>>> >>>>>>> >>>>>>> >>>>>>> Regards >>>>>>> Alex >>>>>>> www.tilogeo.com >>>>>>> >>>>>>> On Mon, Nov 9, 2015 at 9:31 PM, Alex Muir <ale...@gm...> >>>>>>> wrote: >>>>>>> >>>>>>>> I get the same result >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Regards >>>>>>>> Alex >>>>>>>> www.tilogeo.com >>>>>>>> >>>>>>>> On Mon, Nov 9, 2015 at 6:28 PM, Martynas Jusevičius < >>>>>>>> mar...@gr...> wrote: >>>>>>>> >>>>>>>>> Are your query parameters percent-encoded? >>>>>>>>> https://en.wikipedia.org/wiki/Percent-encoding >>>>>>>>> >>>>>>>>> On Mon, Nov 9, 2015 at 7:11 PM, Alex Muir <ale...@gm...> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi Bryan, >>>>>>>>>> >>>>>>>>>> I've tried that and a number of methods. On export though I get >>>>>>>>>> data that I guess is a description for the service. >>>>>>>>>> >>>>>>>>>> Can blazegraph create some specific examples to show how to >>>>>>>>>> accomplish this using curl? The task is to load an rdf xml file and then >>>>>>>>>> export the same file using a named graph. >>>>>>>>>> >>>>>>>>>> I'm evaluating the system for a large client and have completed >>>>>>>>>> this task for other systems but I'm not clear on how to do this with the >>>>>>>>>> given documentation. >>>>>>>>>> >>>>>>>>>> [exec] <rdf:RDF >>>>>>>>>> [exec] xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"> >>>>>>>>>> [exec] >>>>>>>>>> [exec] <rdf:Description rdf:nodeID="service"> >>>>>>>>>> [exec] <rdf:type rdf:resource="http://www.w3.org/ns/sparql-service-description#Service"/> >>>>>>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/sparql"/> >>>>>>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/LBS/sparql"/> >>>>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL10Query"/> >>>>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Query"/> >>>>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Update"/> >>>>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#BasicFederatedQuery"/> >>>>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#UnionDefaultGraph"/> >>>>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.bigdata.com/rdf#/features/KB/Mode/Quads"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://sw.deri.org/2008/07/n-quads/#n-quads"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>>>>>> [exec] <entailmentRegime xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/entailment/Simple"/> >>>>>>>>>> [exec] <defaultDataset xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:nodeID="defaultDataset"/> >>>>>>>>>> [exec] </rdf:Description> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Regards >>>>>>>>>> Alex >>>>>>>>>> www.tilogeo.com >>>>>>>>>> >>>>>>>>>> On Mon, Nov 9, 2015 at 1:46 PM, Bryan Thompson <br...@sy...> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Alex, >>>>>>>>>>> >>>>>>>>>>> I believe that you should be using the parameters defined at [1] >>>>>>>>>>> for SPARQL UPDATE. Notably, replace ?c=... with >>>>>>>>>>> using-named-graph-uriSpecify zero or more named graphs for this >>>>>>>>>>> the update request (protocol option with the same semantics as USING NAMED). >>>>>>>>>>> >>>>>>>>>>> This is per the SPARQL UPDATE specification. >>>>>>>>>>> >>>>>>>>>>> Thanks, >>>>>>>>>>> Bryan >>>>>>>>>>> >>>>>>>>>>> [1] >>>>>>>>>>> https://wiki.blazegraph.com/wiki/index.php/REST_API#UPDATE_.28SPARQL_1.1_UPDATE.29 >>>>>>>>>>> >>>>>>>>>>> ---- >>>>>>>>>>> Bryan Thompson >>>>>>>>>>> Chief Scientist & Founder >>>>>>>>>>> SYSTAP, LLC >>>>>>>>>>> 4501 Tower Road >>>>>>>>>>> Greensboro, NC 27410 >>>>>>>>>>> br...@sy... >>>>>>>>>>> http://blazegraph.com >>>>>>>>>>> http://blog.blazegraph.com >>>>>>>>>>> >>>>>>>>>>> Blazegraph™ <http://www.blazegraph.com/> is our ultra >>>>>>>>>>> high-performance graph database that supports both RDF/SPARQL and >>>>>>>>>>> Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU >>>>>>>>>>> acceleration using our disruptive technology to accelerate >>>>>>>>>>> data-parallel graph analytics and graph query. >>>>>>>>>>> >>>>>>>>>>> CONFIDENTIALITY NOTICE: This email and its contents and >>>>>>>>>>> attachments are for the sole use of the intended recipient(s) and are >>>>>>>>>>> confidential or proprietary to SYSTAP. Any unauthorized review, use, >>>>>>>>>>> disclosure, dissemination or copying of this email or its contents or >>>>>>>>>>> attachments is prohibited. If you have received this communication in >>>>>>>>>>> error, please notify the sender by reply email and permanently delete all >>>>>>>>>>> copies of the email and its contents and attachments. >>>>>>>>>>> >>>>>>>>>>> On Sun, Nov 8, 2015 at 1:49 PM, Alex Muir <ale...@gm... >>>>>>>>>>> > wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> Using the REST API how do I export the same data file that I uploaded? >>>>>>>>>>>> >>>>>>>>>>>> I'm unclear with the BLAZEGRAPH REST API on the method to associate a named graph on upload and export that same named graph. >>>>>>>>>>>> >>>>>>>>>>>> With the following >>>>>>>>>>>> >>>>>>>>>>>> curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf >>>>>>>>>>>> http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz >>>>>>>>>>>> >>>>>>>>>>>> curl -X POST http://62.59.40.122:9999/bigdata/sparql >>>>>>>>>>>> --data-urlencode 'query=named-graph-uri http://abc.com/id/graph/xyz' -H 'Accept: application/rdf+xml" | gzip > data.rdf.gz >>>>>>>>>>>> >>>>>>>>>>>> I get data exported but not the same large file that I >>>>>>>>>>>> inserted. >>>>>>>>>>>> >>>>>>>>>>>> Regards >>>>>>>>>>>> Alex >>>>>>>>>>>> www.tilogeo.com >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Bigdata-developers mailing list >>>>>>>>>>>> Big...@li... >>>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> Presto, an open source distributed SQL query engine for big data, >>>>>>>>>> initially >>>>>>>>>> developed by Facebook, enables you to easily query your data on >>>>>>>>>> Hadoop in a >>>>>>>>>> more interactive manner. Teradata is also now providing full >>>>>>>>>> enterprise >>>>>>>>>> support for Presto. Download a free open source copy now. >>>>>>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >>>>>>>>>> _______________________________________________ >>>>>>>>>> Bigdata-developers mailing list >>>>>>>>>> Big...@li... >>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >>> >>> ------------------------------------------------------------------------------ >>> Presto, an open source distributed SQL query engine for big data, >>> initially >>> developed by Facebook, enables you to easily query your data on Hadoop >>> in a >>> more interactive manner. Teradata is also now providing full enterprise >>> support for Presto. Download a free open source copy now. >>> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >>> _______________________________________________ >>> Bigdata-developers mailing list >>> Big...@li... >>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>> >>> >> >> >> -- >> _______________ >> Brad Bebee >> CEO, Managing Partner >> SYSTAP, LLC >> e: be...@sy... >> m: 202.642.7961 >> f: 571.367.5000 >> w: www.blazegraph.com >> >> Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance >> graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints >> APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new >> technology to use GPUs to accelerate data-parallel graph analytics. >> >> CONFIDENTIALITY NOTICE: This email and its contents and attachments are >> for the sole use of the intended recipient(s) and are confidential or >> proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, >> dissemination or copying of this email or its contents or attachments is >> prohibited. If you have received this communication in error, please notify >> the sender by reply email and permanently delete all copies of the email >> and its contents and attachments. >> > > |
From: Alex M. <ale...@gm...> - 2015-11-10 13:28:29
|
Thanks Brad,, That did work out great.. Thanks Regards Alex www.tilogeo.com On Tue, Nov 10, 2015 at 4:03 AM, Brad Bebee <be...@sy...> wrote: > Alex, > > What should be working is something like: > > curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf > http://62.59.40.122:9999/bigdata/sparql?context-uri=http://abc.com/id/graph/xyz <http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz> > > curl -X POST http://62.59.40.122:9999/bigdata/sparql > --data-urlencode 'query=construct where {?s ?p ?o}' -H 'Accept: application/rdf+xml' --data-urlencode 'named-graph-uri=http://abc.com/id/graph/xyz <http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz>' > > > Let us know how if that works on your end. > > Thanks, --Brad > > > > > > On Mon, Nov 9, 2015 at 5:39 PM, Alex Muir <ale...@gm...> wrote: > >> I've tried various means of encoding the components but I always get the >> same service description results or notifications about encoding and >> expected data. Can't seem to get the right combination. >> >> Hope someone out there can give an example of curl that imports rdf and >> exports the same ref using context... >> >> Is there any difference between these? It seems different names for the >> same concept in the rest api. Is that correct? >> >> named-graph-uri= >> The Context (aka Named Graph) c= >> context-uri= >> >> Thanks >> >> >> >> >> >> Regards >> Alex >> www.tilogeo.com >> >> On Mon, Nov 9, 2015 at 9:44 PM, Martynas Jusevičius < >> mar...@gr...> wrote: >> >>> *Note that.. >>> >>> On Mon, Nov 9, 2015 at 10:39 PM, Martynas Jusevičius < >>> mar...@gr...> wrote: >>> >>>> Not sure what the command line does, better if you send the full >>>> request URI. >>>> >>>> Not that you only have to encode the components, such as querystring >>>> params/values, not the whole URI. >>>> >>>> On Mon, Nov 9, 2015 at 10:37 PM, Alex Muir <ale...@gm...> >>>> wrote: >>>> >>>>> oops.. actually sorry that was not true... had a bug in that.. :) >>>>> >>>>> >>>>> Regards >>>>> Alex >>>>> www.tilogeo.com >>>>> >>>>> On Mon, Nov 9, 2015 at 9:34 PM, Alex Muir <ale...@gm...> >>>>> wrote: >>>>> >>>>>> Hi martynas, >>>>>> >>>>>> Sorry sent that last one by accident.. >>>>>> >>>>>> I get the same result with the following, encoding the url. >>>>>> >>>>>> URLENCODE=$(cat $1?named-graph-uri=$2 | xxd -plain | tr -d '\n' | sed >>>>>> 's/\(..\)/%\1/g') >>>>>> curl -H "Accept: application/rdf+xml" $URLENCODED -o $3/$4.rdf >>>>>> >>>>>> >>>>>> >>>>>> Regards >>>>>> Alex >>>>>> www.tilogeo.com >>>>>> >>>>>> On Mon, Nov 9, 2015 at 9:31 PM, Alex Muir <ale...@gm...> >>>>>> wrote: >>>>>> >>>>>>> I get the same result >>>>>>> >>>>>>> >>>>>>> >>>>>>> Regards >>>>>>> Alex >>>>>>> www.tilogeo.com >>>>>>> >>>>>>> On Mon, Nov 9, 2015 at 6:28 PM, Martynas Jusevičius < >>>>>>> mar...@gr...> wrote: >>>>>>> >>>>>>>> Are your query parameters percent-encoded? >>>>>>>> https://en.wikipedia.org/wiki/Percent-encoding >>>>>>>> >>>>>>>> On Mon, Nov 9, 2015 at 7:11 PM, Alex Muir <ale...@gm...> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi Bryan, >>>>>>>>> >>>>>>>>> I've tried that and a number of methods. On export though I get >>>>>>>>> data that I guess is a description for the service. >>>>>>>>> >>>>>>>>> Can blazegraph create some specific examples to show how to >>>>>>>>> accomplish this using curl? The task is to load an rdf xml file and then >>>>>>>>> export the same file using a named graph. >>>>>>>>> >>>>>>>>> I'm evaluating the system for a large client and have completed >>>>>>>>> this task for other systems but I'm not clear on how to do this with the >>>>>>>>> given documentation. >>>>>>>>> >>>>>>>>> [exec] <rdf:RDF >>>>>>>>> [exec] xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"> >>>>>>>>> [exec] >>>>>>>>> [exec] <rdf:Description rdf:nodeID="service"> >>>>>>>>> [exec] <rdf:type rdf:resource="http://www.w3.org/ns/sparql-service-description#Service"/> >>>>>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/sparql"/> >>>>>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/LBS/sparql"/> >>>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL10Query"/> >>>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Query"/> >>>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Update"/> >>>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#BasicFederatedQuery"/> >>>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#UnionDefaultGraph"/> >>>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.bigdata.com/rdf#/features/KB/Mode/Quads"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://sw.deri.org/2008/07/n-quads/#n-quads"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>>>>> [exec] <entailmentRegime xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/entailment/Simple"/> >>>>>>>>> [exec] <defaultDataset xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:nodeID="defaultDataset"/> >>>>>>>>> [exec] </rdf:Description> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Regards >>>>>>>>> Alex >>>>>>>>> www.tilogeo.com >>>>>>>>> >>>>>>>>> On Mon, Nov 9, 2015 at 1:46 PM, Bryan Thompson <br...@sy...> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Alex, >>>>>>>>>> >>>>>>>>>> I believe that you should be using the parameters defined at [1] >>>>>>>>>> for SPARQL UPDATE. Notably, replace ?c=... with >>>>>>>>>> using-named-graph-uriSpecify zero or more named graphs for this >>>>>>>>>> the update request (protocol option with the same semantics as USING NAMED). >>>>>>>>>> >>>>>>>>>> This is per the SPARQL UPDATE specification. >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> Bryan >>>>>>>>>> >>>>>>>>>> [1] >>>>>>>>>> https://wiki.blazegraph.com/wiki/index.php/REST_API#UPDATE_.28SPARQL_1.1_UPDATE.29 >>>>>>>>>> >>>>>>>>>> ---- >>>>>>>>>> Bryan Thompson >>>>>>>>>> Chief Scientist & Founder >>>>>>>>>> SYSTAP, LLC >>>>>>>>>> 4501 Tower Road >>>>>>>>>> Greensboro, NC 27410 >>>>>>>>>> br...@sy... >>>>>>>>>> http://blazegraph.com >>>>>>>>>> http://blog.blazegraph.com >>>>>>>>>> >>>>>>>>>> Blazegraph™ <http://www.blazegraph.com/> is our ultra >>>>>>>>>> high-performance graph database that supports both RDF/SPARQL and >>>>>>>>>> Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU >>>>>>>>>> acceleration using our disruptive technology to accelerate >>>>>>>>>> data-parallel graph analytics and graph query. >>>>>>>>>> >>>>>>>>>> CONFIDENTIALITY NOTICE: This email and its contents and >>>>>>>>>> attachments are for the sole use of the intended recipient(s) and are >>>>>>>>>> confidential or proprietary to SYSTAP. Any unauthorized review, use, >>>>>>>>>> disclosure, dissemination or copying of this email or its contents or >>>>>>>>>> attachments is prohibited. If you have received this communication in >>>>>>>>>> error, please notify the sender by reply email and permanently delete all >>>>>>>>>> copies of the email and its contents and attachments. >>>>>>>>>> >>>>>>>>>> On Sun, Nov 8, 2015 at 1:49 PM, Alex Muir <ale...@gm...> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Hi, >>>>>>>>>>> >>>>>>>>>>> Using the REST API how do I export the same data file that I uploaded? >>>>>>>>>>> >>>>>>>>>>> I'm unclear with the BLAZEGRAPH REST API on the method to associate a named graph on upload and export that same named graph. >>>>>>>>>>> >>>>>>>>>>> With the following >>>>>>>>>>> >>>>>>>>>>> curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf >>>>>>>>>>> http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz >>>>>>>>>>> >>>>>>>>>>> curl -X POST http://62.59.40.122:9999/bigdata/sparql >>>>>>>>>>> --data-urlencode 'query=named-graph-uri http://abc.com/id/graph/xyz' -H 'Accept: application/rdf+xml" | gzip > data.rdf.gz >>>>>>>>>>> >>>>>>>>>>> I get data exported but not the same large file that I inserted. >>>>>>>>>>> >>>>>>>>>>> Regards >>>>>>>>>>> Alex >>>>>>>>>>> www.tilogeo.com >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Bigdata-developers mailing list >>>>>>>>>>> Big...@li... >>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>> Presto, an open source distributed SQL query engine for big data, >>>>>>>>> initially >>>>>>>>> developed by Facebook, enables you to easily query your data on >>>>>>>>> Hadoop in a >>>>>>>>> more interactive manner. Teradata is also now providing full >>>>>>>>> enterprise >>>>>>>>> support for Presto. Download a free open source copy now. >>>>>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >>>>>>>>> _______________________________________________ >>>>>>>>> Bigdata-developers mailing list >>>>>>>>> Big...@li... >>>>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >> >> ------------------------------------------------------------------------------ >> Presto, an open source distributed SQL query engine for big data, >> initially >> developed by Facebook, enables you to easily query your data on Hadoop in >> a >> more interactive manner. Teradata is also now providing full enterprise >> support for Presto. Download a free open source copy now. >> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >> _______________________________________________ >> Bigdata-developers mailing list >> Big...@li... >> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >> >> > > > -- > _______________ > Brad Bebee > CEO, Managing Partner > SYSTAP, LLC > e: be...@sy... > m: 202.642.7961 > f: 571.367.5000 > w: www.blazegraph.com > > Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance > graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints > APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new > technology to use GPUs to accelerate data-parallel graph analytics. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > |
From: Brad B. <be...@sy...> - 2015-11-10 04:04:01
|
Alex, What should be working is something like: curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf http://62.59.40.122:9999/bigdata/sparql?context-uri=http://abc.com/id/graph/xyz <http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz> curl -X POST http://62.59.40.122:9999/bigdata/sparql --data-urlencode 'query=construct where {?s ?p ?o}' -H 'Accept: application/rdf+xml' --data-urlencode 'named-graph-uri=http://abc.com/id/graph/xyz <http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz>' Let us know how if that works on your end. Thanks, --Brad On Mon, Nov 9, 2015 at 5:39 PM, Alex Muir <ale...@gm...> wrote: > I've tried various means of encoding the components but I always get the > same service description results or notifications about encoding and > expected data. Can't seem to get the right combination. > > Hope someone out there can give an example of curl that imports rdf and > exports the same ref using context... > > Is there any difference between these? It seems different names for the > same concept in the rest api. Is that correct? > > named-graph-uri= > The Context (aka Named Graph) c= > context-uri= > > Thanks > > > > > > Regards > Alex > www.tilogeo.com > > On Mon, Nov 9, 2015 at 9:44 PM, Martynas Jusevičius <mar...@gr... > > wrote: > >> *Note that.. >> >> On Mon, Nov 9, 2015 at 10:39 PM, Martynas Jusevičius < >> mar...@gr...> wrote: >> >>> Not sure what the command line does, better if you send the full request >>> URI. >>> >>> Not that you only have to encode the components, such as querystring >>> params/values, not the whole URI. >>> >>> On Mon, Nov 9, 2015 at 10:37 PM, Alex Muir <ale...@gm...> >>> wrote: >>> >>>> oops.. actually sorry that was not true... had a bug in that.. :) >>>> >>>> >>>> Regards >>>> Alex >>>> www.tilogeo.com >>>> >>>> On Mon, Nov 9, 2015 at 9:34 PM, Alex Muir <ale...@gm...> >>>> wrote: >>>> >>>>> Hi martynas, >>>>> >>>>> Sorry sent that last one by accident.. >>>>> >>>>> I get the same result with the following, encoding the url. >>>>> >>>>> URLENCODE=$(cat $1?named-graph-uri=$2 | xxd -plain | tr -d '\n' | sed >>>>> 's/\(..\)/%\1/g') >>>>> curl -H "Accept: application/rdf+xml" $URLENCODED -o $3/$4.rdf >>>>> >>>>> >>>>> >>>>> Regards >>>>> Alex >>>>> www.tilogeo.com >>>>> >>>>> On Mon, Nov 9, 2015 at 9:31 PM, Alex Muir <ale...@gm...> >>>>> wrote: >>>>> >>>>>> I get the same result >>>>>> >>>>>> >>>>>> >>>>>> Regards >>>>>> Alex >>>>>> www.tilogeo.com >>>>>> >>>>>> On Mon, Nov 9, 2015 at 6:28 PM, Martynas Jusevičius < >>>>>> mar...@gr...> wrote: >>>>>> >>>>>>> Are your query parameters percent-encoded? >>>>>>> https://en.wikipedia.org/wiki/Percent-encoding >>>>>>> >>>>>>> On Mon, Nov 9, 2015 at 7:11 PM, Alex Muir <ale...@gm...> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi Bryan, >>>>>>>> >>>>>>>> I've tried that and a number of methods. On export though I get >>>>>>>> data that I guess is a description for the service. >>>>>>>> >>>>>>>> Can blazegraph create some specific examples to show how to >>>>>>>> accomplish this using curl? The task is to load an rdf xml file and then >>>>>>>> export the same file using a named graph. >>>>>>>> >>>>>>>> I'm evaluating the system for a large client and have completed >>>>>>>> this task for other systems but I'm not clear on how to do this with the >>>>>>>> given documentation. >>>>>>>> >>>>>>>> [exec] <rdf:RDF >>>>>>>> [exec] xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"> >>>>>>>> [exec] >>>>>>>> [exec] <rdf:Description rdf:nodeID="service"> >>>>>>>> [exec] <rdf:type rdf:resource="http://www.w3.org/ns/sparql-service-description#Service"/> >>>>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/sparql"/> >>>>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/LBS/sparql"/> >>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL10Query"/> >>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Query"/> >>>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Update"/> >>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#BasicFederatedQuery"/> >>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#UnionDefaultGraph"/> >>>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.bigdata.com/rdf#/features/KB/Mode/Quads"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://sw.deri.org/2008/07/n-quads/#n-quads"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>>>> [exec] <entailmentRegime xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/entailment/Simple"/> >>>>>>>> [exec] <defaultDataset xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:nodeID="defaultDataset"/> >>>>>>>> [exec] </rdf:Description> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Regards >>>>>>>> Alex >>>>>>>> www.tilogeo.com >>>>>>>> >>>>>>>> On Mon, Nov 9, 2015 at 1:46 PM, Bryan Thompson <br...@sy...> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Alex, >>>>>>>>> >>>>>>>>> I believe that you should be using the parameters defined at [1] >>>>>>>>> for SPARQL UPDATE. Notably, replace ?c=... with >>>>>>>>> using-named-graph-uriSpecify zero or more named graphs for this >>>>>>>>> the update request (protocol option with the same semantics as USING NAMED). >>>>>>>>> >>>>>>>>> This is per the SPARQL UPDATE specification. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Bryan >>>>>>>>> >>>>>>>>> [1] >>>>>>>>> https://wiki.blazegraph.com/wiki/index.php/REST_API#UPDATE_.28SPARQL_1.1_UPDATE.29 >>>>>>>>> >>>>>>>>> ---- >>>>>>>>> Bryan Thompson >>>>>>>>> Chief Scientist & Founder >>>>>>>>> SYSTAP, LLC >>>>>>>>> 4501 Tower Road >>>>>>>>> Greensboro, NC 27410 >>>>>>>>> br...@sy... >>>>>>>>> http://blazegraph.com >>>>>>>>> http://blog.blazegraph.com >>>>>>>>> >>>>>>>>> Blazegraph™ <http://www.blazegraph.com/> is our ultra >>>>>>>>> high-performance graph database that supports both RDF/SPARQL and >>>>>>>>> Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU >>>>>>>>> acceleration using our disruptive technology to accelerate >>>>>>>>> data-parallel graph analytics and graph query. >>>>>>>>> >>>>>>>>> CONFIDENTIALITY NOTICE: This email and its contents and >>>>>>>>> attachments are for the sole use of the intended recipient(s) and are >>>>>>>>> confidential or proprietary to SYSTAP. Any unauthorized review, use, >>>>>>>>> disclosure, dissemination or copying of this email or its contents or >>>>>>>>> attachments is prohibited. If you have received this communication in >>>>>>>>> error, please notify the sender by reply email and permanently delete all >>>>>>>>> copies of the email and its contents and attachments. >>>>>>>>> >>>>>>>>> On Sun, Nov 8, 2015 at 1:49 PM, Alex Muir <ale...@gm...> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> Using the REST API how do I export the same data file that I uploaded? >>>>>>>>>> >>>>>>>>>> I'm unclear with the BLAZEGRAPH REST API on the method to associate a named graph on upload and export that same named graph. >>>>>>>>>> >>>>>>>>>> With the following >>>>>>>>>> >>>>>>>>>> curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf >>>>>>>>>> http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz >>>>>>>>>> >>>>>>>>>> curl -X POST http://62.59.40.122:9999/bigdata/sparql >>>>>>>>>> --data-urlencode 'query=named-graph-uri http://abc.com/id/graph/xyz' -H 'Accept: application/rdf+xml" | gzip > data.rdf.gz >>>>>>>>>> >>>>>>>>>> I get data exported but not the same large file that I inserted. >>>>>>>>>> >>>>>>>>>> Regards >>>>>>>>>> Alex >>>>>>>>>> www.tilogeo.com >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Bigdata-developers mailing list >>>>>>>>>> Big...@li... >>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> ------------------------------------------------------------------------------ >>>>>>>> Presto, an open source distributed SQL query engine for big data, >>>>>>>> initially >>>>>>>> developed by Facebook, enables you to easily query your data on >>>>>>>> Hadoop in a >>>>>>>> more interactive manner. Teradata is also now providing full >>>>>>>> enterprise >>>>>>>> support for Presto. Download a free open source copy now. >>>>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >>>>>>>> _______________________________________________ >>>>>>>> Bigdata-developers mailing list >>>>>>>> Big...@li... >>>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> > > > ------------------------------------------------------------------------------ > Presto, an open source distributed SQL query engine for big data, initially > developed by Facebook, enables you to easily query your data on Hadoop in a > more interactive manner. Teradata is also now providing full enterprise > support for Presto. Download a free open source copy now. > http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > > -- _______________ Brad Bebee CEO, Managing Partner SYSTAP, LLC e: be...@sy... m: 202.642.7961 f: 571.367.5000 w: www.blazegraph.com Blazegraph™ <http://www.blazegraph.com> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Mapgraph™ <http://www.systap.com/mapgraph> is our disruptive new technology to use GPUs to accelerate data-parallel graph analytics. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. |
From: Alex M. <ale...@gm...> - 2015-11-09 22:39:36
|
I've tried various means of encoding the components but I always get the same service description results or notifications about encoding and expected data. Can't seem to get the right combination. Hope someone out there can give an example of curl that imports rdf and exports the same ref using context... Is there any difference between these? It seems different names for the same concept in the rest api. Is that correct? named-graph-uri= The Context (aka Named Graph) c= context-uri= Thanks Regards Alex www.tilogeo.com On Mon, Nov 9, 2015 at 9:44 PM, Martynas Jusevičius <mar...@gr...> wrote: > *Note that.. > > On Mon, Nov 9, 2015 at 10:39 PM, Martynas Jusevičius < > mar...@gr...> wrote: > >> Not sure what the command line does, better if you send the full request >> URI. >> >> Not that you only have to encode the components, such as querystring >> params/values, not the whole URI. >> >> On Mon, Nov 9, 2015 at 10:37 PM, Alex Muir <ale...@gm...> wrote: >> >>> oops.. actually sorry that was not true... had a bug in that.. :) >>> >>> >>> Regards >>> Alex >>> www.tilogeo.com >>> >>> On Mon, Nov 9, 2015 at 9:34 PM, Alex Muir <ale...@gm...> wrote: >>> >>>> Hi martynas, >>>> >>>> Sorry sent that last one by accident.. >>>> >>>> I get the same result with the following, encoding the url. >>>> >>>> URLENCODE=$(cat $1?named-graph-uri=$2 | xxd -plain | tr -d '\n' | sed >>>> 's/\(..\)/%\1/g') >>>> curl -H "Accept: application/rdf+xml" $URLENCODED -o $3/$4.rdf >>>> >>>> >>>> >>>> Regards >>>> Alex >>>> www.tilogeo.com >>>> >>>> On Mon, Nov 9, 2015 at 9:31 PM, Alex Muir <ale...@gm...> >>>> wrote: >>>> >>>>> I get the same result >>>>> >>>>> >>>>> >>>>> Regards >>>>> Alex >>>>> www.tilogeo.com >>>>> >>>>> On Mon, Nov 9, 2015 at 6:28 PM, Martynas Jusevičius < >>>>> mar...@gr...> wrote: >>>>> >>>>>> Are your query parameters percent-encoded? >>>>>> https://en.wikipedia.org/wiki/Percent-encoding >>>>>> >>>>>> On Mon, Nov 9, 2015 at 7:11 PM, Alex Muir <ale...@gm...> >>>>>> wrote: >>>>>> >>>>>>> Hi Bryan, >>>>>>> >>>>>>> I've tried that and a number of methods. On export though I get data >>>>>>> that I guess is a description for the service. >>>>>>> >>>>>>> Can blazegraph create some specific examples to show how to >>>>>>> accomplish this using curl? The task is to load an rdf xml file and then >>>>>>> export the same file using a named graph. >>>>>>> >>>>>>> I'm evaluating the system for a large client and have completed this >>>>>>> task for other systems but I'm not clear on how to do this with the given >>>>>>> documentation. >>>>>>> >>>>>>> [exec] <rdf:RDF >>>>>>> [exec] xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"> >>>>>>> [exec] >>>>>>> [exec] <rdf:Description rdf:nodeID="service"> >>>>>>> [exec] <rdf:type rdf:resource="http://www.w3.org/ns/sparql-service-description#Service"/> >>>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/sparql"/> >>>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/LBS/sparql"/> >>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL10Query"/> >>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Query"/> >>>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Update"/> >>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#BasicFederatedQuery"/> >>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#UnionDefaultGraph"/> >>>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.bigdata.com/rdf#/features/KB/Mode/Quads"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://sw.deri.org/2008/07/n-quads/#n-quads"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>>> [exec] <entailmentRegime xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/entailment/Simple"/> >>>>>>> [exec] <defaultDataset xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:nodeID="defaultDataset"/> >>>>>>> [exec] </rdf:Description> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Regards >>>>>>> Alex >>>>>>> www.tilogeo.com >>>>>>> >>>>>>> On Mon, Nov 9, 2015 at 1:46 PM, Bryan Thompson <br...@sy...> >>>>>>> wrote: >>>>>>> >>>>>>>> Alex, >>>>>>>> >>>>>>>> I believe that you should be using the parameters defined at [1] >>>>>>>> for SPARQL UPDATE. Notably, replace ?c=... with >>>>>>>> using-named-graph-uriSpecify zero or more named graphs for this >>>>>>>> the update request (protocol option with the same semantics as USING NAMED). >>>>>>>> >>>>>>>> This is per the SPARQL UPDATE specification. >>>>>>>> >>>>>>>> Thanks, >>>>>>>> Bryan >>>>>>>> >>>>>>>> [1] >>>>>>>> https://wiki.blazegraph.com/wiki/index.php/REST_API#UPDATE_.28SPARQL_1.1_UPDATE.29 >>>>>>>> >>>>>>>> ---- >>>>>>>> Bryan Thompson >>>>>>>> Chief Scientist & Founder >>>>>>>> SYSTAP, LLC >>>>>>>> 4501 Tower Road >>>>>>>> Greensboro, NC 27410 >>>>>>>> br...@sy... >>>>>>>> http://blazegraph.com >>>>>>>> http://blog.blazegraph.com >>>>>>>> >>>>>>>> Blazegraph™ <http://www.blazegraph.com/> is our ultra >>>>>>>> high-performance graph database that supports both RDF/SPARQL and >>>>>>>> Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU >>>>>>>> acceleration using our disruptive technology to accelerate >>>>>>>> data-parallel graph analytics and graph query. >>>>>>>> >>>>>>>> CONFIDENTIALITY NOTICE: This email and its contents and >>>>>>>> attachments are for the sole use of the intended recipient(s) and are >>>>>>>> confidential or proprietary to SYSTAP. Any unauthorized review, use, >>>>>>>> disclosure, dissemination or copying of this email or its contents or >>>>>>>> attachments is prohibited. If you have received this communication in >>>>>>>> error, please notify the sender by reply email and permanently delete all >>>>>>>> copies of the email and its contents and attachments. >>>>>>>> >>>>>>>> On Sun, Nov 8, 2015 at 1:49 PM, Alex Muir <ale...@gm...> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> Using the REST API how do I export the same data file that I uploaded? >>>>>>>>> >>>>>>>>> I'm unclear with the BLAZEGRAPH REST API on the method to associate a named graph on upload and export that same named graph. >>>>>>>>> >>>>>>>>> With the following >>>>>>>>> >>>>>>>>> curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf >>>>>>>>> http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz >>>>>>>>> >>>>>>>>> curl -X POST http://62.59.40.122:9999/bigdata/sparql >>>>>>>>> --data-urlencode 'query=named-graph-uri http://abc.com/id/graph/xyz' -H 'Accept: application/rdf+xml" | gzip > data.rdf.gz >>>>>>>>> >>>>>>>>> I get data exported but not the same large file that I inserted. >>>>>>>>> >>>>>>>>> Regards >>>>>>>>> Alex >>>>>>>>> www.tilogeo.com >>>>>>>>> >>>>>>>>> >>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Bigdata-developers mailing list >>>>>>>>> Big...@li... >>>>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> Presto, an open source distributed SQL query engine for big data, >>>>>>> initially >>>>>>> developed by Facebook, enables you to easily query your data on >>>>>>> Hadoop in a >>>>>>> more interactive manner. Teradata is also now providing full >>>>>>> enterprise >>>>>>> support for Presto. Download a free open source copy now. >>>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >>>>>>> _______________________________________________ >>>>>>> Bigdata-developers mailing list >>>>>>> Big...@li... >>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> > |
From: Martynas J. <mar...@gr...> - 2015-11-09 22:04:56
|
Not sure what the command line does, better if you send the full request URI. Not that you only have to encode the components, such as querystring params/values, not the whole URI. On Mon, Nov 9, 2015 at 10:37 PM, Alex Muir <ale...@gm...> wrote: > oops.. actually sorry that was not true... had a bug in that.. :) > > > Regards > Alex > www.tilogeo.com > > On Mon, Nov 9, 2015 at 9:34 PM, Alex Muir <ale...@gm...> wrote: > >> Hi martynas, >> >> Sorry sent that last one by accident.. >> >> I get the same result with the following, encoding the url. >> >> URLENCODE=$(cat $1?named-graph-uri=$2 | xxd -plain | tr -d '\n' | sed >> 's/\(..\)/%\1/g') >> curl -H "Accept: application/rdf+xml" $URLENCODED -o $3/$4.rdf >> >> >> >> Regards >> Alex >> www.tilogeo.com >> >> On Mon, Nov 9, 2015 at 9:31 PM, Alex Muir <ale...@gm...> wrote: >> >>> I get the same result >>> >>> >>> >>> Regards >>> Alex >>> www.tilogeo.com >>> >>> On Mon, Nov 9, 2015 at 6:28 PM, Martynas Jusevičius < >>> mar...@gr...> wrote: >>> >>>> Are your query parameters percent-encoded? >>>> https://en.wikipedia.org/wiki/Percent-encoding >>>> >>>> On Mon, Nov 9, 2015 at 7:11 PM, Alex Muir <ale...@gm...> >>>> wrote: >>>> >>>>> Hi Bryan, >>>>> >>>>> I've tried that and a number of methods. On export though I get data >>>>> that I guess is a description for the service. >>>>> >>>>> Can blazegraph create some specific examples to show how to accomplish >>>>> this using curl? The task is to load an rdf xml file and then export the >>>>> same file using a named graph. >>>>> >>>>> I'm evaluating the system for a large client and have completed this >>>>> task for other systems but I'm not clear on how to do this with the given >>>>> documentation. >>>>> >>>>> [exec] <rdf:RDF >>>>> [exec] xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"> >>>>> [exec] >>>>> [exec] <rdf:Description rdf:nodeID="service"> >>>>> [exec] <rdf:type rdf:resource="http://www.w3.org/ns/sparql-service-description#Service"/> >>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/sparql"/> >>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/LBS/sparql"/> >>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL10Query"/> >>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Query"/> >>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Update"/> >>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#BasicFederatedQuery"/> >>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#UnionDefaultGraph"/> >>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.bigdata.com/rdf#/features/KB/Mode/Quads"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://sw.deri.org/2008/07/n-quads/#n-quads"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>> [exec] <entailmentRegime xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/entailment/Simple"/> >>>>> [exec] <defaultDataset xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:nodeID="defaultDataset"/> >>>>> [exec] </rdf:Description> >>>>> >>>>> >>>>> >>>>> >>>>> Regards >>>>> Alex >>>>> www.tilogeo.com >>>>> >>>>> On Mon, Nov 9, 2015 at 1:46 PM, Bryan Thompson <br...@sy...> >>>>> wrote: >>>>> >>>>>> Alex, >>>>>> >>>>>> I believe that you should be using the parameters defined at [1] for >>>>>> SPARQL UPDATE. Notably, replace ?c=... with >>>>>> using-named-graph-uriSpecify zero or more named graphs for this the >>>>>> update request (protocol option with the same semantics as USING NAMED). >>>>>> >>>>>> This is per the SPARQL UPDATE specification. >>>>>> >>>>>> Thanks, >>>>>> Bryan >>>>>> >>>>>> [1] >>>>>> https://wiki.blazegraph.com/wiki/index.php/REST_API#UPDATE_.28SPARQL_1.1_UPDATE.29 >>>>>> >>>>>> ---- >>>>>> Bryan Thompson >>>>>> Chief Scientist & Founder >>>>>> SYSTAP, LLC >>>>>> 4501 Tower Road >>>>>> Greensboro, NC 27410 >>>>>> br...@sy... >>>>>> http://blazegraph.com >>>>>> http://blog.blazegraph.com >>>>>> >>>>>> Blazegraph™ <http://www.blazegraph.com/> is our ultra >>>>>> high-performance graph database that supports both RDF/SPARQL and >>>>>> Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU >>>>>> acceleration using our disruptive technology to accelerate >>>>>> data-parallel graph analytics and graph query. >>>>>> >>>>>> CONFIDENTIALITY NOTICE: This email and its contents and attachments >>>>>> are for the sole use of the intended recipient(s) and are confidential or >>>>>> proprietary to SYSTAP. Any unauthorized review, use, disclosure, >>>>>> dissemination or copying of this email or its contents or attachments is >>>>>> prohibited. If you have received this communication in error, please notify >>>>>> the sender by reply email and permanently delete all copies of the email >>>>>> and its contents and attachments. >>>>>> >>>>>> On Sun, Nov 8, 2015 at 1:49 PM, Alex Muir <ale...@gm...> >>>>>> wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> Using the REST API how do I export the same data file that I uploaded? >>>>>>> >>>>>>> I'm unclear with the BLAZEGRAPH REST API on the method to associate a named graph on upload and export that same named graph. >>>>>>> >>>>>>> With the following >>>>>>> >>>>>>> curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf >>>>>>> http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz >>>>>>> >>>>>>> curl -X POST http://62.59.40.122:9999/bigdata/sparql >>>>>>> --data-urlencode 'query=named-graph-uri http://abc.com/id/graph/xyz' -H 'Accept: application/rdf+xml" | gzip > data.rdf.gz >>>>>>> >>>>>>> I get data exported but not the same large file that I inserted. >>>>>>> >>>>>>> Regards >>>>>>> Alex >>>>>>> www.tilogeo.com >>>>>>> >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Bigdata-developers mailing list >>>>>>> Big...@li... >>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>> >>>>>>> >>>>>> >>>>> >>>>> >>>>> ------------------------------------------------------------------------------ >>>>> Presto, an open source distributed SQL query engine for big data, >>>>> initially >>>>> developed by Facebook, enables you to easily query your data on Hadoop >>>>> in a >>>>> more interactive manner. Teradata is also now providing full enterprise >>>>> support for Presto. Download a free open source copy now. >>>>> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >>>>> _______________________________________________ >>>>> Bigdata-developers mailing list >>>>> Big...@li... >>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>> >>>>> >>>> >>> >> > |
From: Martynas J. <mar...@gr...> - 2015-11-09 21:44:53
|
*Note that.. On Mon, Nov 9, 2015 at 10:39 PM, Martynas Jusevičius <mar...@gr...> wrote: > Not sure what the command line does, better if you send the full request > URI. > > Not that you only have to encode the components, such as querystring > params/values, not the whole URI. > > On Mon, Nov 9, 2015 at 10:37 PM, Alex Muir <ale...@gm...> wrote: > >> oops.. actually sorry that was not true... had a bug in that.. :) >> >> >> Regards >> Alex >> www.tilogeo.com >> >> On Mon, Nov 9, 2015 at 9:34 PM, Alex Muir <ale...@gm...> wrote: >> >>> Hi martynas, >>> >>> Sorry sent that last one by accident.. >>> >>> I get the same result with the following, encoding the url. >>> >>> URLENCODE=$(cat $1?named-graph-uri=$2 | xxd -plain | tr -d '\n' | sed >>> 's/\(..\)/%\1/g') >>> curl -H "Accept: application/rdf+xml" $URLENCODED -o $3/$4.rdf >>> >>> >>> >>> Regards >>> Alex >>> www.tilogeo.com >>> >>> On Mon, Nov 9, 2015 at 9:31 PM, Alex Muir <ale...@gm...> wrote: >>> >>>> I get the same result >>>> >>>> >>>> >>>> Regards >>>> Alex >>>> www.tilogeo.com >>>> >>>> On Mon, Nov 9, 2015 at 6:28 PM, Martynas Jusevičius < >>>> mar...@gr...> wrote: >>>> >>>>> Are your query parameters percent-encoded? >>>>> https://en.wikipedia.org/wiki/Percent-encoding >>>>> >>>>> On Mon, Nov 9, 2015 at 7:11 PM, Alex Muir <ale...@gm...> >>>>> wrote: >>>>> >>>>>> Hi Bryan, >>>>>> >>>>>> I've tried that and a number of methods. On export though I get data >>>>>> that I guess is a description for the service. >>>>>> >>>>>> Can blazegraph create some specific examples to show how to >>>>>> accomplish this using curl? The task is to load an rdf xml file and then >>>>>> export the same file using a named graph. >>>>>> >>>>>> I'm evaluating the system for a large client and have completed this >>>>>> task for other systems but I'm not clear on how to do this with the given >>>>>> documentation. >>>>>> >>>>>> [exec] <rdf:RDF >>>>>> [exec] xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"> >>>>>> [exec] >>>>>> [exec] <rdf:Description rdf:nodeID="service"> >>>>>> [exec] <rdf:type rdf:resource="http://www.w3.org/ns/sparql-service-description#Service"/> >>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/sparql"/> >>>>>> [exec] <endpoint xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://52.89.40.122:9999/bigdata/LBS/sparql"/> >>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL10Query"/> >>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Query"/> >>>>>> [exec] <supportedLanguage xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#SPARQL11Update"/> >>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#BasicFederatedQuery"/> >>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/sparql-service-description#UnionDefaultGraph"/> >>>>>> [exec] <feature xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.bigdata.com/rdf#/features/KB/Mode/Quads"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://sw.deri.org/2008/07/n-quads/#n-quads"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>> [exec] <inputFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/RDF_XML"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N-Triples"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/Turtle"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/N3"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.wiwiss.fu-berlin.de/suhl/bizer/TriG/Spec/"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_XML"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_JSON"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_CSV"/> >>>>>> [exec] <resultFormat xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/formats/SPARQL_Results_TSV"/> >>>>>> [exec] <entailmentRegime xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:resource="http://www.w3.org/ns/entailment/Simple"/> >>>>>> [exec] <defaultDataset xmlns="http://www.w3.org/ns/sparql-service-description#" rdf:nodeID="defaultDataset"/> >>>>>> [exec] </rdf:Description> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Regards >>>>>> Alex >>>>>> www.tilogeo.com >>>>>> >>>>>> On Mon, Nov 9, 2015 at 1:46 PM, Bryan Thompson <br...@sy...> >>>>>> wrote: >>>>>> >>>>>>> Alex, >>>>>>> >>>>>>> I believe that you should be using the parameters defined at [1] for >>>>>>> SPARQL UPDATE. Notably, replace ?c=... with >>>>>>> using-named-graph-uriSpecify zero or more named graphs for this the >>>>>>> update request (protocol option with the same semantics as USING NAMED). >>>>>>> >>>>>>> This is per the SPARQL UPDATE specification. >>>>>>> >>>>>>> Thanks, >>>>>>> Bryan >>>>>>> >>>>>>> [1] >>>>>>> https://wiki.blazegraph.com/wiki/index.php/REST_API#UPDATE_.28SPARQL_1.1_UPDATE.29 >>>>>>> >>>>>>> ---- >>>>>>> Bryan Thompson >>>>>>> Chief Scientist & Founder >>>>>>> SYSTAP, LLC >>>>>>> 4501 Tower Road >>>>>>> Greensboro, NC 27410 >>>>>>> br...@sy... >>>>>>> http://blazegraph.com >>>>>>> http://blog.blazegraph.com >>>>>>> >>>>>>> Blazegraph™ <http://www.blazegraph.com/> is our ultra >>>>>>> high-performance graph database that supports both RDF/SPARQL and >>>>>>> Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU >>>>>>> acceleration using our disruptive technology to accelerate >>>>>>> data-parallel graph analytics and graph query. >>>>>>> >>>>>>> CONFIDENTIALITY NOTICE: This email and its contents and >>>>>>> attachments are for the sole use of the intended recipient(s) and are >>>>>>> confidential or proprietary to SYSTAP. Any unauthorized review, use, >>>>>>> disclosure, dissemination or copying of this email or its contents or >>>>>>> attachments is prohibited. If you have received this communication in >>>>>>> error, please notify the sender by reply email and permanently delete all >>>>>>> copies of the email and its contents and attachments. >>>>>>> >>>>>>> On Sun, Nov 8, 2015 at 1:49 PM, Alex Muir <ale...@gm...> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> Using the REST API how do I export the same data file that I uploaded? >>>>>>>> >>>>>>>> I'm unclear with the BLAZEGRAPH REST API on the method to associate a named graph on upload and export that same named graph. >>>>>>>> >>>>>>>> With the following >>>>>>>> >>>>>>>> curl -X POST -H 'Content-Type:application/xml' --data-binary @data.rdf >>>>>>>> http://62.59.40.122:9999/bigdata/sparql?c=http://abc.com/id/graph/xyz >>>>>>>> >>>>>>>> curl -X POST http://62.59.40.122:9999/bigdata/sparql >>>>>>>> --data-urlencode 'query=named-graph-uri http://abc.com/id/graph/xyz' -H 'Accept: application/rdf+xml" | gzip > data.rdf.gz >>>>>>>> >>>>>>>> I get data exported but not the same large file that I inserted. >>>>>>>> >>>>>>>> Regards >>>>>>>> Alex >>>>>>>> www.tilogeo.com >>>>>>>> >>>>>>>> >>>>>>>> ------------------------------------------------------------------------------ >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Bigdata-developers mailing list >>>>>>>> Big...@li... >>>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> ------------------------------------------------------------------------------ >>>>>> Presto, an open source distributed SQL query engine for big data, >>>>>> initially >>>>>> developed by Facebook, enables you to easily query your data on >>>>>> Hadoop in a >>>>>> more interactive manner. Teradata is also now providing full >>>>>> enterprise >>>>>> support for Presto. Download a free open source copy now. >>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=250295911&iu=/4140 >>>>>> _______________________________________________ >>>>>> Bigdata-developers mailing list >>>>>> Big...@li... >>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>>>> >>>>>> >>>>> >>>> >>> >> > |