|
From: Jeremy C. <jj...@gm...> - 2016-02-01 15:35:00
|
Also on 1.5.3, with %codes as a solution set with half million URIs binding ?x
What does the error message mean?
Jeremy
Feb 01,2016 07:27:13 PST - ERROR: 73992665 qtp1401132667-16779 com.bigdata.rdf.sail.webapp.BigdataRDFServlet.launderThrowable(BigdataRDFServlet.java:214): cause=java.util.concurrent.ExecutionException: java.util.concurrent.ExecutionException: org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info, query=SPARQL-QUERY: queryStr=select (count(?x) as $cnt)^M
{ INCLUDE %codes^M
}
java.util.concurrent.ExecutionException: java.util.concurrent.ExecutionException: org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at com.bigdata.rdf.sail.webapp.BigdataServlet.submitApiTask(BigdataServlet.java:281)
at com.bigdata.rdf.sail.webapp.QueryServlet.doSparqlQuery(QueryServlet.java:636)
at com.bigdata.rdf.sail.webapp.QueryServlet.doPost(QueryServlet.java:263)
at com.bigdata.rdf.sail.webapp.RESTServlet.doPost(RESTServlet.java:248)
at com.bigdata.rdf.sail.webapp.MultiTenancyServlet.doPost(MultiTenancyServlet.java:138)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:707)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:769)
at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585)
at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)
at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577)
at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)
at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1125)
at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515)
at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1059)
at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215)
at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)
at org.eclipse.jetty.server.Server.handle(Server.java:497)
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:311)
at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:248)
at org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540)
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:610)
at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:539)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.ExecutionException: org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:834)
at com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:653)
at com.bigdata.rdf.task.ApiTaskForIndexManager.call(ApiTaskForIndexManager.java:68)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
... 1 more
Caused by: org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:188)
at info.aduna.iteration.IterationWrapper.hasNext(IterationWrapper.java:68)
at org.openrdf.query.QueryResults.report(QueryResults.java:155)
at org.openrdf.repository.sail.SailTupleQuery.evaluate(SailTupleQuery.java:76)
at com.bigdata.rdf.sail.webapp.BigdataRDFContext$TupleQueryTask.doQuery(BigdataRDFContext.java:1710)
at com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.innerCall(BigdataRDFContext.java:1567)
at com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:1532)
at com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:704)
... 4 more
Caused by: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1523)
at com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator._hasNext(BlockingBuffer.java:1710)
at com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.hasNext(BlockingBuffer.java:1563)
at com.bigdata.striterator.AbstractChunkedResolverator._hasNext(AbstractChunkedResolverator.java:365)
at com.bigdata.striterator.AbstractChunkedResolverator.hasNext(AbstractChunkedResolverator.java:341)
at com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:134)
... 11 more
Caused by: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1454)
... 16 more
Caused by: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:59)
at com.bigdata.rdf.sail.RunningQueryCloseableIterator.close(RunningQueryCloseableIterator.java:73)
at com.bigdata.rdf.sail.RunningQueryCloseableIterator.hasNext(RunningQueryCloseableIterator.java:82)
at com.bigdata.striterator.ChunkedWrappedIterator.hasNext(ChunkedWrappedIterator.java:197)
at com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:222)
at com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:197)
... 4 more
Caused by: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at com.bigdata.util.concurrent.Haltable.get(Haltable.java:273)
at com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:1514)
at com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:104)
at com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:46)
... 9 more
Caused by: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1337)
at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTaskWrapper.run(ChunkedRunningQuery.java:896)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at com.bigdata.concurrent.FutureTaskMon.run(FutureTaskMon.java:63)
at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkFutureTask.run(ChunkedRunningQuery.java:791)
... 3 more
Caused by: java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1317)
... 8 more
Caused by: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at com.bigdata.rwstore.RWStore.getData(RWStore.java:2190)
at com.bigdata.rwstore.RWStore.getData(RWStore.java:1989)
at com.bigdata.rwstore.PSInputStream.<init>(PSInputStream.java:75)
at com.bigdata.rwstore.RWStore.getInputStream(RWStore.java:6463)
at com.bigdata.journal.RWStrategy.getInputStream(RWStrategy.java:846)
at com.bigdata.bop.solutions.SolutionSetStream.get(SolutionSetStream.java:237)
at com.bigdata.rdf.sparql.ast.ssets.SolutionSetManager.getSolutions(SolutionSetManager.java:556)
at com.bigdata.bop.NamedSolutionSetRefUtility.getSolutionSet(NamedSolutionSetRefUtility.java:529)
at com.bigdata.bop.BOpContext.getAlternateSource(BOpContext.java:752)
at com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.getRightSolutions(NestedLoopJoinOp.java:263)
at com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:200)
at com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:166)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1316)
... 8 more
Caused by: java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
at com.bigdata.rwstore.RWStore.getData(RWStore.java:2112)
... 21 more
Caused by: com.bigdata.util.ChecksumError: offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633
at com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711)
at com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526)
at com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200)
at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363)
at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347)
at com.bigdata.util.concurrent.Memoizer$1.call(Memoizer.java:77)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at com.bigdata.util.concurrent.Memoizer.compute(Memoizer.java:92)
at com.bigdata.io.writecache.WriteCacheService.loadRecord(WriteCacheService.java:3468)
at com.bigdata.io.writecache.WriteCacheService.read(WriteCacheService.java:3187)
at com.bigdata.rwstore.RWStore.getData(RWStore.java:2106)
... 21 more
|
|
From: Bryan T. <br...@sy...> - 2016-02-01 15:43:15
|
Typically this indicates an actual disk error. It is attempting to read
data from the backing file. The checksum that was stored is not matched by
the data. The only time I have see this was when there was actually a bad
disk.
Caused by: com.bigdata.util.ChecksumError:
offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633
at
com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711)
at
com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526)
at
com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200)
at
com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363)
at
com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347)
Bryan
----
Bryan Thompson
Chief Scientist & Founder
SYSTAP, LLC
4501 Tower Road
Greensboro, NC 27410
br...@sy...
http://blazegraph.com
http://blog.blazegraph.com
Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance
graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints
APIs. Blazegraph is now available with GPU acceleration using our disruptive
technology to accelerate data-parallel graph analytics and graph query.
CONFIDENTIALITY NOTICE: This email and its contents and attachments are
for the sole use of the intended recipient(s) and are confidential or
proprietary to SYSTAP. Any unauthorized review, use, disclosure,
dissemination or copying of this email or its contents or attachments is
prohibited. If you have received this communication in error, please notify
the sender by reply email and permanently delete all copies of the email
and its contents and attachments.
On Mon, Feb 1, 2016 at 10:34 AM, Jeremy Carroll <jj...@gm...> wrote:
>
> Also on 1.5.3, with %codes as a solution set with half million URIs
> binding ?x
>
> What does the error message mean?
>
> Jeremy
>
>
> Feb 01,2016 07:27:13 PST - ERROR: 73992665 qtp1401132667-16779
> com.bigdata.rdf.sail.webapp.BigdataRDFServlet.launderThrowable(BigdataRDFServlet.java:214):
> cause=java.util.concurrent.ExecutionException:
> java.util.concurrent.ExecutionException:
> org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info, query=SPARQL-QUERY: queryStr=select (count(?x) as $cnt)^M
> { INCLUDE %codes^M
> }
> java.util.concurrent.ExecutionException:
> java.util.concurrent.ExecutionException:
> org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info
> at java.util.concurrent.FutureTask.report(FutureTask.java:122)
> at java.util.concurrent.FutureTask.get(FutureTask.java:192)
> at
> com.bigdata.rdf.sail.webapp.BigdataServlet.submitApiTask(BigdataServlet.java:281)
> at
> com.bigdata.rdf.sail.webapp.QueryServlet.doSparqlQuery(QueryServlet.java:636)
> at
> com.bigdata.rdf.sail.webapp.QueryServlet.doPost(QueryServlet.java:263)
> at
> com.bigdata.rdf.sail.webapp.RESTServlet.doPost(RESTServlet.java:248)
> at
> com.bigdata.rdf.sail.webapp.MultiTenancyServlet.doPost(MultiTenancyServlet.java:138)
> at javax.servlet.http.HttpServlet.service(HttpServlet.java:707)
> at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
> at
> org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:769)
> at
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585)
> at
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)
> at
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577)
> at
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)
> at
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1125)
> at
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515)
> at
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
> at
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1059)
> at
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
> at
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215)
> at
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110)
> at
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)
> at org.eclipse.jetty.server.Server.handle(Server.java:497)
> at
> org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:311)
> at
> org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:248)
> at
> org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540)
> at
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:610)
> at
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:539)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.ExecutionException:
> org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info
> at java.util.concurrent.FutureTask.report(FutureTask.java:122)
> at java.util.concurrent.FutureTask.get(FutureTask.java:192)
> at
> com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:834)
> at
> com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:653)
> at
> com.bigdata.rdf.task.ApiTaskForIndexManager.call(ApiTaskForIndexManager.java:68)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> ... 1 more
> Caused by: org.openrdf.query.QueryEvaluationException:
> java.lang.RuntimeException: java.util.concurrent.ExecutionException:
> java.lang.RuntimeException: java.util.concurrent.ExecutionException:
> java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info
> at
> com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:188)
> at
> info.aduna.iteration.IterationWrapper.hasNext(IterationWrapper.java:68)
> at org.openrdf.query.QueryResults.report(QueryResults.java:155)
> at
> org.openrdf.repository.sail.SailTupleQuery.evaluate(SailTupleQuery.java:76)
> at
> com.bigdata.rdf.sail.webapp.BigdataRDFContext$TupleQueryTask.doQuery(BigdataRDFContext.java:1710)
> at
> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.innerCall(BigdataRDFContext.java:1567)
> at
> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:1532)
> at
> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:704)
> ... 4 more
> Caused by: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info
> at
> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1523)
> at
> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator._hasNext(BlockingBuffer.java:1710)
> at
> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.hasNext(BlockingBuffer.java:1563)
> at
> com.bigdata.striterator.AbstractChunkedResolverator._hasNext(AbstractChunkedResolverator.java:365)
> at
> com.bigdata.striterator.AbstractChunkedResolverator.hasNext(AbstractChunkedResolverator.java:341)
> at
> com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:134)
> ... 11 more
> Caused by: java.util.concurrent.ExecutionException:
> java.lang.RuntimeException: java.util.concurrent.ExecutionException:
> java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info
> at java.util.concurrent.FutureTask.report(FutureTask.java:122)
> at java.util.concurrent.FutureTask.get(FutureTask.java:192)
> at
> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1454)
> ... 16 more
> Caused by: java.lang.RuntimeException:
> java.util.concurrent.ExecutionException: java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info
> at
> com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:59)
> at
> com.bigdata.rdf.sail.RunningQueryCloseableIterator.close(RunningQueryCloseableIterator.java:73)
> at
> com.bigdata.rdf.sail.RunningQueryCloseableIterator.hasNext(RunningQueryCloseableIterator.java:82)
> at
> com.bigdata.striterator.ChunkedWrappedIterator.hasNext(ChunkedWrappedIterator.java:197)
> at
> com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:222)
> at
> com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:197)
> ... 4 more
> Caused by: java.util.concurrent.ExecutionException: java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info
> at com.bigdata.util.concurrent.Haltable.get(Haltable.java:273)
> at
> com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:1514)
> at
> com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:104)
> at
> com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:46)
> ... 9 more
> Caused by: java.lang.Exception:
> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null},
> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException:
> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from
> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache
> debug info
> at
> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1337)
> at
> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTaskWrapper.run(ChunkedRunningQuery.java:896)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at com.bigdata.concurrent.FutureTaskMon.run(FutureTaskMon.java:63)
> at
> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkFutureTask.run(ChunkedRunningQuery.java:791)
> ... 3 more
> Caused by: java.util.concurrent.ExecutionException:
> java.lang.RuntimeException: addr=-403683228 :
> cause=java.lang.IllegalStateException: Error reading from WriteCache addr:
> 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
> at java.util.concurrent.FutureTask.report(FutureTask.java:122)
> at java.util.concurrent.FutureTask.get(FutureTask.java:192)
> at
> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1317)
> ... 8 more
> Caused by: java.lang.RuntimeException: addr=-403683228 :
> cause=java.lang.IllegalStateException: Error reading from WriteCache addr:
> 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2190)
> at com.bigdata.rwstore.RWStore.getData(RWStore.java:1989)
> at com.bigdata.rwstore.PSInputStream.<init>(PSInputStream.java:75)
> at com.bigdata.rwstore.RWStore.getInputStream(RWStore.java:6463)
> at
> com.bigdata.journal.RWStrategy.getInputStream(RWStrategy.java:846)
> at
> com.bigdata.bop.solutions.SolutionSetStream.get(SolutionSetStream.java:237)
> at
> com.bigdata.rdf.sparql.ast.ssets.SolutionSetManager.getSolutions(SolutionSetManager.java:556)
> at
> com.bigdata.bop.NamedSolutionSetRefUtility.getSolutionSet(NamedSolutionSetRefUtility.java:529)
> at
> com.bigdata.bop.BOpContext.getAlternateSource(BOpContext.java:752)
> at
> com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.getRightSolutions(NestedLoopJoinOp.java:263)
> at
> com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:200)
> at
> com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:166)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1316)
> ... 8 more
> Caused by: java.lang.IllegalStateException: Error reading from WriteCache
> addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info
> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2112)
> ... 21 more
> Caused by: com.bigdata.util.ChecksumError:
> offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633
> at
> com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711)
> at
> com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526)
> at
> com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200)
> at
> com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363)
> at
> com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347)
> at com.bigdata.util.concurrent.Memoizer$1.call(Memoizer.java:77)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at com.bigdata.util.concurrent.Memoizer.compute(Memoizer.java:92)
> at
> com.bigdata.io.writecache.WriteCacheService.loadRecord(WriteCacheService.java:3468)
> at
> com.bigdata.io.writecache.WriteCacheService.read(WriteCacheService.java:3187)
> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2106)
> ... 21 more
>
>
>
> ------------------------------------------------------------------------------
> Site24x7 APM Insight: Get Deep Visibility into Application Performance
> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
> Monitor end-to-end web transactions and take corrective actions now
> Troubleshoot faster and improve end-user experience. Signup Now!
> http://pubads.g.doubleclick.net/gampad/clk?id=267308311&iu=/4140
> _______________________________________________
> Bigdata-developers mailing list
> Big...@li...
> https://lists.sourceforge.net/lists/listinfo/bigdata-developers
>
>
|
|
From: Matthew R. <mr...@ca...> - 2016-02-01 18:11:32
|
I saw this error this morning as well after testing against the 2.0.0 release code. Was running code from around 1.5.0 previously. Caused by: com.bigdata.util.ChecksumError: offset=18124800,nbytes=4044,expected=0,actual=1696870497 at com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3761) ~[bigdata-core-2.0.0.jar:na] at com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3576) ~[bigdata-core-2.0.0.jar:na] at com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) ~[bigdata-core-2.0.0.jar:na] at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3413) ~[bigdata-core-2.0.0.jar:na] at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3397) ~[bigdata-core-2.0.0.jar:na] at com.bigdata.util.concurrent.Memoizer$1.call(Memoizer.java:77) ~[bigdata-core-2.0.0.jar:na] at java.util.concurrent.FutureTask.run(FutureTask.java:266) [na:1.8.0_60] at com.bigdata.util.concurrent.Memoizer.compute(Memoizer.java:92) ~[bigdata-core-2.0.0.jar:na] at com.bigdata.io.writecache.WriteCacheService.loadRecord(WriteCacheService.java:3518) ~[bigdata-core-2.0.0.jar:na] at com.bigdata.io.writecache.WriteCacheService.read(WriteCacheService.java:3237) ~[bigdata-core-2.0.0.jar:na] at com.bigdata.rwstore.RWStore.getData(RWStore.java:2052) ~[bigdata-core-2.0.0.jar:na] ... 24 common frames omitted and after reopening the journal file get: java.lang.Error: Two allocators at same address at com.bigdata.rwstore.FixedAllocator.compareTo(FixedAllocator.java:102) at java.util.ComparableTimSort.countRunAndMakeAscending(ComparableTimSort.java:295) at java.util.ComparableTimSort.sort(ComparableTimSort.java:157) at java.util.ComparableTimSort.sort(ComparableTimSort.java:146) at java.util.Arrays.sort(Arrays.java:472) at java.util.Collections.sort(Collections.java:155) at com.bigdata.rwstore.RWStore.readAllocationBlocks(RWStore.java:1682) at com.bigdata.rwstore.RWStore.initfromRootBlock(RWStore.java:1557) at com.bigdata.rwstore.RWStore.<init>(RWStore.java:969) at com.bigdata.journal.RWStrategy.<init>(RWStrategy.java:137) Can't tell exactly what was going on query/update wise when the error occurred. Will let you know if I can reproduce the error again. Matt ------ Original Message ------ From: "Bryan Thompson" <br...@sy...> To: "Jeremy Carroll" <jj...@gm...>; "Martyn Cutcher" <ma...@sy...> Cc: "Big...@li..." <Big...@li...> Sent: 2/1/2016 10:43:04 AM Subject: Re: [Bigdata-developers] " No WriteCache debug info" >Typically this indicates an actual disk error. It is attempting to >read data from the backing file. The checksum that was stored is not >matched by the data. The only time I have see this was when there was >actually a bad disk. > >Caused by: com.bigdata.util.ChecksumError: >offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633 > at >com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711) > at >com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526) > at >com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) > at >com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363) > at >com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347) > >Bryan > >---- >Bryan Thompson >Chief Scientist & Founder >SYSTAP, LLC >4501 Tower Road >Greensboro, NC 27410 >br...@sy... >http://blazegraph.com/ >http://blog.blazegraph.com > >Blazegraph™ is our ultra high-performance graph database that supports >both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now >available with GPU acceleration using our disruptive technology to >accelerate data-parallel graph analytics and graph query. >CONFIDENTIALITY NOTICE: This email and its contents and attachments >are for the sole use of the intended recipient(s) and are confidential >or proprietary to SYSTAP. Any unauthorized review, use, disclosure, >dissemination or copying of this email or its contents or attachments >is prohibited. If you have received this communication in error, please >notify the sender by reply email and permanently delete all copies of >the email and its contents and attachments. > > >On Mon, Feb 1, 2016 at 10:34 AM, Jeremy Carroll <jj...@gm...> >wrote: >> >>Also on 1.5.3, with %codes as a solution set with half million URIs >>binding ?x >> >>What does the error message mean? >> >>Jeremy >> >> >>Feb 01,2016 07:27:13 PST - ERROR: 73992665 qtp1401132667-16779 >>com.bigdata.rdf.sail.webapp.BigdataRDFServlet.launderThrowable(BigdataRDFServlet.java:214): >>cause=java.util.concurrent.ExecutionException: >>java.util.concurrent.ExecutionException: >>org.openrdf.query.QueryEvaluationException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info, query=SPARQL-QUERY: queryStr=select (count(?x) as $cnt)^M >>{ INCLUDE %codes^M >>} >>java.util.concurrent.ExecutionException: >>java.util.concurrent.ExecutionException: >>org.openrdf.query.QueryEvaluationException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at >>com.bigdata.rdf.sail.webapp.BigdataServlet.submitApiTask(BigdataServlet.java:281) >> at >>com.bigdata.rdf.sail.webapp.QueryServlet.doSparqlQuery(QueryServlet.java:636) >> at >>com.bigdata.rdf.sail.webapp.QueryServlet.doPost(QueryServlet.java:263) >> at >>com.bigdata.rdf.sail.webapp.RESTServlet.doPost(RESTServlet.java:248) >> at >>com.bigdata.rdf.sail.webapp.MultiTenancyServlet.doPost(MultiTenancyServlet.java:138) >> at >>javax.servlet.http.HttpServlet.service(HttpServlet.java:707) >> at >>javax.servlet.http.HttpServlet.service(HttpServlet.java:790) >> at >>org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:769) >> at >>org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) >> at >>org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) >> at >>org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) >> at >>org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223) >> at >>org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1125) >> at >>org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) >> at >>org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) >> at >>org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1059) >> at >>org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) >> at >>org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215) >> at >>org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110) >> at >>org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) >> at org.eclipse.jetty.server.Server.handle(Server.java:497) >> at >>org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:311) >> at >>org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:248) >> at >>org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) >> at >>org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:610) >> at >>org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:539) >> at java.lang.Thread.run(Thread.java:745) >>Caused by: java.util.concurrent.ExecutionException: >>org.openrdf.query.QueryEvaluationException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at >>com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:834) >> at >>com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:653) >> at >>com.bigdata.rdf.task.ApiTaskForIndexManager.call(ApiTaskForIndexManager.java:68) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at >>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >> at >>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >> ... 1 more >>Caused by: org.openrdf.query.QueryEvaluationException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at >>com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:188) >> at >>info.aduna.iteration.IterationWrapper.hasNext(IterationWrapper.java:68) >> at >>org.openrdf.query.QueryResults.report(QueryResults.java:155) >> at >>org.openrdf.repository.sail.SailTupleQuery.evaluate(SailTupleQuery.java:76) >> at >>com.bigdata.rdf.sail.webapp.BigdataRDFContext$TupleQueryTask.doQuery(BigdataRDFContext.java:1710) >> at >>com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.innerCall(BigdataRDFContext.java:1567) >> at >>com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:1532) >> at >>com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:704) >> ... 4 more >>Caused by: java.lang.RuntimeException: >>java.util.concurrent.ExecutionException: java.lang.RuntimeException: >>java.util.concurrent.ExecutionException: java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at >>com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1523) >> at >>com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator._hasNext(BlockingBuffer.java:1710) >> at >>com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.hasNext(BlockingBuffer.java:1563) >> at >>com.bigdata.striterator.AbstractChunkedResolverator._hasNext(AbstractChunkedResolverator.java:365) >> at >>com.bigdata.striterator.AbstractChunkedResolverator.hasNext(AbstractChunkedResolverator.java:341) >> at >>com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:134) >> ... 11 more >>Caused by: java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: java.util.concurrent.ExecutionException: >>java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at >>com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1454) >> ... 16 more >>Caused by: java.lang.RuntimeException: >>java.util.concurrent.ExecutionException: java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at >>com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:59) >> at >>com.bigdata.rdf.sail.RunningQueryCloseableIterator.close(RunningQueryCloseableIterator.java:73) >> at >>com.bigdata.rdf.sail.RunningQueryCloseableIterator.hasNext(RunningQueryCloseableIterator.java:82) >> at >>com.bigdata.striterator.ChunkedWrappedIterator.hasNext(ChunkedWrappedIterator.java:197) >> at >>com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:222) >> at >>com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:197) >> ... 4 more >>Caused by: java.util.concurrent.ExecutionException: >>java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at com.bigdata.util.concurrent.Haltable.get(Haltable.java:273) >> at >>com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:1514) >> at >>com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:104) >> at >>com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:46) >> ... 9 more >>Caused by: java.lang.Exception: >>task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >>cause=java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at >>com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1337) >> at >>com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTaskWrapper.run(ChunkedRunningQuery.java:896) >> at >>java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at >>com.bigdata.concurrent.FutureTaskMon.run(FutureTaskMon.java:63) >> at >>com.bigdata.bop.engine.ChunkedRunningQuery$ChunkFutureTask.run(ChunkedRunningQuery.java:791) >> ... 3 more >>Caused by: java.util.concurrent.ExecutionException: >>java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at >>com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1317) >> ... 8 more >>Caused by: java.lang.RuntimeException: addr=-403683228 : >>cause=java.lang.IllegalStateException: Error reading from WriteCache >>addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug >>info >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2190) >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:1989) >> at >>com.bigdata.rwstore.PSInputStream.<init>(PSInputStream.java:75) >> at >>com.bigdata.rwstore.RWStore.getInputStream(RWStore.java:6463) >> at >>com.bigdata.journal.RWStrategy.getInputStream(RWStrategy.java:846) >> at >>com.bigdata.bop.solutions.SolutionSetStream.get(SolutionSetStream.java:237) >> at >>com.bigdata.rdf.sparql.ast.ssets.SolutionSetManager.getSolutions(SolutionSetManager.java:556) >> at >>com.bigdata.bop.NamedSolutionSetRefUtility.getSolutionSet(NamedSolutionSetRefUtility.java:529) >> at >>com.bigdata.bop.BOpContext.getAlternateSource(BOpContext.java:752) >> at >>com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.getRightSolutions(NestedLoopJoinOp.java:263) >> at >>com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:200) >> at >>com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:166) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at >>com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1316) >> ... 8 more >>Caused by: java.lang.IllegalStateException: Error reading from >>WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No >>WriteCache debug info >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2112) >> ... 21 more >>Caused by: com.bigdata.util.ChecksumError: >>offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633 >> at >>com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711) >> at >>com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526) >> at >>com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) >> at >>com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363) >> at >>com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347) >> at >>com.bigdata.util.concurrent.Memoizer$1.call(Memoizer.java:77) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at >>com.bigdata.util.concurrent.Memoizer.compute(Memoizer.java:92) >> at >>com.bigdata.io.writecache.WriteCacheService.loadRecord(WriteCacheService.java:3468) >> at >>com.bigdata.io.writecache.WriteCacheService.read(WriteCacheService.java:3187) >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2106) >> ... 21 more >> >> >>------------------------------------------------------------------------------ >>Site24x7 APM Insight: Get Deep Visibility into Application Performance >>APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>Monitor end-to-end web transactions and take corrective actions now >>Troubleshoot faster and improve end-user experience. Signup Now! >>http://pubads.g.doubleclick.net/gampad/clk?id=267308311&iu=/4140 >>_______________________________________________ >>Bigdata-developers mailing list >>Big...@li... >>https://lists.sourceforge.net/lists/listinfo/bigdata-developers >> > |
|
From: Bryan T. <br...@sy...> - 2016-02-01 18:02:40
|
Interesting. There were a few RWStore changes that were held back to get more experience with them. One of the ones that did go in is: - https://jira.blazegraph.com/browse/BLZG-1667 (Growth in RWStore.alloc() cumulative time) I do not see any reason offhand why this might be related. We did hold back some changes to accelerate deferred free processing. This change is mainly of benefit to very large stores and very large commits. Thanks, Bryan ---- Bryan Thompson Chief Scientist & Founder SYSTAP, LLC 4501 Tower Road Greensboro, NC 27410 br...@sy... http://blazegraph.com http://blog.blazegraph.com Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Mon, Feb 1, 2016 at 12:48 PM, Matthew Roy <mr...@ca...> wrote: > I saw this error this morning as well after testing against the 2.0.0 > release code. Was running code from around 1.5.0 previously. > > Caused by: com.bigdata.util.ChecksumError: > offset=18124800,nbytes=4044,expected=0,actual=1696870497 > at > com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3761) > ~[bigdata-core-2.0.0.jar:na] > at > com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3576) > ~[bigdata-core-2.0.0.jar:na] > at > com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) > ~[bigdata-core-2.0.0.jar:na] > at > com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3413) > ~[bigdata-core-2.0.0.jar:na] > at > com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3397) > ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.util.concurrent.Memoizer$1.call(Memoizer.java:77) > ~[bigdata-core-2.0.0.jar:na] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) [na:1.8.0_60] > at com.bigdata.util.concurrent.Memoizer.compute(Memoizer.java:92) > ~[bigdata-core-2.0.0.jar:na] > at > com.bigdata.io.writecache.WriteCacheService.loadRecord(WriteCacheService.java:3518) > ~[bigdata-core-2.0.0.jar:na] > at > com.bigdata.io.writecache.WriteCacheService.read(WriteCacheService.java:3237) > ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.rwstore.RWStore.getData(RWStore.java:2052) > ~[bigdata-core-2.0.0.jar:na] > ... 24 common frames omitted > and after reopening the journal file get: > > java.lang.Error: Two allocators at same address > at com.bigdata.rwstore.FixedAllocator.compareTo(FixedAllocator.java:102) > at > java.util.ComparableTimSort.countRunAndMakeAscending(ComparableTimSort.java:295) > at java.util.ComparableTimSort.sort(ComparableTimSort.java:157) > at java.util.ComparableTimSort.sort(ComparableTimSort.java:146) > at java.util.Arrays.sort(Arrays.java:472) > at java.util.Collections.sort(Collections.java:155) > at com.bigdata.rwstore.RWStore.readAllocationBlocks(RWStore.java:1682) > at com.bigdata.rwstore.RWStore.initfromRootBlock(RWStore.java:1557) > at com.bigdata.rwstore.RWStore.<init>(RWStore.java:969) > at com.bigdata.journal.RWStrategy.<init>(RWStrategy.java:137) > Can't tell exactly what was going on query/update wise when the error > occurred. > Will let you know if I can reproduce the error again. > Matt > > ------ Original Message ------ > From: "Bryan Thompson" <br...@sy...> > To: "Jeremy Carroll" <jj...@gm...>; "Martyn Cutcher" < > ma...@sy...> > Cc: "Big...@li..." < > Big...@li...> > Sent: 2/1/2016 10:43:04 AM > Subject: Re: [Bigdata-developers] " No WriteCache debug info" > > > Typically this indicates an actual disk error. It is attempting to read > data from the backing file. The checksum that was stored is not matched by > the data. The only time I have see this was when there was actually a bad > disk. > > Caused by: com.bigdata.util.ChecksumError: > offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633 > at > com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711) > at > com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526) > at > com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) > at > com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363) > at > com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347) > > Bryan > > ---- > Bryan Thompson > Chief Scientist & Founder > SYSTAP, LLC > 4501 Tower Road > Greensboro, NC 27410 > br...@sy... > http://blazegraph.com/ > http://blog.blazegraph.com > > Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance > graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints > APIs. Blazegraph is now available with GPU acceleration using our disruptive > technology to accelerate data-parallel graph analytics and graph query. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP. Any unauthorized review, use, disclosure, > dissemination or copying of this email or its contents or attachments is > prohibited. If you have received this communication in error, please notify > the sender by reply email and permanently delete all copies of the email > and its contents and attachments. > > On Mon, Feb 1, 2016 at 10:34 AM, Jeremy Carroll <jj...@gm...> wrote: > >> >> Also on 1.5.3, with %codes as a solution set with half million URIs >> binding ?x >> >> What does the error message mean? >> >> Jeremy >> >> >> Feb 01,2016 07:27:13 PST - ERROR: 73992665 qtp1401132667-16779 >> com.bigdata.rdf.sail.webapp.BigdataRDFServlet.launderThrowable(BigdataRDFServlet.java:214): >> cause=java.util.concurrent.ExecutionException: >> java.util.concurrent.ExecutionException: >> org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info, query=SPARQL-QUERY: queryStr=select (count(?x) as $cnt)^M >> { INCLUDE %codes^M >> } >> java.util.concurrent.ExecutionException: >> java.util.concurrent.ExecutionException: >> org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at >> com.bigdata.rdf.sail.webapp.BigdataServlet.submitApiTask(BigdataServlet.java:281) >> at >> com.bigdata.rdf.sail.webapp.QueryServlet.doSparqlQuery(QueryServlet.java:636) >> at >> com.bigdata.rdf.sail.webapp.QueryServlet.doPost(QueryServlet.java:263) >> at >> com.bigdata.rdf.sail.webapp.RESTServlet.doPost(RESTServlet.java:248) >> at >> com.bigdata.rdf.sail.webapp.MultiTenancyServlet.doPost(MultiTenancyServlet.java:138) >> at javax.servlet.http.HttpServlet.service(HttpServlet.java:707) >> at javax.servlet.http.HttpServlet.service(HttpServlet.java:790) >> at >> org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:769) >> at >> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) >> at >> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) >> at >> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) >> at >> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223) >> at >> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1125) >> at >> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) >> at >> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) >> at >> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1059) >> at >> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) >> at >> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215) >> at >> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110) >> at >> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) >> at org.eclipse.jetty.server.Server.handle(Server.java:497) >> at >> org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:311) >> at >> org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:248) >> at >> org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) >> at >> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:610) >> at >> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:539) >> at java.lang.Thread.run(Thread.java:745) >> Caused by: java.util.concurrent.ExecutionException: >> org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at >> com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:834) >> at >> com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:653) >> at >> com.bigdata.rdf.task.ApiTaskForIndexManager.call(ApiTaskForIndexManager.java:68) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >> ... 1 more >> Caused by: org.openrdf.query.QueryEvaluationException: >> java.lang.RuntimeException: java.util.concurrent.ExecutionException: >> java.lang.RuntimeException: java.util.concurrent.ExecutionException: >> java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info >> at >> com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:188) >> at >> info.aduna.iteration.IterationWrapper.hasNext(IterationWrapper.java:68) >> at org.openrdf.query.QueryResults.report(QueryResults.java:155) >> at >> org.openrdf.repository.sail.SailTupleQuery.evaluate(SailTupleQuery.java:76) >> at >> com.bigdata.rdf.sail.webapp.BigdataRDFContext$TupleQueryTask.doQuery(BigdataRDFContext.java:1710) >> at >> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.innerCall(BigdataRDFContext.java:1567) >> at >> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:1532) >> at >> com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:704) >> ... 4 more >> Caused by: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info >> at >> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1523) >> at >> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator._hasNext(BlockingBuffer.java:1710) >> at >> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.hasNext(BlockingBuffer.java:1563) >> at >> com.bigdata.striterator.AbstractChunkedResolverator._hasNext(AbstractChunkedResolverator.java:365) >> at >> com.bigdata.striterator.AbstractChunkedResolverator.hasNext(AbstractChunkedResolverator.java:341) >> at >> com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:134) >> ... 11 more >> Caused by: java.util.concurrent.ExecutionException: >> java.lang.RuntimeException: java.util.concurrent.ExecutionException: >> java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at >> com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1454) >> ... 16 more >> Caused by: java.lang.RuntimeException: >> java.util.concurrent.ExecutionException: java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info >> at >> com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:59) >> at >> com.bigdata.rdf.sail.RunningQueryCloseableIterator.close(RunningQueryCloseableIterator.java:73) >> at >> com.bigdata.rdf.sail.RunningQueryCloseableIterator.hasNext(RunningQueryCloseableIterator.java:82) >> at >> com.bigdata.striterator.ChunkedWrappedIterator.hasNext(ChunkedWrappedIterator.java:197) >> at >> com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:222) >> at >> com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:197) >> ... 4 more >> Caused by: java.util.concurrent.ExecutionException: java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info >> at com.bigdata.util.concurrent.Haltable.get(Haltable.java:273) >> at >> com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:1514) >> at >> com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:104) >> at >> com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:46) >> ... 9 more >> Caused by: java.lang.Exception: >> task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, >> cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: >> addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from >> WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache >> debug info >> at >> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1337) >> at >> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTaskWrapper.run(ChunkedRunningQuery.java:896) >> at >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at com.bigdata.concurrent.FutureTaskMon.run(FutureTaskMon.java:63) >> at >> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkFutureTask.run(ChunkedRunningQuery.java:791) >> ... 3 more >> Caused by: java.util.concurrent.ExecutionException: >> java.lang.RuntimeException: addr=-403683228 : >> cause=java.lang.IllegalStateException: Error reading from WriteCache addr: >> 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at >> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1317) >> ... 8 more >> Caused by: java.lang.RuntimeException: addr=-403683228 : >> cause=java.lang.IllegalStateException: Error reading from WriteCache addr: >> 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2190) >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:1989) >> at com.bigdata.rwstore.PSInputStream.<init>(PSInputStream.java:75) >> at com.bigdata.rwstore.RWStore.getInputStream(RWStore.java:6463) >> at >> com.bigdata.journal.RWStrategy.getInputStream(RWStrategy.java:846) >> at >> com.bigdata.bop.solutions.SolutionSetStream.get(SolutionSetStream.java:237) >> at >> com.bigdata.rdf.sparql.ast.ssets.SolutionSetManager.getSolutions(SolutionSetManager.java:556) >> at >> com.bigdata.bop.NamedSolutionSetRefUtility.getSolutionSet(NamedSolutionSetRefUtility.java:529) >> at >> com.bigdata.bop.BOpContext.getAlternateSource(BOpContext.java:752) >> at >> com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.getRightSolutions(NestedLoopJoinOp.java:263) >> at >> com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:200) >> at >> com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:166) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at >> com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1316) >> ... 8 more >> Caused by: java.lang.IllegalStateException: Error reading from WriteCache >> addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2112) >> ... 21 more >> Caused by: com.bigdata.util.ChecksumError: >> offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633 >> at >> com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711) >> at >> com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526) >> at >> com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) >> at >> com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363) >> at >> com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347) >> at com.bigdata.util.concurrent.Memoizer$1.call(Memoizer.java:77) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at com.bigdata.util.concurrent.Memoizer.compute(Memoizer.java:92) >> at >> com.bigdata.io.writecache.WriteCacheService.loadRecord(WriteCacheService.java:3468) >> at >> com.bigdata.io.writecache.WriteCacheService.read(WriteCacheService.java:3187) >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2106) >> ... 21 more >> >> >> >> ------------------------------------------------------------------------------ >> Site24x7 APM Insight: Get Deep Visibility into Application Performance >> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >> Monitor end-to-end web transactions and take corrective actions now >> Troubleshoot faster and improve end-user experience. Signup Now! >> http://pubads.g.doubleclick.net/gampad/clk?id=267308311&iu=/4140 >> _______________________________________________ >> Bigdata-developers mailing list >> Big...@li... >> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >> >> > |
|
From: Jeremy C. <jj...@gm...> - 2016-02-01 19:57:54
|
My experience is that blazegraph checks for disk integrity much more than most software, and after an error of this sort from blazegraph, sooner or later other things stop working too, because the h/w is defective. Jeremy > On Feb 1, 2016, at 9:48 AM, Matthew Roy <mr...@ca...> wrote: > > I saw this error this morning as well after testing against the 2.0.0 release code. Was running code from around 1.5.0 previously. > > Caused by: com.bigdata.util.ChecksumError: offset=18124800,nbytes=4044,expected=0,actual=1696870497 > at com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3761) ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3576) ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3413) ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3397) ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.util.concurrent.Memoizer$1.call(Memoizer.java:77) ~[bigdata-core-2.0.0.jar:na] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) [na:1.8.0_60] > at com.bigdata.util.concurrent.Memoizer.compute(Memoizer.java:92) ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.io.writecache.WriteCacheService.loadRecord(WriteCacheService.java:3518) ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.io.writecache.WriteCacheService.read(WriteCacheService.java:3237) ~[bigdata-core-2.0.0.jar:na] > at com.bigdata.rwstore.RWStore.getData(RWStore.java:2052) ~[bigdata-core-2.0.0.jar:na] > ... 24 common frames omitted > and after reopening the journal file get: > > java.lang.Error: Two allocators at same address > at com.bigdata.rwstore.FixedAllocator.compareTo(FixedAllocator.java:102) > at java.util.ComparableTimSort.countRunAndMakeAscending(ComparableTimSort.java:295) > at java.util.ComparableTimSort.sort(ComparableTimSort.java:157) > at java.util.ComparableTimSort.sort(ComparableTimSort.java:146) > at java.util.Arrays.sort(Arrays.java:472) > at java.util.Collections.sort(Collections.java:155) > at com.bigdata.rwstore.RWStore.readAllocationBlocks(RWStore.java:1682) > at com.bigdata.rwstore.RWStore.initfromRootBlock(RWStore.java:1557) > at com.bigdata.rwstore.RWStore.<init>(RWStore.java:969) > at com.bigdata.journal.RWStrategy.<init>(RWStrategy.java:137) > Can't tell exactly what was going on query/update wise when the error occurred. > Will let you know if I can reproduce the error again. > Matt > > ------ Original Message ------ > From: "Bryan Thompson" <br...@sy... <mailto:br...@sy...>> > To: "Jeremy Carroll" <jj...@gm... <mailto:jj...@gm...>>; "Martyn Cutcher" <ma...@sy... <mailto:ma...@sy...>> > Cc: "Big...@li..." <Big...@li... <mailto:Big...@li...>> > Sent: 2/1/2016 10:43:04 AM > Subject: Re: [Bigdata-developers] " No WriteCache debug info" > >> Typically this indicates an actual disk error. It is attempting to read data from the backing file. The checksum that was stored is not matched by the data. The only time I have see this was when there was actually a bad disk. >> >> Caused by: com.bigdata.util.ChecksumError: offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633 >> at com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711) >> at com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526) >> at com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) >> at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363) >> at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347) >> >> Bryan >> >> ---- >> Bryan Thompson >> Chief Scientist & Founder >> SYSTAP, LLC >> 4501 Tower Road >> Greensboro, NC 27410 >> br...@sy... <mailto:br...@sy...> >> http://blazegraph.com/ <http://blazegraph.com/> >> http://blog.blazegraph.com <http://blog.blazegraph.com/> >> >> Blazegraph™ <http://www.blazegraph.com/> is our ultra high-performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph is now available with GPU acceleration using our disruptive technology to accelerate data-parallel graph analytics and graph query. >> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. >> >> >> >> On Mon, Feb 1, 2016 at 10:34 AM, Jeremy Carroll <jj...@gm... <mailto:jj...@gm...>> wrote: >> >> Also on 1.5.3, with %codes as a solution set with half million URIs binding ?x >> >> What does the error message mean? >> >> Jeremy >> >> >> Feb 01,2016 07:27:13 PST - ERROR: 73992665 qtp1401132667-16779 com.bigdata.rdf.sail.webapp.BigdataRDFServlet.launderThrowable(BigdataRDFServlet.java:214): cause=java.util.concurrent.ExecutionException: java.util.concurrent.ExecutionException: org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info, query=SPARQL-QUERY: queryStr=select (count(?x) as $cnt)^M >> { INCLUDE %codes^M >> } >> java.util.concurrent.ExecutionException: java.util.concurrent.ExecutionException: org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at com.bigdata.rdf.sail.webapp.BigdataServlet.submitApiTask(BigdataServlet.java:281) >> at com.bigdata.rdf.sail.webapp.QueryServlet.doSparqlQuery(QueryServlet.java:636) >> at com.bigdata.rdf.sail.webapp.QueryServlet.doPost(QueryServlet.java:263) >> at com.bigdata.rdf.sail.webapp.RESTServlet.doPost(RESTServlet.java:248) >> at com.bigdata.rdf.sail.webapp.MultiTenancyServlet.doPost(MultiTenancyServlet.java:138) >> at javax.servlet.http.HttpServlet.service(HttpServlet.java:707) >> at javax.servlet.http.HttpServlet.service(HttpServlet.java:790) >> at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:769) >> at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) >> at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) >> at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) >> at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223) >> at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1125) >> at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) >> at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) >> at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1059) >> at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) >> at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215) >> at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110) >> at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) >> at org.eclipse.jetty.server.Server.handle(Server.java:497) >> at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:311) >> at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:248) >> at org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) >> at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:610) >> at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:539) >> at java.lang.Thread.run(Thread.java:745) >> Caused by: java.util.concurrent.ExecutionException: org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:834) >> at com.bigdata.rdf.sail.webapp.QueryServlet$SparqlQueryTask.call(QueryServlet.java:653) >> at com.bigdata.rdf.task.ApiTaskForIndexManager.call(ApiTaskForIndexManager.java:68) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >> ... 1 more >> Caused by: org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:188) >> at info.aduna.iteration.IterationWrapper.hasNext(IterationWrapper.java:68) >> at org.openrdf.query.QueryResults.report(QueryResults.java:155) >> at org.openrdf.repository.sail.SailTupleQuery.evaluate(SailTupleQuery.java:76) >> at com.bigdata.rdf.sail.webapp.BigdataRDFContext$TupleQueryTask.doQuery(BigdataRDFContext.java:1710) >> at com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.innerCall(BigdataRDFContext.java:1567) >> at com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:1532) >> at com.bigdata.rdf.sail.webapp.BigdataRDFContext$AbstractQueryTask.call(BigdataRDFContext.java:704) >> ... 4 more >> Caused by: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1523) >> at com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator._hasNext(BlockingBuffer.java:1710) >> at com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.hasNext(BlockingBuffer.java:1563) >> at com.bigdata.striterator.AbstractChunkedResolverator._hasNext(AbstractChunkedResolverator.java:365) >> at com.bigdata.striterator.AbstractChunkedResolverator.hasNext(AbstractChunkedResolverator.java:341) >> at com.bigdata.rdf.sail.Bigdata2Sesame2BindingSetIterator.hasNext(Bigdata2Sesame2BindingSetIterator.java:134) >> ... 11 more >> Caused by: java.util.concurrent.ExecutionException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at com.bigdata.relation.accesspath.BlockingBuffer$BlockingIterator.checkFuture(BlockingBuffer.java:1454) >> ... 16 more >> Caused by: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:59) >> at com.bigdata.rdf.sail.RunningQueryCloseableIterator.close(RunningQueryCloseableIterator.java:73) >> at com.bigdata.rdf.sail.RunningQueryCloseableIterator.hasNext(RunningQueryCloseableIterator.java:82) >> at com.bigdata.striterator.ChunkedWrappedIterator.hasNext(ChunkedWrappedIterator.java:197) >> at com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:222) >> at com.bigdata.striterator.AbstractChunkedResolverator$ChunkConsumerTask.call(AbstractChunkedResolverator.java:197) >> ... 4 more >> Caused by: java.util.concurrent.ExecutionException: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.util.concurrent.Haltable.get(Haltable.java:273) >> at com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:1514) >> at com.bigdata.bop.engine.AbstractRunningQuery.get(AbstractRunningQuery.java:104) >> at com.bigdata.rdf.sail.RunningQueryCloseableIterator.checkFuture(RunningQueryCloseableIterator.java:46) >> ... 9 more >> Caused by: java.lang.Exception: task=ChunkTask{query=57386638-1f48-4826-966a-84a6b36b5427,bopId=1,partitionId=-1,sinkId=2,altSinkId=null}, cause=java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1337) >> at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTaskWrapper.run(ChunkedRunningQuery.java:896) >> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at com.bigdata.concurrent.FutureTaskMon.run(FutureTaskMon.java:63) >> at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkFutureTask.run(ChunkedRunningQuery.java:791) >> ... 3 more >> Caused by: java.util.concurrent.ExecutionException: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >> at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1317) >> ... 8 more >> Caused by: java.lang.RuntimeException: addr=-403683228 : cause=java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2190) >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:1989) >> at com.bigdata.rwstore.PSInputStream.<init>(PSInputStream.java:75) >> at com.bigdata.rwstore.RWStore.getInputStream(RWStore.java:6463) >> at com.bigdata.journal.RWStrategy.getInputStream(RWStrategy.java:846) >> at com.bigdata.bop.solutions.SolutionSetStream.get(SolutionSetStream.java:237) >> at com.bigdata.rdf.sparql.ast.ssets.SolutionSetManager.getSolutions(SolutionSetManager.java:556) >> at com.bigdata.bop.NamedSolutionSetRefUtility.getSolutionSet(NamedSolutionSetRefUtility.java:529) >> at com.bigdata.bop.BOpContext.getAlternateSource(BOpContext.java:752) >> at com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.getRightSolutions(NestedLoopJoinOp.java:263) >> at com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:200) >> at com.bigdata.bop.join.NestedLoopJoinOp$ChunkTask.call(NestedLoopJoinOp.java:166) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at com.bigdata.bop.engine.ChunkedRunningQuery$ChunkTask.call(ChunkedRunningQuery.java:1316) >> ... 8 more >> Caused by: java.lang.IllegalStateException: Error reading from WriteCache addr: 404849739776 length: 1152, writeCacheDebug: No WriteCache debug info >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2112) >> ... 21 more >> Caused by: com.bigdata.util.ChecksumError: offset=404849739776,nbytes=1156,expected=-402931822,actual=1830389633 >> at com.bigdata.io.writecache.WriteCacheService._readFromLocalDiskIntoNewHeapByteBuffer(WriteCacheService.java:3711) >> at com.bigdata.io.writecache.WriteCacheService._getRecord(WriteCacheService.java:3526) >> at com.bigdata.io.writecache.WriteCacheService.access$2500(WriteCacheService.java:200) >> at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3363) >> at com.bigdata.io.writecache.WriteCacheService$1.compute(WriteCacheService.java:3347) >> at com.bigdata.util.concurrent.Memoizer$1.call(Memoizer.java:77) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> at com.bigdata.util.concurrent.Memoizer.compute(Memoizer.java:92) >> at com.bigdata.io.writecache.WriteCacheService.loadRecord(WriteCacheService.java:3468) >> at com.bigdata.io.writecache.WriteCacheService.read(WriteCacheService.java:3187) >> at com.bigdata.rwstore.RWStore.getData(RWStore.java:2106) >> ... 21 more >> >> >> ------------------------------------------------------------------------------ >> Site24x7 APM Insight: Get Deep Visibility into Application Performance >> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >> Monitor end-to-end web transactions and take corrective actions now >> Troubleshoot faster and improve end-user experience. Signup Now! >> http://pubads.g.doubleclick.net/gampad/clk?id=267308311&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=267308311&iu=/4140> >> _______________________________________________ >> Bigdata-developers mailing list >> Big...@li... <mailto:Big...@li...> >> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> >> >> |
|
From: Joakim S. <joa...@bl...> - 2016-02-15 22:47:19
|
Hello, I am trying to set the logger from the command prompt like this without success: java -server -Xmx4g -Dbigdata.propertyFile=RWStore.properties -Djava.util.logging.config.file=log4j.properties -jar blazegraph.jar Does anyone know how configure the log util when running from the command line? |
|
From: Bryan T. <br...@sy...> - 2016-02-15 23:59:11
|
Blazegraph is configured using log4j, not java util logging. Do something like this: -Dlog4j.configuration=/Users/bryan/git/bigdata-gpu/src/main/resources/log4j.properties ---- Bryan Thompson Chief Scientist & Founder Blazegraph e: br...@bl... w: http://blazegraph.com Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Mon, Feb 15, 2016 at 5:18 PM, Joakim Soderberg < joa...@bl...> wrote: > Hello, > > I am trying to set the logger from the command prompt like this without > success: > > java -server -Xmx4g -Dbigdata.propertyFile=RWStore.properties > -Djava.util.logging.config.file=log4j.properties -jar blazegraph.jar > > Does anyone know how configure the log util when running from the command > line? > > > > ------------------------------------------------------------------------------ > Site24x7 APM Insight: Get Deep Visibility into Application Performance > APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month > Monitor end-to-end web transactions and take corrective actions now > Troubleshoot faster and improve end-user experience. Signup Now! > http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > |
|
From: Joakim S. <joa...@bl...> - 2016-02-16 06:13:47
|
Thanks, now it works just fine. > On Feb 15, 2016, at 3:34 PM, Bryan Thompson <br...@sy...> wrote: > > Blazegraph is configured using log4j, not java util logging. Do something like this: > > -Dlog4j.configuration=/Users/bryan/git/bigdata-gpu/src/main/resources/log4j.properties > > > ---- > Bryan Thompson > Chief Scientist & Founder > Blazegraph > e: br...@bl... <mailto:br...@bl...> > w: http://blazegraph.com <http://blazegraph.com/> > > Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. > > Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > On Mon, Feb 15, 2016 at 5:18 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: > Hello, > > I am trying to set the logger from the command prompt like this without success: > > java -server -Xmx4g -Dbigdata.propertyFile=RWStore.properties -Djava.util.logging.config.file=log4j.properties -jar blazegraph.jar > > Does anyone know how configure the log util when running from the command line? > > > ------------------------------------------------------------------------------ > Site24x7 APM Insight: Get Deep Visibility into Application Performance > APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month > Monitor end-to-end web transactions and take corrective actions now > Troubleshoot faster and improve end-user experience. Signup Now! > http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140> > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> > |
|
From: Joakim S. <joa...@bl...> - 2016-02-16 06:42:47
|
Has anyone succeeded to load a folder of .nt files? I can load one by one: LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt> INTO GRAPH <http://dbpedia2015> But it doesn’t like a folder name LOAD <file:///mydata/dbpedia2015/core/> INTO GRAPH <http://dbpedia2015> |
|
From: Jeremy J C. <jj...@sy...> - 2016-02-16 15:40:11
|
> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <joa...@bl...> wrote: > > Has anyone succeeded to load a folder of .nt files? I can load one by one: > > LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> > > But it doesn’t like a folder name > LOAD <file:///mydata/dbpedia2015/core/ <file:///mydata/dbpedia2015/core/>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> That is correct. If you look at the spec for LOAD: https://www.w3.org/TR/sparql11-update/#load <https://www.w3.org/TR/sparql11-update/#load> then it takes an IRI as where you are loading from, and the concept of folder is simply not applicable. A few schemes such as file: and ftp: may have such a notion, but the operation you are looking for is local to your machine on the client and you should probably implement it yourself. In particular, do you want each file loaded into a different graph or the same graph: probably best for you to make up your own mind. I have had success loading trig files into multiple graphs, using a simple POST to the endpoint. Jeremy |
|
From: Bryan T. <br...@sy...> - 2016-02-16 15:55:17
|
2.0 includes support for bulk data load with a number of interesting features, including durable queue patterns, folders, etc. See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load ---- Bryan Thompson Chief Scientist & Founder Blazegraph e: br...@bl... w: http://blazegraph.com Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy...> wrote: > > > On Feb 15, 2016, at 10:42 PM, Joakim Soderberg < > joa...@bl...> wrote: > > Has anyone succeeded to load a folder of .nt files? I can load one by one: > > LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt> INTO > GRAPH <http://dbpedia2015> > > But it doesn’t like a folder name > LOAD <file:///mydata/dbpedia2015/core/> INTO GRAPH <http://dbpedia2015> > > > > That is correct. If you look at the spec for LOAD: > https://www.w3.org/TR/sparql11-update/#load > then it takes an IRI as where you are loading from, and the concept of > folder is simply not applicable. > A few schemes such as file: and ftp: may have such a notion, but the > operation you are looking for is local to your machine on the client and > you should probably implement it yourself. > > In particular, do you want each file loaded into a different graph or the > same graph: probably best for you to make up your own mind. > > I have had success loading trig files into multiple graphs, using a > simple POST to the endpoint. > > > Jeremy > > > > ------------------------------------------------------------------------------ > Site24x7 APM Insight: Get Deep Visibility into Application Performance > APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month > Monitor end-to-end web transactions and take corrective actions now > Troubleshoot faster and improve end-user experience. Signup Now! > http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > > |
|
From: Jeremy J C. <jj...@sy...> - 2016-02-16 16:10:05
|
> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load> That looks very interesting: I read: "Parsing, insert, and removal on the database are now decoupled from the index writes” One behavior we have is that we have small inserts concurrent with other activity (typically but not exclusively read activity). Does the enhanced configurability in 2.0 give us options that may allow us to improve performance of these writes. E.g. this week we have many (millions? at least hundreds of thousands) of such small writes (10 - 100 quads) and we also are trying to delete 25 million quads using about 100 delete/insert requests (that I take to be not impacted by this change). I am currently suggesting we should do one or the other at any one time, and not try to mix: but frankly I am guessing, and guessing conservatively. We have to maintain an always-on read performance at the same time. Total store size approx 3billion. [Unfortunately this machine is still a 1.5.3 machine, but for future reference I am trying to have better sense of how to organize such activity] Jeremy > On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy...> wrote: > > 2.0 includes support for bulk data load with a number of interesting features, including durable queue patterns, folders, etc. See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load> > > ---- > Bryan Thompson > Chief Scientist & Founder > Blazegraph > e: br...@bl... <mailto:br...@bl...> > w: http://blazegraph.com <http://blazegraph.com/> > > Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. > > Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote: > > >> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >> >> Has anyone succeeded to load a folder of .nt files? I can load one by one: >> >> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> >> >> But it doesn’t like a folder name >> LOAD <file:///mydata/dbpedia2015/core/ <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> > > > That is correct. If you look at the spec for LOAD: > https://www.w3.org/TR/sparql11-update/#load <https://www.w3.org/TR/sparql11-update/#load> > then it takes an IRI as where you are loading from, and the concept of folder is simply not applicable. > A few schemes such as file: and ftp: may have such a notion, but the operation you are looking for is local to your machine on the client and you should probably implement it yourself. > > In particular, do you want each file loaded into a different graph or the same graph: probably best for you to make up your own mind. > > I have had success loading trig files into multiple graphs, using a simple POST to the endpoint. > > > Jeremy > > > ------------------------------------------------------------------------------ > Site24x7 APM Insight: Get Deep Visibility into Application Performance > APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month > Monitor end-to-end web transactions and take corrective actions now > Troubleshoot faster and improve end-user experience. Signup Now! > http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140> > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> > > |
|
From: Bryan T. <br...@sy...> - 2016-02-16 16:17:55
|
Jeremy, The bulk data loader will not help with that scenario. It is desired for high throughput load. It can be used concurrent with query, but it can not really be mixed with concurrent small updates. In general, mixing concurrent small updates and large updates does not work well. Updates against a single graph must be serialized using the unisolated connection. So at some point the small updates will block for the large update. Thanks, Bryan ---- Bryan Thompson Chief Scientist & Founder Blazegraph e: br...@bl... w: http://blazegraph.com Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Tue, Feb 16, 2016 at 11:09 AM, Jeremy J Carroll <jj...@sy...> wrote: > See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load > > > > That looks very interesting: > > I read: > > "Parsing, insert, and removal on the database are now decoupled from the > index writes” > > One behavior we have is that we have small inserts concurrent with other > activity (typically but not exclusively read activity). Does the > enhanced configurability in 2.0 give us options that may allow us to > improve performance of these writes. > > E.g. this week we have many (millions? at least hundreds of thousands) of > such small writes (10 - 100 quads) and we also are trying to delete 25 > million quads using about 100 delete/insert requests (that I take to be not > impacted by this change). I am currently suggesting we should do one or the > other at any one time, and not try to mix: but frankly I am guessing, and > guessing conservatively. We have to maintain an always-on read > performance at the same time. Total store size approx 3billion. > > [Unfortunately this machine is still a 1.5.3 machine, but for future > reference I am trying to have better sense of how to organize such activity] > > Jeremy > > > > > > On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy...> wrote: > > 2.0 includes support for bulk data load with a number of interesting > features, including durable queue patterns, folders, etc. See > https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load > > ---- > Bryan Thompson > Chief Scientist & Founder > Blazegraph > e: br...@bl... > w: http://blazegraph.com > > Blazegraph products help to solve the Graph Cache Thrash to achieve large > scale processing for graph and predictive analytics. Blazegraph is the > creator of the industry’s first GPU-accelerated high-performance database > for large graphs, has been named as one of the “10 Companies and > Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. > > > Blazegraph Database <https://www.blazegraph.com/> is our ultra-high > performance graph database that supports both RDF/SPARQL and > Tinkerpop/Blueprints APIs. Blazegraph GPU > <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS > <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new > technologies that use GPUs to enable extreme scaling that is thousands of > times faster and 40 times more affordable than CPU-based solutions. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, > disclosure, dissemination or copying of this email or its contents or > attachments is prohibited. If you have received this communication in > error, please notify the sender by reply email and permanently delete all > copies of the email and its contents and attachments. > > On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy...> wrote: > >> >> >> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg < >> joa...@bl...> wrote: >> >> Has anyone succeeded to load a folder of .nt files? I can load one by one: >> >> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt> INTO >> GRAPH <http://dbpedia2015> >> >> But it doesn’t like a folder name >> LOAD <file:///mydata/dbpedia2015/core/> INTO GRAPH <http://dbpedia2015> >> >> >> >> That is correct. If you look at the spec for LOAD: >> https://www.w3.org/TR/sparql11-update/#load >> then it takes an IRI as where you are loading from, and the concept of >> folder is simply not applicable. >> A few schemes such as file: and ftp: may have such a notion, but the >> operation you are looking for is local to your machine on the client and >> you should probably implement it yourself. >> >> In particular, do you want each file loaded into a different graph or the >> same graph: probably best for you to make up your own mind. >> >> I have had success loading trig files into multiple graphs, using a >> simple POST to the endpoint. >> >> >> Jeremy >> >> >> >> ------------------------------------------------------------------------------ >> Site24x7 APM Insight: Get Deep Visibility into Application Performance >> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >> Monitor end-to-end web transactions and take corrective actions now >> Troubleshoot faster and improve end-user experience. Signup Now! >> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 >> _______________________________________________ >> Bigdata-developers mailing list >> Big...@li... >> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >> >> > > |
|
From: Joakim S. <joa...@bl...> - 2016-02-16 16:36:06
|
I knew there is a DataLoader class, but I wasn’t aware it was available as a service in NanoSparql server. I will try it immediately Thanks Joakim > On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy...> wrote: > >> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load> > > > That looks very interesting: > > I read: > > "Parsing, insert, and removal on the database are now decoupled from the index writes” > > One behavior we have is that we have small inserts concurrent with other activity (typically but not exclusively read activity). Does the enhanced configurability in 2.0 give us options that may allow us to improve performance of these writes. > > E.g. this week we have many (millions? at least hundreds of thousands) of such small writes (10 - 100 quads) and we also are trying to delete 25 million quads using about 100 delete/insert requests (that I take to be not impacted by this change). I am currently suggesting we should do one or the other at any one time, and not try to mix: but frankly I am guessing, and guessing conservatively. We have to maintain an always-on read performance at the same time. Total store size approx 3billion. > > [Unfortunately this machine is still a 1.5.3 machine, but for future reference I am trying to have better sense of how to organize such activity] > > Jeremy > > > > > >> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy... <mailto:br...@sy...>> wrote: >> >> 2.0 includes support for bulk data load with a number of interesting features, including durable queue patterns, folders, etc. See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load> >> >> ---- >> Bryan Thompson >> Chief Scientist & Founder >> Blazegraph >> e: br...@bl... <mailto:br...@bl...> >> w: http://blazegraph.com <http://blazegraph.com/> >> >> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. >> >> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. >> >> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. >> >> >> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote: >> >> >>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >>> >>> Has anyone succeeded to load a folder of .nt files? I can load one by one: >>> >>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> >>> >>> But it doesn’t like a folder name >>> LOAD <file:///mydata/dbpedia2015/core/ <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> >> >> >> That is correct. If you look at the spec for LOAD: >> https://www.w3.org/TR/sparql11-update/#load <https://www.w3.org/TR/sparql11-update/#load> >> then it takes an IRI as where you are loading from, and the concept of folder is simply not applicable. >> A few schemes such as file: and ftp: may have such a notion, but the operation you are looking for is local to your machine on the client and you should probably implement it yourself. >> >> In particular, do you want each file loaded into a different graph or the same graph: probably best for you to make up your own mind. >> >> I have had success loading trig files into multiple graphs, using a simple POST to the endpoint. >> >> >> Jeremy >> >> >> ------------------------------------------------------------------------------ >> Site24x7 APM Insight: Get Deep Visibility into Application Performance >> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >> Monitor end-to-end web transactions and take corrective actions now >> Troubleshoot faster and improve end-user experience. Signup Now! >> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140> >> _______________________________________________ >> Bigdata-developers mailing list >> Big...@li... <mailto:Big...@li...> >> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> >> >> > |
|
From: Jeremy J C. <jj...@sy...> - 2016-02-17 16:20:15
|
> On Feb 16, 2016, at 8:17 AM, Bryan Thompson <br...@sy...> wrote: > > > Updates against a single graph must be serialized using the unisolated connection. perhaps reading too much into this … if in quads mode I have many small graphs, do updates against different graphs not interface (at the blazegraph level, rather than disk I/O level where there is obvious contention) Jeremy |
|
From: Bryan T. <br...@sy...> - 2016-02-17 16:22:28
|
Updates against different quads contexts in the same blazegraph namespace are updating the same 6 backing indices in blazegraph. So these updates would be serialized. For improved throughput, you can batched together a number of small updates against different contexts. Bryan ---- Bryan Thompson Chief Scientist & Founder Blazegraph e: br...@bl... w: http://blazegraph.com Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Wed, Feb 17, 2016 at 11:20 AM, Jeremy J Carroll <jj...@sy...> wrote: > > > > On Feb 16, 2016, at 8:17 AM, Bryan Thompson <br...@sy...> wrote: > > > Updates against a single graph must be serialized using the unisolated > connection. > > > perhaps reading too much into this … > if in quads mode I have many small graphs, do updates against different > graphs not interface (at the blazegraph level, rather than disk I/O level > where there is obvious contention) > > Jeremy > > > > |
|
From: Joakim S. <joa...@bl...> - 2016-02-17 20:21:47
|
I am calling:
curl -X POST --data-binary @dataloader.xml --header 'Content-Type:application/xml' http:/__.__.__:9999/blazegraph/dataloader
I can see the size of the JNL-file is increasing, but when I query number of statements in the dashboard the data doesn’t show up.
select (count(*) as ?num) { ?s ?p ?o }
Do I need to Flush the StatementBuffer to the backing store after the curl?
This is my config file:
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd">
<properties>
<!-- RDF Format (Default is rdf/xml) -->
<entry key="format">N-Triples</entry>
<!-- Base URI (Optional) -->
<entry key="baseURI"></entry>
<!-- Default Graph URI (Optional - Required for quads mode namespace) -->
<entry key="defaultGraph"></entry>
<!-- Suppress all stdout messages (Optional) -->
<entry key="quiet">false</entry>
<!-- Show additional messages detailing the load performance. (Optional) -->
<entry key="verbose">3</entry>
<!-- Compute the RDF(S)+ closure. (Optional) -->
<entry key="closure">false</entry>
<!-- Files will be renamed to either .good or .fail as they are processed.
The files will remain in the same directory. -->
<entry key="durableQueues">true</entry>
<!-- The namespace of the KB instance. Defaults to kb. -->
<entry key="namespace">kb</entry>
<!-- The configuration file for the database instance. It must be readable by the web application. -->
<entry key="propertyFile">RWStore.properties</entry>
<!-- Zero or more files or directories containing the data to be loaded.
This should be a comma delimited list. The files must be readable by the web application. -->
<entry key="fileOrDirs">/mydata/dbpedia2015/core/</entry>
</properties>
> On Feb 16, 2016, at 8:35 AM, Joakim Soderberg <joa...@bl...> wrote:
>
> I knew there is a DataLoader class, but I wasn’t aware it was available as a service in NanoSparql server. I will try it immediately
>
>
> Thanks
> Joakim
>
>> On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote:
>>
>>> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load>
>>
>>
>> That looks very interesting:
>>
>> I read:
>>
>> "Parsing, insert, and removal on the database are now decoupled from the index writes”
>>
>> One behavior we have is that we have small inserts concurrent with other activity (typically but not exclusively read activity). Does the enhanced configurability in 2.0 give us options that may allow us to improve performance of these writes.
>>
>> E.g. this week we have many (millions? at least hundreds of thousands) of such small writes (10 - 100 quads) and we also are trying to delete 25 million quads using about 100 delete/insert requests (that I take to be not impacted by this change). I am currently suggesting we should do one or the other at any one time, and not try to mix: but frankly I am guessing, and guessing conservatively. We have to maintain an always-on read performance at the same time. Total store size approx 3billion.
>>
>> [Unfortunately this machine is still a 1.5.3 machine, but for future reference I am trying to have better sense of how to organize such activity]
>>
>> Jeremy
>>
>>
>>
>>
>>
>>> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy... <mailto:br...@sy...>> wrote:
>>>
>>> 2.0 includes support for bulk data load with a number of interesting features, including durable queue patterns, folders, etc. See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load>
>>>
>>> ----
>>> Bryan Thompson
>>> Chief Scientist & Founder
>>> Blazegraph
>>> e: br...@bl... <mailto:br...@bl...>
>>> w: http://blazegraph.com <http://blazegraph.com/>
>>>
>>> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
>>>
>>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions.
>>>
>>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments.
>>>
>>>
>>> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote:
>>>
>>>
>>>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote:
>>>>
>>>> Has anyone succeeded to load a folder of .nt files? I can load one by one:
>>>>
>>>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>>
>>>>
>>>> But it doesn’t like a folder name
>>>> LOAD <file:///mydata/dbpedia2015/core/ <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>>
>>>
>>>
>>> That is correct. If you look at the spec for LOAD:
>>> https://www.w3.org/TR/sparql11-update/#load <https://www.w3.org/TR/sparql11-update/#load>
>>> then it takes an IRI as where you are loading from, and the concept of folder is simply not applicable.
>>> A few schemes such as file: and ftp: may have such a notion, but the operation you are looking for is local to your machine on the client and you should probably implement it yourself.
>>>
>>> In particular, do you want each file loaded into a different graph or the same graph: probably best for you to make up your own mind.
>>>
>>> I have had success loading trig files into multiple graphs, using a simple POST to the endpoint.
>>>
>>>
>>> Jeremy
>>>
>>>
>>> ------------------------------------------------------------------------------
>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance
>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
>>> Monitor end-to-end web transactions and take corrective actions now
>>> Troubleshoot faster and improve end-user experience. Signup Now!
>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140>
>>> _______________________________________________
>>> Bigdata-developers mailing list
>>> Big...@li... <mailto:Big...@li...>
>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers>
>>>
>>>
>>
>
|
|
From: Brad B. <be...@bl...> - 2016-02-17 20:27:06
|
Joakim,
With the DataLoader, the commit is after all of the data is loaded. Once
the load is complete, all of the statements will be visible.
Thanks, --Brad
On Wed, Feb 17, 2016 at 3:21 PM, Joakim Soderberg <
joa...@bl...> wrote:
> I am calling:
>
> curl -X POST --data-binary @dataloader.xml --header
> 'Content-Type:application/xml' http:/__.__.__:9999/blazegraph/dataloader
>
> I can see the size of the JNL-file is increasing, but when I query number
> of statements in the dashboard the data doesn’t show up.
>
> select (count(*) as ?num) { ?s ?p ?o }
>
> Do I need to Flush the StatementBuffer to the backing store after the curl?
>
> This is my config file:
>
> <?xml version="1.0" encoding="UTF-8" standalone="no"?>
> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd">
> <properties>
> <!-- RDF Format (Default is rdf/xml) -->
> <entry key="format">N-Triples</entry>
> <!-- Base URI (Optional) -->
> <entry key="baseURI"></entry>
> <!-- Default Graph URI (Optional -
> Required for quads mode namespace) -->
> <entry key="defaultGraph"></entry>
> <!-- Suppress all stdout
> messages (Optional) -->
> <entry
> key="quiet">false</entry>
> <!-- Show
> additional messages detailing the load performance. (Optional) -->
> <entry
> key="verbose">3</entry>
> <!--
> Compute the RDF(S)+ closure. (Optional) -->
> <entry key="closure">false</entry>
> <!-- Files will be renamed to either .good or .fail as
> they are processed.
> The files will remain in the same directory. -->
> <entry key="durableQueues">true</entry>
> <!-- The namespace of the KB instance.
> Defaults to kb. -->
> <entry key="namespace">kb</entry>
> <!-- The configuration file for the
> database instance. It must be readable by the web application. -->
> <entry key="propertyFile">RWStore.properties</entry>
> <!-- Zero or more files or directories containing the
> data to be loaded.
> This should be a comma delimited list. The files must
> be readable by the web application. -->
> <entry key="fileOrDirs">/mydata/dbpedia2015/core/</entry>
> </properties>
>
>
>
> On Feb 16, 2016, at 8:35 AM, Joakim Soderberg <
> joa...@bl...> wrote:
>
> I knew there is a DataLoader class, but I wasn’t aware it was available as
> a service in NanoSparql server. I will try it immediately
>
>
> Thanks
> Joakim
>
> On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy...> wrote:
>
> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load
>
>
>
> That looks very interesting:
>
> I read:
>
> "Parsing, insert, and removal on the database are now decoupled from the
> index writes”
>
> One behavior we have is that we have small inserts concurrent with other
> activity (typically but not exclusively read activity). Does the
> enhanced configurability in 2.0 give us options that may allow us to
> improve performance of these writes.
>
> E.g. this week we have many (millions? at least hundreds of thousands) of
> such small writes (10 - 100 quads) and we also are trying to delete 25
> million quads using about 100 delete/insert requests (that I take to be not
> impacted by this change). I am currently suggesting we should do one or the
> other at any one time, and not try to mix: but frankly I am guessing, and
> guessing conservatively. We have to maintain an always-on read
> performance at the same time. Total store size approx 3billion.
>
> [Unfortunately this machine is still a 1.5.3 machine, but for future
> reference I am trying to have better sense of how to organize such activity]
>
> Jeremy
>
>
>
>
>
> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy...> wrote:
>
> 2.0 includes support for bulk data load with a number of interesting
> features, including durable queue patterns, folders, etc. See
> https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load
>
> ----
> Bryan Thompson
> Chief Scientist & Founder
> Blazegraph
> e: br...@bl...
> w: http://blazegraph.com
>
> Blazegraph products help to solve the Graph Cache Thrash to achieve large
> scale processing for graph and predictive analytics. Blazegraph is the
> creator of the industry’s first GPU-accelerated high-performance database
> for large graphs, has been named as one of the “10 Companies and
> Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
>
>
> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high
> performance graph database that supports both RDF/SPARQL and
> Tinkerpop/Blueprints APIs. Blazegraph GPU
> <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS
> <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new
> technologies that use GPUs to enable extreme scaling that is thousands of
> times faster and 40 times more affordable than CPU-based solutions.
>
> CONFIDENTIALITY NOTICE: This email and its contents and attachments are
> for the sole use of the intended recipient(s) and are confidential or
> proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use,
> disclosure, dissemination or copying of this email or its contents or
> attachments is prohibited. If you have received this communication in
> error, please notify the sender by reply email and permanently delete all
> copies of the email and its contents and attachments.
>
> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy...> wrote:
>
>>
>>
>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <
>> joa...@bl...> wrote:
>>
>> Has anyone succeeded to load a folder of .nt files? I can load one by one:
>>
>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt> INTO
>> GRAPH <http://dbpedia2015>
>>
>> But it doesn’t like a folder name
>> LOAD <file:///mydata/dbpedia2015/core/> INTO GRAPH <http://dbpedia2015>
>>
>>
>>
>> That is correct. If you look at the spec for LOAD:
>> https://www.w3.org/TR/sparql11-update/#load
>> then it takes an IRI as where you are loading from, and the concept of
>> folder is simply not applicable.
>> A few schemes such as file: and ftp: may have such a notion, but the
>> operation you are looking for is local to your machine on the client and
>> you should probably implement it yourself.
>>
>> In particular, do you want each file loaded into a different graph or the
>> same graph: probably best for you to make up your own mind.
>>
>> I have had success loading trig files into multiple graphs, using a
>> simple POST to the endpoint.
>>
>>
>> Jeremy
>>
>>
>>
>> ------------------------------------------------------------------------------
>> Site24x7 APM Insight: Get Deep Visibility into Application Performance
>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
>> Monitor end-to-end web transactions and take corrective actions now
>> Troubleshoot faster and improve end-user experience. Signup Now!
>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140
>> _______________________________________________
>> Bigdata-developers mailing list
>> Big...@li...
>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers
>>
>>
>
>
>
>
>
> ------------------------------------------------------------------------------
> Site24x7 APM Insight: Get Deep Visibility into Application Performance
> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
> Monitor end-to-end web transactions and take corrective actions now
> Troubleshoot faster and improve end-user experience. Signup Now!
> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140
> _______________________________________________
> Bigdata-developers mailing list
> Big...@li...
> https://lists.sourceforge.net/lists/listinfo/bigdata-developers
>
>
--
_______________
Brad Bebee
CEO
Blazegraph
e: be...@bl...
m: 202.642.7961
w: www.blazegraph.com
Blazegraph products help to solve the Graph Cache Thrash to achieve large
scale processing for graph and predictive analytics. Blazegraph is the
creator of the industry’s first GPU-accelerated high-performance database
for large graphs, has been named as one of the “10 Companies and
Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
Blazegraph Database <https://www.blazegraph.com/> is our ultra-high
performance graph database that supports both RDF/SPARQL and
Tinkerpop/Blueprints APIs. Blazegraph GPU
<https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS
<https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new
technologies that use GPUs to enable extreme scaling that is thousands of
times faster and 40 times more affordable than CPU-based solutions.
CONFIDENTIALITY NOTICE: This email and its contents and attachments are
for the sole use of the intended recipient(s) and are confidential or
proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use,
disclosure, dissemination or copying of this email or its contents or
attachments is prohibited. If you have received this communication in
error, please notify the sender by reply email and permanently delete all
copies of the email and its contents and attachments.
|
|
From: Jeremy J C. <jj...@sy...> - 2016-02-22 17:04:32
|
Try looking on the status tab of the blazegraph UI in the browser. In the detail view of your particular task, there might be a counter showing how many triples have been updated.
(I am unsure as to which tasks support this under which versions …)
Jeremy
> On Feb 17, 2016, at 12:26 PM, Brad Bebee <be...@bl...> wrote:
>
> Joakim,
>
> With the DataLoader, the commit is after all of the data is loaded. Once the load is complete, all of the statements will be visible.
>
> Thanks, --Brad
>
> On Wed, Feb 17, 2016 at 3:21 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote:
> I am calling:
>
> curl -X POST --data-binary @dataloader.xml --header 'Content-Type:application/xml' http:/__.__.__:9999/blazegraph/dataloader
>
> I can see the size of the JNL-file is increasing, but when I query number of statements in the dashboard the data doesn’t show up.
>
> select (count(*) as ?num) { ?s ?p ?o }
>
> Do I need to Flush the StatementBuffer to the backing store after the curl?
>
> This is my config file:
>
> <?xml version="1.0" encoding="UTF-8" standalone="no"?>
> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd <http://java.sun.com/dtd/properties.dtd>">
> <properties>
> <!-- RDF Format (Default is rdf/xml) -->
> <entry key="format">N-Triples</entry>
> <!-- Base URI (Optional) -->
> <entry key="baseURI"></entry>
> <!-- Default Graph URI (Optional - Required for quads mode namespace) -->
> <entry key="defaultGraph"></entry>
> <!-- Suppress all stdout messages (Optional) -->
> <entry key="quiet">false</entry>
> <!-- Show additional messages detailing the load performance. (Optional) -->
> <entry key="verbose">3</entry>
> <!-- Compute the RDF(S)+ closure. (Optional) -->
> <entry key="closure">false</entry>
> <!-- Files will be renamed to either .good or .fail as they are processed.
> The files will remain in the same directory. -->
> <entry key="durableQueues">true</entry>
> <!-- The namespace of the KB instance. Defaults to kb. -->
> <entry key="namespace">kb</entry>
> <!-- The configuration file for the database instance. It must be readable by the web application. -->
> <entry key="propertyFile">RWStore.properties</entry>
> <!-- Zero or more files or directories containing the data to be loaded.
> This should be a comma delimited list. The files must be readable by the web application. -->
> <entry key="fileOrDirs">/mydata/dbpedia2015/core/</entry>
> </properties>
>
>
>
>
>> On Feb 16, 2016, at 8:35 AM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote:
>>
>> I knew there is a DataLoader class, but I wasn’t aware it was available as a service in NanoSparql server. I will try it immediately
>>
>>
>> Thanks
>> Joakim
>>
>>> On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote:
>>>
>>>> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load>
>>>
>>>
>>> That looks very interesting:
>>>
>>> I read:
>>>
>>> "Parsing, insert, and removal on the database are now decoupled from the index writes”
>>>
>>> One behavior we have is that we have small inserts concurrent with other activity (typically but not exclusively read activity). Does the enhanced configurability in 2.0 give us options that may allow us to improve performance of these writes.
>>>
>>> E.g. this week we have many (millions? at least hundreds of thousands) of such small writes (10 - 100 quads) and we also are trying to delete 25 million quads using about 100 delete/insert requests (that I take to be not impacted by this change). I am currently suggesting we should do one or the other at any one time, and not try to mix: but frankly I am guessing, and guessing conservatively. We have to maintain an always-on read performance at the same time. Total store size approx 3billion.
>>>
>>> [Unfortunately this machine is still a 1.5.3 machine, but for future reference I am trying to have better sense of how to organize such activity]
>>>
>>> Jeremy
>>>
>>>
>>>
>>>
>>>
>>>> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy... <mailto:br...@sy...>> wrote:
>>>>
>>>> 2.0 includes support for bulk data load with a number of interesting features, including durable queue patterns, folders, etc. See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load>
>>>>
>>>> ----
>>>> Bryan Thompson
>>>> Chief Scientist & Founder
>>>> Blazegraph
>>>> e: br...@bl... <mailto:br...@bl...>
>>>> w: http://blazegraph.com <http://blazegraph.com/>
>>>>
>>>> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
>>>>
>>>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions.
>>>>
>>>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments.
>>>>
>>>>
>>>> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote:
>>>>
>>>>
>>>>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote:
>>>>>
>>>>> Has anyone succeeded to load a folder of .nt files? I can load one by one:
>>>>>
>>>>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>>
>>>>>
>>>>> But it doesn’t like a folder name
>>>>> LOAD <file:///mydata/dbpedia2015/core/ <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>>
>>>>
>>>>
>>>> That is correct. If you look at the spec for LOAD:
>>>> https://www.w3.org/TR/sparql11-update/#load <https://www.w3.org/TR/sparql11-update/#load>
>>>> then it takes an IRI as where you are loading from, and the concept of folder is simply not applicable.
>>>> A few schemes such as file: and ftp: may have such a notion, but the operation you are looking for is local to your machine on the client and you should probably implement it yourself.
>>>>
>>>> In particular, do you want each file loaded into a different graph or the same graph: probably best for you to make up your own mind.
>>>>
>>>> I have had success loading trig files into multiple graphs, using a simple POST to the endpoint.
>>>>
>>>>
>>>> Jeremy
>>>>
>>>>
>>>> ------------------------------------------------------------------------------
>>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance
>>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
>>>> Monitor end-to-end web transactions and take corrective actions now
>>>> Troubleshoot faster and improve end-user experience. Signup Now!
>>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140>
>>>> _______________________________________________
>>>> Bigdata-developers mailing list
>>>> Big...@li... <mailto:Big...@li...>
>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers>
>>>>
>>>>
>>>
>>
>
>
> ------------------------------------------------------------------------------
> Site24x7 APM Insight: Get Deep Visibility into Application Performance
> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
> Monitor end-to-end web transactions and take corrective actions now
> Troubleshoot faster and improve end-user experience. Signup Now!
> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140>
> _______________________________________________
> Bigdata-developers mailing list
> Big...@li... <mailto:Big...@li...>
> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers>
>
>
>
>
> --
> _______________
> Brad Bebee
> CEO
> Blazegraph
> e: be...@bl... <mailto:be...@bl...>
> m: 202.642.7961 <tel:202.642.7961>
> w: www.blazegraph.com <http://www.blazegraph.com/>
>
> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
>
> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions.
>
> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments.
>
> ------------------------------------------------------------------------------
> Site24x7 APM Insight: Get Deep Visibility into Application Performance
> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
> Monitor end-to-end web transactions and take corrective actions now
> Troubleshoot faster and improve end-user experience. Signup Now!
> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________
> Bigdata-developers mailing list
> Big...@li...
> https://lists.sourceforge.net/lists/listinfo/bigdata-developers
|
|
From: Joakim S. <joa...@bl...> - 2016-02-22 19:59:17
|
Thanks for the advice. Now it has been indexing for several days and I have no idea what it’s doing.
> On Feb 22, 2016, at 9:04 AM, Jeremy J Carroll <jj...@sy...> wrote:
>
> Try looking on the status tab of the blazegraph UI in the browser. In the detail view of your particular task, there might be a counter showing how many triples have been updated.
>
> (I am unsure as to which tasks support this under which versions …)
>
> Jeremy
>
>
>
>> On Feb 17, 2016, at 12:26 PM, Brad Bebee <be...@bl... <mailto:be...@bl...>> wrote:
>>
>> Joakim,
>>
>> With the DataLoader, the commit is after all of the data is loaded. Once the load is complete, all of the statements will be visible.
>>
>> Thanks, --Brad
>>
>> On Wed, Feb 17, 2016 at 3:21 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote:
>> I am calling:
>>
>> curl -X POST --data-binary @dataloader.xml --header 'Content-Type:application/xml' http:/__.__.__:9999/blazegraph/dataloader
>>
>> I can see the size of the JNL-file is increasing, but when I query number of statements in the dashboard the data doesn’t show up.
>>
>> select (count(*) as ?num) { ?s ?p ?o }
>>
>> Do I need to Flush the StatementBuffer to the backing store after the curl?
>>
>> This is my config file:
>>
>> <?xml version="1.0" encoding="UTF-8" standalone="no"?>
>> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd <http://java.sun.com/dtd/properties.dtd>">
>> <properties>
>> <!-- RDF Format (Default is rdf/xml) -->
>> <entry key="format">N-Triples</entry>
>> <!-- Base URI (Optional) -->
>> <entry key="baseURI"></entry>
>> <!-- Default Graph URI (Optional - Required for quads mode namespace) -->
>> <entry key="defaultGraph"></entry>
>> <!-- Suppress all stdout messages (Optional) -->
>> <entry key="quiet">false</entry>
>> <!-- Show additional messages detailing the load performance. (Optional) -->
>> <entry key="verbose">3</entry>
>> <!-- Compute the RDF(S)+ closure. (Optional) -->
>> <entry key="closure">false</entry>
>> <!-- Files will be renamed to either .good or .fail as they are processed.
>> The files will remain in the same directory. -->
>> <entry key="durableQueues">true</entry>
>> <!-- The namespace of the KB instance. Defaults to kb. -->
>> <entry key="namespace">kb</entry>
>> <!-- The configuration file for the database instance. It must be readable by the web application. -->
>> <entry key="propertyFile">RWStore.properties</entry>
>> <!-- Zero or more files or directories containing the data to be loaded.
>> This should be a comma delimited list. The files must be readable by the web application. -->
>> <entry key="fileOrDirs">/mydata/dbpedia2015/core/</entry>
>> </properties>
>>
>>
>>
>>
>>> On Feb 16, 2016, at 8:35 AM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote:
>>>
>>> I knew there is a DataLoader class, but I wasn’t aware it was available as a service in NanoSparql server. I will try it immediately
>>>
>>>
>>> Thanks
>>> Joakim
>>>
>>>> On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote:
>>>>
>>>>> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load>
>>>>
>>>>
>>>> That looks very interesting:
>>>>
>>>> I read:
>>>>
>>>> "Parsing, insert, and removal on the database are now decoupled from the index writes”
>>>>
>>>> One behavior we have is that we have small inserts concurrent with other activity (typically but not exclusively read activity). Does the enhanced configurability in 2.0 give us options that may allow us to improve performance of these writes.
>>>>
>>>> E.g. this week we have many (millions? at least hundreds of thousands) of such small writes (10 - 100 quads) and we also are trying to delete 25 million quads using about 100 delete/insert requests (that I take to be not impacted by this change). I am currently suggesting we should do one or the other at any one time, and not try to mix: but frankly I am guessing, and guessing conservatively. We have to maintain an always-on read performance at the same time. Total store size approx 3billion.
>>>>
>>>> [Unfortunately this machine is still a 1.5.3 machine, but for future reference I am trying to have better sense of how to organize such activity]
>>>>
>>>> Jeremy
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy... <mailto:br...@sy...>> wrote:
>>>>>
>>>>> 2.0 includes support for bulk data load with a number of interesting features, including durable queue patterns, folders, etc. See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load>
>>>>>
>>>>> ----
>>>>> Bryan Thompson
>>>>> Chief Scientist & Founder
>>>>> Blazegraph
>>>>> e: br...@bl... <mailto:br...@bl...>
>>>>> w: http://blazegraph.com <http://blazegraph.com/>
>>>>>
>>>>> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
>>>>>
>>>>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions.
>>>>>
>>>>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments.
>>>>>
>>>>>
>>>>> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote:
>>>>>
>>>>>
>>>>>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote:
>>>>>>
>>>>>> Has anyone succeeded to load a folder of .nt files? I can load one by one:
>>>>>>
>>>>>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>>
>>>>>>
>>>>>> But it doesn’t like a folder name
>>>>>> LOAD <file:///mydata/dbpedia2015/core/ <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>>
>>>>>
>>>>>
>>>>> That is correct. If you look at the spec for LOAD:
>>>>> https://www.w3.org/TR/sparql11-update/#load <https://www.w3.org/TR/sparql11-update/#load>
>>>>> then it takes an IRI as where you are loading from, and the concept of folder is simply not applicable.
>>>>> A few schemes such as file: and ftp: may have such a notion, but the operation you are looking for is local to your machine on the client and you should probably implement it yourself.
>>>>>
>>>>> In particular, do you want each file loaded into a different graph or the same graph: probably best for you to make up your own mind.
>>>>>
>>>>> I have had success loading trig files into multiple graphs, using a simple POST to the endpoint.
>>>>>
>>>>>
>>>>> Jeremy
>>>>>
>>>>>
>>>>> ------------------------------------------------------------------------------
>>>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance
>>>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
>>>>> Monitor end-to-end web transactions and take corrective actions now
>>>>> Troubleshoot faster and improve end-user experience. Signup Now!
>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140>
>>>>> _______________________________________________
>>>>> Bigdata-developers mailing list
>>>>> Big...@li... <mailto:Big...@li...>
>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers>
>>>>>
>>>>>
>>>>
>>>
>>
>>
>> ------------------------------------------------------------------------------
>> Site24x7 APM Insight: Get Deep Visibility into Application Performance
>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
>> Monitor end-to-end web transactions and take corrective actions now
>> Troubleshoot faster and improve end-user experience. Signup Now!
>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140>
>> _______________________________________________
>> Bigdata-developers mailing list
>> Big...@li... <mailto:Big...@li...>
>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers>
>>
>>
>>
>>
>> --
>> _______________
>> Brad Bebee
>> CEO
>> Blazegraph
>> e: be...@bl... <mailto:be...@bl...>
>> m: 202.642.7961 <tel:202.642.7961>
>> w: www.blazegraph.com <http://www.blazegraph.com/>
>>
>> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
>>
>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions.
>>
>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments.
>>
>> ------------------------------------------------------------------------------
>> Site24x7 APM Insight: Get Deep Visibility into Application Performance
>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
>> Monitor end-to-end web transactions and take corrective actions now
>> Troubleshoot faster and improve end-user experience. Signup Now!
>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________ <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________>
>> Bigdata-developers mailing list
>> Big...@li...
>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers
>
|
|
From: Brad B. <be...@bl...> - 2016-02-22 20:06:13
|
Joakim,
You should see log output as the statements are loaded. How much data are
you loading at once?
Thanks, --Brad
On Mon, Feb 22, 2016 at 2:59 PM, Joakim Soderberg <
joa...@bl...> wrote:
> Thanks for the advice. Now it has been indexing for several days and I
> have no idea what it’s doing.
>
> On Feb 22, 2016, at 9:04 AM, Jeremy J Carroll <jj...@sy...> wrote:
>
> Try looking on the status tab of the blazegraph UI in the browser. In the
> detail view of your particular task, there might be a counter showing how
> many triples have been updated.
>
> (I am unsure as to which tasks support this under which versions …)
>
> Jeremy
>
>
>
> On Feb 17, 2016, at 12:26 PM, Brad Bebee <be...@bl...> wrote:
>
> Joakim,
>
> With the DataLoader, the commit is after all of the data is loaded. Once
> the load is complete, all of the statements will be visible.
>
> Thanks, --Brad
>
> On Wed, Feb 17, 2016 at 3:21 PM, Joakim Soderberg <
> joa...@bl...> wrote:
>
>> I am calling:
>>
>> curl -X POST --data-binary @dataloader.xml --header
>> 'Content-Type:application/xml' http:/__.__.__:9999/blazegraph/dataloader
>>
>> I can see the size of the JNL-file is increasing, but when I query number
>> of statements in the dashboard the data doesn’t show up.
>>
>> select (count(*) as ?num) { ?s ?p ?o }
>>
>> Do I need to Flush the StatementBuffer to the backing store after the
>> curl?
>>
>> This is my config file:
>>
>> <?xml version="1.0" encoding="UTF-8" standalone="no"?>
>> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd">
>> <properties>
>> <!-- RDF Format (Default is rdf/xml) -->
>> <entry key="format">N-Triples</entry>
>> <!-- Base URI (Optional) -->
>> <entry key="baseURI"></entry>
>> <!-- Default Graph URI (Optional -
>> Required for quads mode namespace) -->
>> <entry key="defaultGraph"></entry>
>> <!-- Suppress all stdout
>> messages (Optional) -->
>> <entry
>> key="quiet">false</entry>
>> <!-- Show
>> additional messages detailing the load performance. (Optional) -->
>> <entry
>> key="verbose">3</entry>
>> <!--
>> Compute the RDF(S)+ closure. (Optional) -->
>> <entry key="closure">false</entry>
>> <!-- Files will be renamed to either .good or .fail as
>> they are processed.
>> The files will remain in the same directory. -->
>> <entry key="durableQueues">true</entry>
>> <!-- The namespace of the KB instance.
>> Defaults to kb. -->
>> <entry key="namespace">kb</entry>
>> <!-- The configuration file for the
>> database instance. It must be readable by the web application. -->
>> <entry key="propertyFile">RWStore.properties</entry>
>> <!-- Zero or more files or directories containing the
>> data to be loaded.
>> This should be a comma delimited list. The files must
>> be readable by the web application. -->
>> <entry key="fileOrDirs">/mydata/dbpedia2015/core/</entry>
>> </properties>
>>
>>
>>
>> On Feb 16, 2016, at 8:35 AM, Joakim Soderberg <
>> joa...@bl...> wrote:
>>
>> I knew there is a DataLoader class, but I wasn’t aware it was available
>> as a service in NanoSparql server. I will try it immediately
>>
>>
>> Thanks
>> Joakim
>>
>> On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy...> wrote:
>>
>> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load
>>
>>
>>
>> That looks very interesting:
>>
>> I read:
>>
>> "Parsing, insert, and removal on the database are now decoupled from the
>> index writes”
>>
>> One behavior we have is that we have small inserts concurrent with other
>> activity (typically but not exclusively read activity). Does the
>> enhanced configurability in 2.0 give us options that may allow us to
>> improve performance of these writes.
>>
>> E.g. this week we have many (millions? at least hundreds of thousands) of
>> such small writes (10 - 100 quads) and we also are trying to delete 25
>> million quads using about 100 delete/insert requests (that I take to be not
>> impacted by this change). I am currently suggesting we should do one or the
>> other at any one time, and not try to mix: but frankly I am guessing, and
>> guessing conservatively. We have to maintain an always-on read
>> performance at the same time. Total store size approx 3billion.
>>
>> [Unfortunately this machine is still a 1.5.3 machine, but for future
>> reference I am trying to have better sense of how to organize such activity]
>>
>> Jeremy
>>
>>
>>
>>
>>
>> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy...> wrote:
>>
>> 2.0 includes support for bulk data load with a number of interesting
>> features, including durable queue patterns, folders, etc. See
>> https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load
>>
>> ----
>> Bryan Thompson
>> Chief Scientist & Founder
>> Blazegraph
>> e: br...@bl...
>> w: http://blazegraph.com
>>
>> Blazegraph products help to solve the Graph Cache Thrash to achieve large
>> scale processing for graph and predictive analytics. Blazegraph is the
>> creator of the industry’s first GPU-accelerated high-performance database
>> for large graphs, has been named as one of the “10 Companies and
>> Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
>>
>>
>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high
>> performance graph database that supports both RDF/SPARQL and
>> Tinkerpop/Blueprints APIs. Blazegraph GPU
>> <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS
>> <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive
>> new technologies that use GPUs to enable extreme scaling that is thousands
>> of times faster and 40 times more affordable than CPU-based solutions.
>>
>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are
>> for the sole use of the intended recipient(s) and are confidential or
>> proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use,
>> disclosure, dissemination or copying of this email or its contents or
>> attachments is prohibited. If you have received this communication in
>> error, please notify the sender by reply email and permanently delete all
>> copies of the email and its contents and attachments.
>>
>> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy...>
>> wrote:
>>
>>>
>>>
>>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <
>>> joa...@bl...> wrote:
>>>
>>> Has anyone succeeded to load a folder of .nt files? I can load one by
>>> one:
>>>
>>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt> INTO
>>> GRAPH <http://dbpedia2015>
>>>
>>> But it doesn’t like a folder name
>>> LOAD <file:///mydata/dbpedia2015/core/> INTO GRAPH <http://dbpedia2015>
>>>
>>>
>>>
>>> That is correct. If you look at the spec for LOAD:
>>> https://www.w3.org/TR/sparql11-update/#load
>>> then it takes an IRI as where you are loading from, and the concept of
>>> folder is simply not applicable.
>>> A few schemes such as file: and ftp: may have such a notion, but the
>>> operation you are looking for is local to your machine on the client and
>>> you should probably implement it yourself.
>>>
>>> In particular, do you want each file loaded into a different graph or
>>> the same graph: probably best for you to make up your own mind.
>>>
>>> I have had success loading trig files into multiple graphs, using a
>>> simple POST to the endpoint.
>>>
>>>
>>> Jeremy
>>>
>>>
>>>
>>> ------------------------------------------------------------------------------
>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance
>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
>>> Monitor end-to-end web transactions and take corrective actions now
>>> Troubleshoot faster and improve end-user experience. Signup Now!
>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140
>>> _______________________________________________
>>> Bigdata-developers mailing list
>>> Big...@li...
>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers
>>>
>>>
>>
>>
>>
>>
>>
>> ------------------------------------------------------------------------------
>> Site24x7 APM Insight: Get Deep Visibility into Application Performance
>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
>> Monitor end-to-end web transactions and take corrective actions now
>> Troubleshoot faster and improve end-user experience. Signup Now!
>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140
>> _______________________________________________
>> Bigdata-developers mailing list
>> Big...@li...
>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers
>>
>>
>
>
> --
> _______________
> Brad Bebee
> CEO
> Blazegraph
> e: be...@bl...
> m: 202.642.7961
> w: www.blazegraph.com
>
> Blazegraph products help to solve the Graph Cache Thrash to achieve large
> scale processing for graph and predictive analytics. Blazegraph is the
> creator of the industry’s first GPU-accelerated high-performance database
> for large graphs, has been named as one of the “10 Companies and
> Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
>
>
> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high
> performance graph database that supports both RDF/SPARQL and
> Tinkerpop/Blueprints APIs. Blazegraph GPU
> <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS
> <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new
> technologies that use GPUs to enable extreme scaling that is thousands of
> times faster and 40 times more affordable than CPU-based solutions.
>
> CONFIDENTIALITY NOTICE: This email and its contents and attachments are
> for the sole use of the intended recipient(s) and are confidential or
> proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use,
> disclosure, dissemination or copying of this email or its contents or
> attachments is prohibited. If you have received this communication in
> error, please notify the sender by reply email and permanently delete all
> copies of the email and its contents and attachments.
>
> ------------------------------------------------------------------------------
> Site24x7 APM Insight: Get Deep Visibility into Application Performance
> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month
> Monitor end-to-end web transactions and take corrective actions now
> Troubleshoot faster and improve end-user experience. Signup Now!
>
> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________
> Bigdata-developers mailing list
> Big...@li...
> https://lists.sourceforge.net/lists/listinfo/bigdata-developers
>
>
>
>
--
_______________
Brad Bebee
CEO
Blazegraph
e: be...@bl...
m: 202.642.7961
w: www.blazegraph.com
Blazegraph products help to solve the Graph Cache Thrash to achieve large
scale processing for graph and predictive analytics. Blazegraph is the
creator of the industry’s first GPU-accelerated high-performance database
for large graphs, has been named as one of the “10 Companies and
Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>.
Blazegraph Database <https://www.blazegraph.com/> is our ultra-high
performance graph database that supports both RDF/SPARQL and
Tinkerpop/Blueprints APIs. Blazegraph GPU
<https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS
<https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new
technologies that use GPUs to enable extreme scaling that is thousands of
times faster and 40 times more affordable than CPU-based solutions.
CONFIDENTIALITY NOTICE: This email and its contents and attachments are
for the sole use of the intended recipient(s) and are confidential or
proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use,
disclosure, dissemination or copying of this email or its contents or
attachments is prohibited. If you have received this communication in
error, please notify the sender by reply email and permanently delete all
copies of the email and its contents and attachments.
|
|
From: Joakim S. <joa...@bl...> - 2016-02-22 20:16:38
|
Brad, Thats’s right, in my log i get a steady stream of this: -02-22 20:11:11,639) INFO : StatementBuffer.java:1773: term: http://pl.dbpedia.org/resource/Melbourne_Zoo, iv: null (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://pt.dbpedia.org/resource/Zoológico_de_Melbourne, iv: null (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://ru.dbpedia.org/resource/Мельбурнский_зоопарк, iv: null (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://uk.dbpedia.org/resource/Мельбурнський_зоопарк, iv: null (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://vi.dbpedia.org/resource/Sở_thú_Melbourne, iv: null (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://dbpedia.org/resource/Nova_Air, iv: null (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://wikidata.org/entity/Q578032, iv: null (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://wikidata.dbpedia.org/resource/Q578032, iv: null (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://es.dbpedia.org/resource/Nova_Air, iv: null (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://pl.dbpedia.org/resource/Nova_Air, iv: null (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://dbpedia.org/resource/Milton_Work, iv: null (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://wikidata.org/entity/Q578085, iv: null (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://wikidata.dbpedia.org/resource/Q578085, iv: null (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://fr.dbpedia.org/resource/Milton_Work, iv: null (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://pl.dbpedia.org/resource/Milton_Work, iv: null (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://dbpedia.org/resource/Lisa_Nandy, iv: null (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://wikidata.org/entity/Q578037, iv: null (2016-02-22 20:11:11,642) INFO : StatementBuffer.java:1773: term: http://wikidata.dbpedia.org/resource/Q578037, iv: null Is “iv:null” bad? I am loading 53 ttl-files of 150G /Joakim > On Feb 22, 2016, at 12:06 PM, Brad Bebee <be...@bl...> wrote: > > Joakim, > > You should see log output as the statements are loaded. How much data are you loading at once? > > Thanks, --Brad > > On Mon, Feb 22, 2016 at 2:59 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: > Thanks for the advice. Now it has been indexing for several days and I have no idea what it’s doing. > >> On Feb 22, 2016, at 9:04 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote: >> >> Try looking on the status tab of the blazegraph UI in the browser. In the detail view of your particular task, there might be a counter showing how many triples have been updated. >> >> (I am unsure as to which tasks support this under which versions …) >> >> Jeremy >> >> >> >>> On Feb 17, 2016, at 12:26 PM, Brad Bebee <be...@bl... <mailto:be...@bl...>> wrote: >>> >>> Joakim, >>> >>> With the DataLoader, the commit is after all of the data is loaded. Once the load is complete, all of the statements will be visible. >>> >>> Thanks, --Brad >>> >>> On Wed, Feb 17, 2016 at 3:21 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >>> I am calling: >>> >>> curl -X POST --data-binary @dataloader.xml --header 'Content-Type:application/xml' http:/__.__.__:9999/blazegraph/dataloader >>> >>> I can see the size of the JNL-file is increasing, but when I query number of statements in the dashboard the data doesn’t show up. >>> >>> select (count(*) as ?num) { ?s ?p ?o } >>> >>> Do I need to Flush the StatementBuffer to the backing store after the curl? >>> >>> This is my config file: >>> >>> <?xml version="1.0" encoding="UTF-8" standalone="no"?> >>> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd <http://java.sun.com/dtd/properties.dtd>"> >>> <properties> >>> <!-- RDF Format (Default is rdf/xml) --> >>> <entry key="format">N-Triples</entry> >>> <!-- Base URI (Optional) --> >>> <entry key="baseURI"></entry> >>> <!-- Default Graph URI (Optional - Required for quads mode namespace) --> >>> <entry key="defaultGraph"></entry> >>> <!-- Suppress all stdout messages (Optional) --> >>> <entry key="quiet">false</entry> >>> <!-- Show additional messages detailing the load performance. (Optional) --> >>> <entry key="verbose">3</entry> >>> <!-- Compute the RDF(S)+ closure. (Optional) --> >>> <entry key="closure">false</entry> >>> <!-- Files will be renamed to either .good or .fail as they are processed. >>> The files will remain in the same directory. --> >>> <entry key="durableQueues">true</entry> >>> <!-- The namespace of the KB instance. Defaults to kb. --> >>> <entry key="namespace">kb</entry> >>> <!-- The configuration file for the database instance. It must be readable by the web application. --> >>> <entry key="propertyFile">RWStore.properties</entry> >>> <!-- Zero or more files or directories containing the data to be loaded. >>> This should be a comma delimited list. The files must be readable by the web application. --> >>> <entry key="fileOrDirs">/mydata/dbpedia2015/core/</entry> >>> </properties> >>> >>> >>> >>> >>>> On Feb 16, 2016, at 8:35 AM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >>>> >>>> I knew there is a DataLoader class, but I wasn’t aware it was available as a service in NanoSparql server. I will try it immediately >>>> >>>> >>>> Thanks >>>> Joakim >>>> >>>>> On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote: >>>>> >>>>>> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load> >>>>> >>>>> >>>>> That looks very interesting: >>>>> >>>>> I read: >>>>> >>>>> "Parsing, insert, and removal on the database are now decoupled from the index writes” >>>>> >>>>> One behavior we have is that we have small inserts concurrent with other activity (typically but not exclusively read activity). Does the enhanced configurability in 2.0 give us options that may allow us to improve performance of these writes. >>>>> >>>>> E.g. this week we have many (millions? at least hundreds of thousands) of such small writes (10 - 100 quads) and we also are trying to delete 25 million quads using about 100 delete/insert requests (that I take to be not impacted by this change). I am currently suggesting we should do one or the other at any one time, and not try to mix: but frankly I am guessing, and guessing conservatively. We have to maintain an always-on read performance at the same time. Total store size approx 3billion. >>>>> >>>>> [Unfortunately this machine is still a 1.5.3 machine, but for future reference I am trying to have better sense of how to organize such activity] >>>>> >>>>> Jeremy >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy... <mailto:br...@sy...>> wrote: >>>>>> >>>>>> 2.0 includes support for bulk data load with a number of interesting features, including durable queue patterns, folders, etc. See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load> >>>>>> >>>>>> ---- >>>>>> Bryan Thompson >>>>>> Chief Scientist & Founder >>>>>> Blazegraph >>>>>> e: br...@bl... <mailto:br...@bl...> >>>>>> w: http://blazegraph.com <http://blazegraph.com/> >>>>>> >>>>>> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. >>>>>> >>>>>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. >>>>>> >>>>>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. >>>>>> >>>>>> >>>>>> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote: >>>>>> >>>>>> >>>>>>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >>>>>>> >>>>>>> Has anyone succeeded to load a folder of .nt files? I can load one by one: >>>>>>> >>>>>>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> >>>>>>> >>>>>>> But it doesn’t like a folder name >>>>>>> LOAD <file:///mydata/dbpedia2015/core/ <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> >>>>>> >>>>>> >>>>>> That is correct. If you look at the spec for LOAD: >>>>>> https://www.w3.org/TR/sparql11-update/#load <https://www.w3.org/TR/sparql11-update/#load> >>>>>> then it takes an IRI as where you are loading from, and the concept of folder is simply not applicable. >>>>>> A few schemes such as file: and ftp: may have such a notion, but the operation you are looking for is local to your machine on the client and you should probably implement it yourself. >>>>>> >>>>>> In particular, do you want each file loaded into a different graph or the same graph: probably best for you to make up your own mind. >>>>>> >>>>>> I have had success loading trig files into multiple graphs, using a simple POST to the endpoint. >>>>>> >>>>>> >>>>>> Jeremy >>>>>> >>>>>> >>>>>> ------------------------------------------------------------------------------ >>>>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance >>>>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>>>>> Monitor end-to-end web transactions and take corrective actions now >>>>>> Troubleshoot faster and improve end-user experience. Signup Now! >>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140> >>>>>> _______________________________________________ >>>>>> Bigdata-developers mailing list >>>>>> Big...@li... <mailto:Big...@li...> >>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> >>>>>> >>>>>> >>>>> >>>> >>> >>> >>> ------------------------------------------------------------------------------ >>> Site24x7 APM Insight: Get Deep Visibility into Application Performance >>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>> Monitor end-to-end web transactions and take corrective actions now >>> Troubleshoot faster and improve end-user experience. Signup Now! >>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140> >>> _______________________________________________ >>> Bigdata-developers mailing list >>> Big...@li... <mailto:Big...@li...> >>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> >>> >>> >>> >>> >>> -- >>> _______________ >>> Brad Bebee >>> CEO >>> Blazegraph >>> e: be...@bl... <mailto:be...@bl...> >>> m: 202.642.7961 <tel:202.642.7961> >>> w: www.blazegraph.com <http://www.blazegraph.com/> >>> >>> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. >>> >>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. >>> >>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. >>> >>> ------------------------------------------------------------------------------ >>> Site24x7 APM Insight: Get Deep Visibility into Application Performance >>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>> Monitor end-to-end web transactions and take corrective actions now >>> Troubleshoot faster and improve end-user experience. Signup Now! >>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________ <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________> >>> Bigdata-developers mailing list >>> Big...@li... <mailto:Big...@li...> >>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> >> > > > > > -- > _______________ > Brad Bebee > CEO > Blazegraph > e: be...@bl... <mailto:be...@bl...> > m: 202.642.7961 <tel:202.642.7961> > w: www.blazegraph.com <http://www.blazegraph.com/> > > Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. > > Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > |
|
From: Bryan T. <br...@sy...> - 2016-02-22 20:20:42
|
Do not have log @ INFO for blazegraph. It will kill performance. Put it at WARN. There is a bug in the DataLoaderServlet. If you have to abort a load, make sure that you terminate the blazegraph process since that servlet does not correctly unwind a partial commit. Bryan ---- Bryan Thompson Chief Scientist & Founder Blazegraph e: br...@bl... w: http://blazegraph.com Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. On Mon, Feb 22, 2016 at 3:16 PM, Joakim Soderberg < joa...@bl...> wrote: > Brad, > Thats’s right, in my log i get a steady stream of this: > > -02-22 20:11:11,639) INFO : StatementBuffer.java:1773: term: > http://pl.dbpedia.org/resource/Melbourne_Zoo, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: > http://pt.dbpedia.org/resource/Zoológico_de_Melbourne, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: > http://ru.dbpedia.org/resource/Мельбурнский_зоопарк, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: > http://uk.dbpedia.org/resource/Мельбурнський_зоопарк, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: > http://vi.dbpedia.org/resource/Sở_thú_Melbourne, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: > http://dbpedia.org/resource/Nova_Air, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: > http://wikidata.org/entity/Q578032, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: > http://wikidata.dbpedia.org/resource/Q578032, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: > http://es.dbpedia.org/resource/Nova_Air, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: > http://pl.dbpedia.org/resource/Nova_Air, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: > http://dbpedia.org/resource/Milton_Work, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: > http://wikidata.org/entity/Q578085, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: > http://wikidata.dbpedia.org/resource/Q578085, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: > http://fr.dbpedia.org/resource/Milton_Work, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: > http://pl.dbpedia.org/resource/Milton_Work, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: > http://dbpedia.org/resource/Lisa_Nandy, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: > http://wikidata.org/entity/Q578037, iv: null > (2016-02-22 20:11:11,642) INFO : StatementBuffer.java:1773: term: > http://wikidata.dbpedia.org/resource/Q578037, iv: null > > > Is “iv:null” bad? > > I am loading 53 ttl-files of 150G > > /Joakim > > > > On Feb 22, 2016, at 12:06 PM, Brad Bebee <be...@bl...> wrote: > > Joakim, > > You should see log output as the statements are loaded. How much data > are you loading at once? > > Thanks, --Brad > > On Mon, Feb 22, 2016 at 2:59 PM, Joakim Soderberg < > joa...@bl...> wrote: > >> Thanks for the advice. Now it has been indexing for several days and I >> have no idea what it’s doing. >> >> On Feb 22, 2016, at 9:04 AM, Jeremy J Carroll <jj...@sy...> wrote: >> >> Try looking on the status tab of the blazegraph UI in the browser. In the >> detail view of your particular task, there might be a counter showing how >> many triples have been updated. >> >> (I am unsure as to which tasks support this under which versions …) >> >> Jeremy >> >> >> >> On Feb 17, 2016, at 12:26 PM, Brad Bebee <be...@bl...> wrote: >> >> Joakim, >> >> With the DataLoader, the commit is after all of the data is loaded. Once >> the load is complete, all of the statements will be visible. >> >> Thanks, --Brad >> >> On Wed, Feb 17, 2016 at 3:21 PM, Joakim Soderberg < >> joa...@bl...> wrote: >> >>> I am calling: >>> >>> curl -X POST --data-binary @dataloader.xml --header >>> 'Content-Type:application/xml' http:/__.__.__:9999/blazegraph/dataloader >>> >>> I can see the size of the JNL-file is increasing, but when I query >>> number of statements in the dashboard the data doesn’t show up. >>> >>> select (count(*) as ?num) { ?s ?p ?o } >>> >>> Do I need to Flush the StatementBuffer to the backing store after the >>> curl? >>> >>> This is my config file: >>> >>> <?xml version="1.0" encoding="UTF-8" standalone="no"?> >>> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd"> >>> <properties> >>> <!-- RDF Format (Default is rdf/xml) --> >>> <entry key="format">N-Triples</entry> >>> <!-- Base URI (Optional) --> >>> <entry key="baseURI"></entry> >>> <!-- Default Graph URI (Optional - >>> Required for quads mode namespace) --> >>> <entry >>> key="defaultGraph"></entry> >>> <!-- Suppress all stdout >>> messages (Optional) --> >>> <entry >>> key="quiet">false</entry> >>> <!-- Show >>> additional messages detailing the load performance. (Optional) --> >>> <entry >>> key="verbose">3</entry> >>> >>> <!-- Compute the RDF(S)+ closure. (Optional) --> >>> <entry key="closure">false</entry> >>> <!-- Files will be renamed to either .good or .fail as >>> they are processed. >>> The files will remain in the same directory. --> >>> <entry key="durableQueues">true</entry> >>> <!-- The namespace of the KB instance. >>> Defaults to kb. --> >>> <entry key="namespace">kb</entry> >>> <!-- The configuration file for >>> the database instance. It must be readable by the web application. --> >>> <entry key="propertyFile">RWStore.properties</entry> >>> <!-- Zero or more files or directories containing the >>> data to be loaded. >>> This should be a comma delimited list. The files must >>> be readable by the web application. --> >>> <entry key="fileOrDirs">/mydata/dbpedia2015/core/</entry> >>> </properties> >>> >>> >>> >>> On Feb 16, 2016, at 8:35 AM, Joakim Soderberg < >>> joa...@bl...> wrote: >>> >>> I knew there is a DataLoader class, but I wasn’t aware it was available >>> as a service in NanoSparql server. I will try it immediately >>> >>> >>> Thanks >>> Joakim >>> >>> On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy...> wrote: >>> >>> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load >>> >>> >>> >>> That looks very interesting: >>> >>> I read: >>> >>> "Parsing, insert, and removal on the database are now decoupled from the >>> index writes” >>> >>> One behavior we have is that we have small inserts concurrent with other >>> activity (typically but not exclusively read activity). Does the >>> enhanced configurability in 2.0 give us options that may allow us to >>> improve performance of these writes. >>> >>> E.g. this week we have many (millions? at least hundreds of thousands) >>> of such small writes (10 - 100 quads) and we also are trying to delete 25 >>> million quads using about 100 delete/insert requests (that I take to be not >>> impacted by this change). I am currently suggesting we should do one or the >>> other at any one time, and not try to mix: but frankly I am guessing, and >>> guessing conservatively. We have to maintain an always-on read >>> performance at the same time. Total store size approx 3billion. >>> >>> [Unfortunately this machine is still a 1.5.3 machine, but for future >>> reference I am trying to have better sense of how to organize such activity] >>> >>> Jeremy >>> >>> >>> >>> >>> >>> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy...> wrote: >>> >>> 2.0 includes support for bulk data load with a number of interesting >>> features, including durable queue patterns, folders, etc. See >>> https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load >>> >>> ---- >>> Bryan Thompson >>> Chief Scientist & Founder >>> Blazegraph >>> e: br...@bl... >>> w: http://blazegraph.com >>> >>> Blazegraph products help to solve the Graph Cache Thrash to achieve >>> large scale processing for graph and predictive analytics. Blazegraph is >>> the creator of the industry’s first GPU-accelerated high-performance >>> database for large graphs, has been named as one of the “10 Companies >>> and Technologies to Watch in 2016” >>> <http://insideanalysis.com/2016/01/20535/>. >>> >>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high >>> performance graph database that supports both RDF/SPARQL and >>> Tinkerpop/Blueprints APIs. Blazegraph GPU >>> <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS >>> <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive >>> new technologies that use GPUs to enable extreme scaling that is thousands >>> of times faster and 40 times more affordable than CPU-based solutions. >>> >>> CONFIDENTIALITY NOTICE: This email and its contents and attachments >>> are for the sole use of the intended recipient(s) and are confidential or >>> proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, >>> disclosure, dissemination or copying of this email or its contents or >>> attachments is prohibited. If you have received this communication in >>> error, please notify the sender by reply email and permanently delete all >>> copies of the email and its contents and attachments. >>> >>> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy...> >>> wrote: >>> >>>> >>>> >>>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg < >>>> joa...@bl...> wrote: >>>> >>>> Has anyone succeeded to load a folder of .nt files? I can load one by >>>> one: >>>> >>>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt> INTO >>>> GRAPH <http://dbpedia2015> >>>> >>>> But it doesn’t like a folder name >>>> LOAD <file:///mydata/dbpedia2015/core/> INTO GRAPH <http://dbpedia2015> >>>> >>>> >>>> >>>> That is correct. If you look at the spec for LOAD: >>>> https://www.w3.org/TR/sparql11-update/#load >>>> then it takes an IRI as where you are loading from, and the concept of >>>> folder is simply not applicable. >>>> A few schemes such as file: and ftp: may have such a notion, but the >>>> operation you are looking for is local to your machine on the client and >>>> you should probably implement it yourself. >>>> >>>> In particular, do you want each file loaded into a different graph or >>>> the same graph: probably best for you to make up your own mind. >>>> >>>> I have had success loading trig files into multiple graphs, using a >>>> simple POST to the endpoint. >>>> >>>> >>>> Jeremy >>>> >>>> >>>> >>>> ------------------------------------------------------------------------------ >>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance >>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>>> Monitor end-to-end web transactions and take corrective actions now >>>> Troubleshoot faster and improve end-user experience. Signup Now! >>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 >>>> _______________________________________________ >>>> Bigdata-developers mailing list >>>> Big...@li... >>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>>> >>>> >>> >>> >>> >>> >>> >>> ------------------------------------------------------------------------------ >>> Site24x7 APM Insight: Get Deep Visibility into Application Performance >>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>> Monitor end-to-end web transactions and take corrective actions now >>> Troubleshoot faster and improve end-user experience. Signup Now! >>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 >>> _______________________________________________ >>> Bigdata-developers mailing list >>> Big...@li... >>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >>> >>> >> >> >> -- >> _______________ >> Brad Bebee >> CEO >> Blazegraph >> e: be...@bl... >> m: 202.642.7961 >> w: www.blazegraph.com >> >> Blazegraph products help to solve the Graph Cache Thrash to achieve large >> scale processing for graph and predictive analytics. Blazegraph is the >> creator of the industry’s first GPU-accelerated high-performance database >> for large graphs, has been named as one of the “10 Companies and >> Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. >> >> >> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high >> performance graph database that supports both RDF/SPARQL and >> Tinkerpop/Blueprints APIs. Blazegraph GPU >> <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS >> <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive >> new technologies that use GPUs to enable extreme scaling that is thousands >> of times faster and 40 times more affordable than CPU-based solutions. >> >> CONFIDENTIALITY NOTICE: This email and its contents and attachments are >> for the sole use of the intended recipient(s) and are confidential or >> proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, >> disclosure, dissemination or copying of this email or its contents or >> attachments is prohibited. If you have received this communication in >> error, please notify the sender by reply email and permanently delete all >> copies of the email and its contents and attachments. >> >> ------------------------------------------------------------------------------ >> Site24x7 APM Insight: Get Deep Visibility into Application Performance >> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >> Monitor end-to-end web transactions and take corrective actions now >> Troubleshoot faster and improve end-user experience. Signup Now! >> >> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________ >> Bigdata-developers mailing list >> Big...@li... >> https://lists.sourceforge.net/lists/listinfo/bigdata-developers >> >> >> >> > > > -- > _______________ > Brad Bebee > CEO > Blazegraph > e: be...@bl... > m: 202.642.7961 > w: www.blazegraph.com > > Blazegraph products help to solve the Graph Cache Thrash to achieve large > scale processing for graph and predictive analytics. Blazegraph is the > creator of the industry’s first GPU-accelerated high-performance database > for large graphs, has been named as one of the “10 Companies and > Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. > > > Blazegraph Database <https://www.blazegraph.com/> is our ultra-high > performance graph database that supports both RDF/SPARQL and > Tinkerpop/Blueprints APIs. Blazegraph GPU > <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS > <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new > technologies that use GPUs to enable extreme scaling that is thousands of > times faster and 40 times more affordable than CPU-based solutions. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are > for the sole use of the intended recipient(s) and are confidential or > proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, > disclosure, dissemination or copying of this email or its contents or > attachments is prohibited. If you have received this communication in > error, please notify the sender by reply email and permanently delete all > copies of the email and its contents and attachments. > > > > > ------------------------------------------------------------------------------ > Site24x7 APM Insight: Get Deep Visibility into Application Performance > APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month > Monitor end-to-end web transactions and take corrective actions now > Troubleshoot faster and improve end-user experience. Signup Now! > http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 > _______________________________________________ > Bigdata-developers mailing list > Big...@li... > https://lists.sourceforge.net/lists/listinfo/bigdata-developers > > |
|
From: Joakim S. <joa...@bl...> - 2016-02-22 21:41:01
|
Bryan, Thanks for the reminder. I changed the loggers to: log4j.logger.com.bigdata=WARN log4j.logger.com.bigdata.btree=WARN log4j.rootCategory=INFO, devDest, fileDev and restarted the indexing. How long time should I expect to index dbpedia core on a 8 CPU 61 GiB memory machine? > On Feb 22, 2016, at 12:20 PM, Bryan Thompson <br...@sy...> wrote: > > Do not have log @ INFO for blazegraph. It will kill performance. Put it at WARN. > > There is a bug in the DataLoaderServlet. If you have to abort a load, make sure that you terminate the blazegraph process since that servlet does not correctly unwind a partial commit. > > Bryan > > ---- > Bryan Thompson > Chief Scientist & Founder > Blazegraph > e: br...@bl... <mailto:br...@bl...> > w: http://blazegraph.com <http://blazegraph.com/> > > Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. > > Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. > > CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. > > > On Mon, Feb 22, 2016 at 3:16 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: > Brad, > Thats’s right, in my log i get a steady stream of this: > > -02-22 20:11:11,639) INFO : StatementBuffer.java:1773: term: http://pl.dbpedia.org/resource/Melbourne_Zoo <http://pl.dbpedia.org/resource/Melbourne_Zoo>, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://pt.dbpedia.org/resource/Zoológico_de_Melbourne <http://pt.dbpedia.org/resource/Zool%C3%B3gico_de_Melbourne>, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://ru.dbpedia.org/resource/Мельбурнский_зоопарк <http://ru.dbpedia.org/resource/%D0%9C%D0%B5%D0%BB%D1%8C%D0%B1%D1%83%D1%80%D0%BD%D1%81%D0%BA%D0%B8%D0%B9_%D0%B7%D0%BE%D0%BE%D0%BF%D0%B0%D1%80%D0%BA>, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://uk.dbpedia.org/resource/Мельбурнський_зоопарк <http://uk.dbpedia.org/resource/%D0%9C%D0%B5%D0%BB%D1%8C%D0%B1%D1%83%D1%80%D0%BD%D1%81%D1%8C%D0%BA%D0%B8%D0%B9_%D0%B7%D0%BE%D0%BE%D0%BF%D0%B0%D1%80%D0%BA>, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://vi.dbpedia.org/resource/Sở_thú_Melbourne <http://vi.dbpedia.org/resource/S%E1%BB%9F_th%C3%BA_Melbourne>, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://dbpedia.org/resource/Nova_Air <http://dbpedia.org/resource/Nova_Air>, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://wikidata.org/entity/Q578032 <http://wikidata.org/entity/Q578032>, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://wikidata.dbpedia.org/resource/Q578032 <http://wikidata.dbpedia.org/resource/Q578032>, iv: null > (2016-02-22 20:11:11,640) INFO : StatementBuffer.java:1773: term: http://es.dbpedia.org/resource/Nova_Air <http://es.dbpedia.org/resource/Nova_Air>, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://pl.dbpedia.org/resource/Nova_Air <http://pl.dbpedia.org/resource/Nova_Air>, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://dbpedia.org/resource/Milton_Work <http://dbpedia.org/resource/Milton_Work>, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://wikidata.org/entity/Q578085 <http://wikidata.org/entity/Q578085>, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://wikidata.dbpedia.org/resource/Q578085 <http://wikidata.dbpedia.org/resource/Q578085>, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://fr.dbpedia.org/resource/Milton_Work <http://fr.dbpedia.org/resource/Milton_Work>, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://pl.dbpedia.org/resource/Milton_Work <http://pl.dbpedia.org/resource/Milton_Work>, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://dbpedia.org/resource/Lisa_Nandy <http://dbpedia.org/resource/Lisa_Nandy>, iv: null > (2016-02-22 20:11:11,641) INFO : StatementBuffer.java:1773: term: http://wikidata.org/entity/Q578037 <http://wikidata.org/entity/Q578037>, iv: null > (2016-02-22 20:11:11,642) INFO : StatementBuffer.java:1773: term: http://wikidata.dbpedia.org/resource/Q578037 <http://wikidata.dbpedia.org/resource/Q578037>, iv: null > > > Is “iv:null” bad? > > I am loading 53 ttl-files of 150G > > /Joakim > > > >> On Feb 22, 2016, at 12:06 PM, Brad Bebee <be...@bl... <mailto:be...@bl...>> wrote: >> >> Joakim, >> >> You should see log output as the statements are loaded. How much data are you loading at once? >> >> Thanks, --Brad >> >> On Mon, Feb 22, 2016 at 2:59 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >> Thanks for the advice. Now it has been indexing for several days and I have no idea what it’s doing. >> >>> On Feb 22, 2016, at 9:04 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote: >>> >>> Try looking on the status tab of the blazegraph UI in the browser. In the detail view of your particular task, there might be a counter showing how many triples have been updated. >>> >>> (I am unsure as to which tasks support this under which versions …) >>> >>> Jeremy >>> >>> >>> >>>> On Feb 17, 2016, at 12:26 PM, Brad Bebee <be...@bl... <mailto:be...@bl...>> wrote: >>>> >>>> Joakim, >>>> >>>> With the DataLoader, the commit is after all of the data is loaded. Once the load is complete, all of the statements will be visible. >>>> >>>> Thanks, --Brad >>>> >>>> On Wed, Feb 17, 2016 at 3:21 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >>>> I am calling: >>>> >>>> curl -X POST --data-binary @dataloader.xml --header 'Content-Type:application/xml' http:/__.__.__:9999/blazegraph/dataloader >>>> >>>> I can see the size of the JNL-file is increasing, but when I query number of statements in the dashboard the data doesn’t show up. >>>> >>>> select (count(*) as ?num) { ?s ?p ?o } >>>> >>>> Do I need to Flush the StatementBuffer to the backing store after the curl? >>>> >>>> This is my config file: >>>> >>>> <?xml version="1.0" encoding="UTF-8" standalone="no"?> >>>> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd <http://java.sun.com/dtd/properties.dtd>"> >>>> <properties> >>>> <!-- RDF Format (Default is rdf/xml) --> >>>> <entry key="format">N-Triples</entry> >>>> <!-- Base URI (Optional) --> >>>> <entry key="baseURI"></entry> >>>> <!-- Default Graph URI (Optional - Required for quads mode namespace) --> >>>> <entry key="defaultGraph"></entry> >>>> <!-- Suppress all stdout messages (Optional) --> >>>> <entry key="quiet">false</entry> >>>> <!-- Show additional messages detailing the load performance. (Optional) --> >>>> <entry key="verbose">3</entry> >>>> <!-- Compute the RDF(S)+ closure. (Optional) --> >>>> <entry key="closure">false</entry> >>>> <!-- Files will be renamed to either .good or .fail as they are processed. >>>> The files will remain in the same directory. --> >>>> <entry key="durableQueues">true</entry> >>>> <!-- The namespace of the KB instance. Defaults to kb. --> >>>> <entry key="namespace">kb</entry> >>>> <!-- The configuration file for the database instance. It must be readable by the web application. --> >>>> <entry key="propertyFile">RWStore.properties</entry> >>>> <!-- Zero or more files or directories containing the data to be loaded. >>>> This should be a comma delimited list. The files must be readable by the web application. --> >>>> <entry key="fileOrDirs">/mydata/dbpedia2015/core/</entry> >>>> </properties> >>>> >>>> >>>> >>>> >>>>> On Feb 16, 2016, at 8:35 AM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >>>>> >>>>> I knew there is a DataLoader class, but I wasn’t aware it was available as a service in NanoSparql server. I will try it immediately >>>>> >>>>> >>>>> Thanks >>>>> Joakim >>>>> >>>>>> On Feb 16, 2016, at 8:09 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote: >>>>>> >>>>>>> See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load> >>>>>> >>>>>> >>>>>> That looks very interesting: >>>>>> >>>>>> I read: >>>>>> >>>>>> "Parsing, insert, and removal on the database are now decoupled from the index writes” >>>>>> >>>>>> One behavior we have is that we have small inserts concurrent with other activity (typically but not exclusively read activity). Does the enhanced configurability in 2.0 give us options that may allow us to improve performance of these writes. >>>>>> >>>>>> E.g. this week we have many (millions? at least hundreds of thousands) of such small writes (10 - 100 quads) and we also are trying to delete 25 million quads using about 100 delete/insert requests (that I take to be not impacted by this change). I am currently suggesting we should do one or the other at any one time, and not try to mix: but frankly I am guessing, and guessing conservatively. We have to maintain an always-on read performance at the same time. Total store size approx 3billion. >>>>>> >>>>>> [Unfortunately this machine is still a 1.5.3 machine, but for future reference I am trying to have better sense of how to organize such activity] >>>>>> >>>>>> Jeremy >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>> On Feb 16, 2016, at 7:55 AM, Bryan Thompson <br...@sy... <mailto:br...@sy...>> wrote: >>>>>>> >>>>>>> 2.0 includes support for bulk data load with a number of interesting features, including durable queue patterns, folders, etc. See https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load <https://wiki.blazegraph.com/wiki/index.php/REST_API#Bulk_Data_Load> >>>>>>> >>>>>>> ---- >>>>>>> Bryan Thompson >>>>>>> Chief Scientist & Founder >>>>>>> Blazegraph >>>>>>> e: br...@bl... <mailto:br...@bl...> >>>>>>> w: http://blazegraph.com <http://blazegraph.com/> >>>>>>> >>>>>>> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. >>>>>>> >>>>>>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. >>>>>>> >>>>>>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. >>>>>>> >>>>>>> >>>>>>> On Tue, Feb 16, 2016 at 10:40 AM, Jeremy J Carroll <jj...@sy... <mailto:jj...@sy...>> wrote: >>>>>>> >>>>>>> >>>>>>>> On Feb 15, 2016, at 10:42 PM, Joakim Soderberg <joa...@bl... <mailto:joa...@bl...>> wrote: >>>>>>>> >>>>>>>> Has anyone succeeded to load a folder of .nt files? I can load one by one: >>>>>>>> >>>>>>>> LOAD <file:///mydata/dbpedia2015/core/amsterdammuseum_links.nt <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> >>>>>>>> >>>>>>>> But it doesn’t like a folder name >>>>>>>> LOAD <file:///mydata/dbpedia2015/core/ <>> INTO GRAPH <http://dbpedia2015 <http://dbpedia2015/>> >>>>>>> >>>>>>> >>>>>>> That is correct. If you look at the spec for LOAD: >>>>>>> https://www.w3.org/TR/sparql11-update/#load <https://www.w3.org/TR/sparql11-update/#load> >>>>>>> then it takes an IRI as where you are loading from, and the concept of folder is simply not applicable. >>>>>>> A few schemes such as file: and ftp: may have such a notion, but the operation you are looking for is local to your machine on the client and you should probably implement it yourself. >>>>>>> >>>>>>> In particular, do you want each file loaded into a different graph or the same graph: probably best for you to make up your own mind. >>>>>>> >>>>>>> I have had success loading trig files into multiple graphs, using a simple POST to the endpoint. >>>>>>> >>>>>>> >>>>>>> Jeremy >>>>>>> >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance >>>>>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>>>>>> Monitor end-to-end web transactions and take corrective actions now >>>>>>> Troubleshoot faster and improve end-user experience. Signup Now! >>>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140> >>>>>>> _______________________________________________ >>>>>>> Bigdata-developers mailing list >>>>>>> Big...@li... <mailto:Big...@li...> >>>>>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>>> >>>> ------------------------------------------------------------------------------ >>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance >>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>>> Monitor end-to-end web transactions and take corrective actions now >>>> Troubleshoot faster and improve end-user experience. Signup Now! >>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140> >>>> _______________________________________________ >>>> Bigdata-developers mailing list >>>> Big...@li... <mailto:Big...@li...> >>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> >>>> >>>> >>>> >>>> >>>> -- >>>> _______________ >>>> Brad Bebee >>>> CEO >>>> Blazegraph >>>> e: be...@bl... <mailto:be...@bl...> >>>> m: 202.642.7961 <tel:202.642.7961> >>>> w: www.blazegraph.com <http://www.blazegraph.com/> >>>> >>>> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. >>>> >>>> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. >>>> >>>> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. >>>> >>>> ------------------------------------------------------------------------------ >>>> Site24x7 APM Insight: Get Deep Visibility into Application Performance >>>> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >>>> Monitor end-to-end web transactions and take corrective actions now >>>> Troubleshoot faster and improve end-user experience. Signup Now! >>>> http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________ <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140_______________________________________________> >>>> Bigdata-developers mailing list >>>> Big...@li... <mailto:Big...@li...> >>>> https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> >>> >> >> >> >> >> -- >> _______________ >> Brad Bebee >> CEO >> Blazegraph >> e: be...@bl... <mailto:be...@bl...> >> m: 202.642.7961 <tel:202.642.7961> >> w: www.blazegraph.com <http://www.blazegraph.com/> >> >> Blazegraph products help to solve the Graph Cache Thrash to achieve large scale processing for graph and predictive analytics. Blazegraph is the creator of the industry’s first GPU-accelerated high-performance database for large graphs, has been named as one of the “10 Companies and Technologies to Watch in 2016” <http://insideanalysis.com/2016/01/20535/>. >> >> Blazegraph Database <https://www.blazegraph.com/> is our ultra-high performance graph database that supports both RDF/SPARQL and Tinkerpop/Blueprints APIs. Blazegraph GPU <https://www.blazegraph.com/product/gpu-accelerated/> andBlazegraph DAS <https://www.blazegraph.com/product/gpu-accelerated/>L are disruptive new technologies that use GPUs to enable extreme scaling that is thousands of times faster and 40 times more affordable than CPU-based solutions. >> >> CONFIDENTIALITY NOTICE: This email and its contents and attachments are for the sole use of the intended recipient(s) and are confidential or proprietary to SYSTAP, LLC DBA Blazegraph. Any unauthorized review, use, disclosure, dissemination or copying of this email or its contents or attachments is prohibited. If you have received this communication in error, please notify the sender by reply email and permanently delete all copies of the email and its contents and attachments. >> > > > ------------------------------------------------------------------------------ > Site24x7 APM Insight: Get Deep Visibility into Application Performance > APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month > Monitor end-to-end web transactions and take corrective actions now > Troubleshoot faster and improve end-user experience. Signup Now! > http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140 <http://pubads.g.doubleclick.net/gampad/clk?id=272487151&iu=/4140> > _______________________________________________ > Bigdata-developers mailing list > Big...@li... <mailto:Big...@li...> > https://lists.sourceforge.net/lists/listinfo/bigdata-developers <https://lists.sourceforge.net/lists/listinfo/bigdata-developers> > > |
|
From: Joakim S. <joa...@bl...> - 2016-03-08 17:24:12
|
Hi, I have a client running org.openrdf.repository.sparql.SPARQLRepository After some thousands queries I get the following error: - *** error closing:org.openrdf.http.client.BackgroundTupleResult@70beb599 org.openrdf.query.QueryEvaluationException: org.apache.http.MalformedChunkCodingException: CRLF expected at end of chunk at org.openrdf.http.client.BackgroundTupleResult.handleClose(BackgroundTupleResult.java:79) at info.aduna.iteration.AbstractCloseableIteration.close(AbstractCloseableIteration.java:60) at bwIdEntityClassPopulate.main(bwIdEntityClassPopulate.java:167) Caused by: org.apache.http.MalformedChunkCodingException: CRLF expected at end of chunk at org.apache.http.impl.io.ChunkedInputStream.getChunkSize(ChunkedInputStream.java:255) at org.apache.http.impl.io.ChunkedInputStream.nextChunk(ChunkedInputStream.java:227) at org.apache.http.impl.io.ChunkedInputStream.read(ChunkedInputStream.java:186) at org.apache.http.impl.io.ChunkedInputStream.read(ChunkedInputStream.java:215) at org.apache.http.impl.io.ChunkedInputStream.close(ChunkedInputStream.java:316) at org.apache.http.impl.execchain.ResponseEntityProxy.streamClosed(ResponseEntityProxy.java:128) at org.apache.http.conn.EofSensorInputStream.checkClose(EofSensorInputStream.java:228) at org.apache.http.conn.EofSensorInputStream.close(EofSensorInputStream.java:174) at org.openrdf.http.client.BackgroundTupleResult.handleClose(BackgroundTupleResult.java:75) ... 2 more And the server is running WDQS: blazegraph-service-0.2.0-SNAPSHOT-dist.war data/ docs/ gui/ jetty-runner-9.2.9.v20150224.jar jolokia.sh* lib/ loadData.sh* munge.sh* nohup.out rules.log runBlazegraph.sh* runBlazegraph.sh~* #runUpdate.sh#* runUpdate.sh* runUpdate.sh~* I’m not sure if the problem originates from the server or the client? |