|
From: Rob V. <rv...@do...> - 2014-06-16 16:02:50
|
Then please provide a minimal reproducible test case that shows the
problem so I can look into it
Thanks,
Rob
On 16/06/2014 16:55, "Tomasz Pluskiewicz" <tom...@gm...>
wrote:
>On Mon, Jun 16, 2014 at 4:55 PM, Rob Vesse <rv...@do...> wrote:
>> Which version is this? I'm assuming pre 1.0.4 or lower?
>>
>
>No, I got that behaviour in 1.0.5.x as well.
>
>>
>> Since you are talking about Store Manager then this sounds like
>>TOOLS-409
>> which was already reported and fixed for the 1.0.5 release. It was a
>>bug
>> in how Store Manager passed the data to the underlying writers.
>>
>
>This isn't only the Store Manager but also instances of TripleStore.
>When LoadFromFile() of LOAD <x> is used it's fine. But after updating
>a fresh store with INSERT DATA, wrong dataset will be serialized.
>
>>
>> Note that serializing an in-memory store directly in code was not
>>affected
>> in any way.
>>
>> The creation of the empty default graph is an implementation detail, any
>> SPARQL Update which inserts data potentially causes the default graph to
>> be created because the SPARQL specification states that a dataset always
>> contains an unnamed default graph and some parts of the implementation
>> assume graphs will already exist so it is safer and faster to pre-create
>> any graphs that will potentially be affected.
>>
>> LOAD is a special case because it is really just a shim to the parser
>> sub-system and the way the parser sub-system generates data means that
>> only specifically mentioned graphs are ever created.
>>
>> Rob
>>
>> On 16/06/2014 15:14, "Tomasz Pluskiewicz" <tom...@gm...>
>> wrote:
>>
>>>Hi
>>>
>>>We've noticed weird behaviour with in-memory triple store, when
>>>serializing to NQuads. Here's what happens:
>>>
>>>1. Create an empty TripleStore
>>>2. Run UPDATE
>>>
>>>INSERT DATA {
>>>GRAPH <http://test.org/user> {
>>><http://test.org/user>
>>><http://www.w3.org/1999/02/22-rdf-syntax-ns#type>
>>><http://schema.org/Person> .
>>><http://test.org/user> <http://some/ontology/favorite>
>>><http://test.org/product/name> .
>>>}
>>>GRAPH <http://test.org/prodList/> {
>>><http://test.org/user> <http://xmlns.com/foaf/0.1/primaryTopic>
>>><http://test.org/user> .
>>>}
>>>}
>>>
>>>3. Serialize to NQuads. Store Manager correctly informs that 3 triple
>>>were serialized in 3 graphs (including empty default graph).
>>>
>>>The output file contains all triples, but without graph names. Thus
>>>they are all serialized in the default graph. It's not a problem with
>>>the store in-memory. The insert creates correct graph with data.
>>>I've confirmed this occurs in all version since 1.0.0.
>>>
>>>Curiously only when data is loaded with a LOAD <x> INTO GRAPH <y>
>>>command or with dNetRDF API, the store is serialized correctly. Only
>>>INSERT DATA causes the problem.
>>>
>>>Is this a known problem?
>>>
>>>And by the way. Why INSERT DATA creates an empty default graph in the
>>>store while loading or LOAD <x> only creates those graphs actually
>>>included in the source files?
>>>
>>>Greets,
>>>Tom
>>>
>>>------------------------------------------------------------------------
>>>--
>>>----
>>>HPCC Systems Open Source Big Data Platform from LexisNexis Risk
>>>Solutions
>>>Find What Matters Most in Your Big Data with HPCC Systems
>>>Open Source. Fast. Scalable. Simple. Ideal for Dirty Data.
>>>Leverages Graph Analysis for Fast Processing & Easy Data Exploration
>>>http://p.sf.net/sfu/hpccsystems
>>>_______________________________________________
>>>dotNetRDF-bugs mailing list
>>>dot...@li...
>>>https://lists.sourceforge.net/lists/listinfo/dotnetrdf-bugs
>>
>>
>>
>>
>>
>>
>>-------------------------------------------------------------------------
>>-----
>> HPCC Systems Open Source Big Data Platform from LexisNexis Risk
>>Solutions
>> Find What Matters Most in Your Big Data with HPCC Systems
>> Open Source. Fast. Scalable. Simple. Ideal for Dirty Data.
>> Leverages Graph Analysis for Fast Processing & Easy Data Exploration
>> http://p.sf.net/sfu/hpccsystems
>> _______________________________________________
>> dotNetRDF-bugs mailing list
>> dot...@li...
>> https://lists.sourceforge.net/lists/listinfo/dotnetrdf-bugs
>
>--------------------------------------------------------------------------
>----
>HPCC Systems Open Source Big Data Platform from LexisNexis Risk Solutions
>Find What Matters Most in Your Big Data with HPCC Systems
>Open Source. Fast. Scalable. Simple. Ideal for Dirty Data.
>Leverages Graph Analysis for Fast Processing & Easy Data Exploration
>http://p.sf.net/sfu/hpccsystems
>_______________________________________________
>dotNetRDF-bugs mailing list
>dot...@li...
>https://lists.sourceforge.net/lists/listinfo/dotnetrdf-bugs
|