From: Tomasz P. <tom...@gm...> - 2014-06-16 16:05:09
|
Will do tomorrow. I've got an example on my work computer. Tom On Mon, Jun 16, 2014 at 6:01 PM, Rob Vesse <rv...@do...> wrote: > Then please provide a minimal reproducible test case that shows the > problem so I can look into it > > Thanks, > > Rob > > On 16/06/2014 16:55, "Tomasz Pluskiewicz" <tom...@gm...> > wrote: > >>On Mon, Jun 16, 2014 at 4:55 PM, Rob Vesse <rv...@do...> wrote: >>> Which version is this? I'm assuming pre 1.0.4 or lower? >>> >> >>No, I got that behaviour in 1.0.5.x as well. >> >>> >>> Since you are talking about Store Manager then this sounds like >>>TOOLS-409 >>> which was already reported and fixed for the 1.0.5 release. It was a >>>bug >>> in how Store Manager passed the data to the underlying writers. >>> >> >>This isn't only the Store Manager but also instances of TripleStore. >>When LoadFromFile() of LOAD <x> is used it's fine. But after updating >>a fresh store with INSERT DATA, wrong dataset will be serialized. >> >>> >>> Note that serializing an in-memory store directly in code was not >>>affected >>> in any way. >>> >>> The creation of the empty default graph is an implementation detail, any >>> SPARQL Update which inserts data potentially causes the default graph to >>> be created because the SPARQL specification states that a dataset always >>> contains an unnamed default graph and some parts of the implementation >>> assume graphs will already exist so it is safer and faster to pre-create >>> any graphs that will potentially be affected. >>> >>> LOAD is a special case because it is really just a shim to the parser >>> sub-system and the way the parser sub-system generates data means that >>> only specifically mentioned graphs are ever created. >>> >>> Rob >>> >>> On 16/06/2014 15:14, "Tomasz Pluskiewicz" <tom...@gm...> >>> wrote: >>> >>>>Hi >>>> >>>>We've noticed weird behaviour with in-memory triple store, when >>>>serializing to NQuads. Here's what happens: >>>> >>>>1. Create an empty TripleStore >>>>2. Run UPDATE >>>> >>>>INSERT DATA { >>>>GRAPH <http://test.org/user> { >>>><http://test.org/user> >>>><http://www.w3.org/1999/02/22-rdf-syntax-ns#type> >>>><http://schema.org/Person> . >>>><http://test.org/user> <http://some/ontology/favorite> >>>><http://test.org/product/name> . >>>>} >>>>GRAPH <http://test.org/prodList/> { >>>><http://test.org/user> <http://xmlns.com/foaf/0.1/primaryTopic> >>>><http://test.org/user> . >>>>} >>>>} >>>> >>>>3. Serialize to NQuads. Store Manager correctly informs that 3 triple >>>>were serialized in 3 graphs (including empty default graph). >>>> >>>>The output file contains all triples, but without graph names. Thus >>>>they are all serialized in the default graph. It's not a problem with >>>>the store in-memory. The insert creates correct graph with data. >>>>I've confirmed this occurs in all version since 1.0.0. >>>> >>>>Curiously only when data is loaded with a LOAD <x> INTO GRAPH <y> >>>>command or with dNetRDF API, the store is serialized correctly. Only >>>>INSERT DATA causes the problem. >>>> >>>>Is this a known problem? >>>> >>>>And by the way. Why INSERT DATA creates an empty default graph in the >>>>store while loading or LOAD <x> only creates those graphs actually >>>>included in the source files? >>>> >>>>Greets, >>>>Tom >>>> >>>>------------------------------------------------------------------------ >>>>-- >>>>---- >>>>HPCC Systems Open Source Big Data Platform from LexisNexis Risk >>>>Solutions >>>>Find What Matters Most in Your Big Data with HPCC Systems >>>>Open Source. Fast. Scalable. Simple. Ideal for Dirty Data. >>>>Leverages Graph Analysis for Fast Processing & Easy Data Exploration >>>>http://p.sf.net/sfu/hpccsystems >>>>_______________________________________________ >>>>dotNetRDF-bugs mailing list >>>>dot...@li... >>>>https://lists.sourceforge.net/lists/listinfo/dotnetrdf-bugs >>> >>> >>> >>> >>> >>> >>>------------------------------------------------------------------------- >>>----- >>> HPCC Systems Open Source Big Data Platform from LexisNexis Risk >>>Solutions >>> Find What Matters Most in Your Big Data with HPCC Systems >>> Open Source. Fast. Scalable. Simple. Ideal for Dirty Data. >>> Leverages Graph Analysis for Fast Processing & Easy Data Exploration >>> http://p.sf.net/sfu/hpccsystems >>> _______________________________________________ >>> dotNetRDF-bugs mailing list >>> dot...@li... >>> https://lists.sourceforge.net/lists/listinfo/dotnetrdf-bugs >> >>-------------------------------------------------------------------------- >>---- >>HPCC Systems Open Source Big Data Platform from LexisNexis Risk Solutions >>Find What Matters Most in Your Big Data with HPCC Systems >>Open Source. Fast. Scalable. Simple. Ideal for Dirty Data. >>Leverages Graph Analysis for Fast Processing & Easy Data Exploration >>http://p.sf.net/sfu/hpccsystems >>_______________________________________________ >>dotNetRDF-bugs mailing list >>dot...@li... >>https://lists.sourceforge.net/lists/listinfo/dotnetrdf-bugs > > > > > > ------------------------------------------------------------------------------ > HPCC Systems Open Source Big Data Platform from LexisNexis Risk Solutions > Find What Matters Most in Your Big Data with HPCC Systems > Open Source. Fast. Scalable. Simple. Ideal for Dirty Data. > Leverages Graph Analysis for Fast Processing & Easy Data Exploration > http://p.sf.net/sfu/hpccsystems > _______________________________________________ > dotNetRDF-bugs mailing list > dot...@li... > https://lists.sourceforge.net/lists/listinfo/dotnetrdf-bugs |