DELETE {?s ?p ?oldValue} INSERT {?s ?p ?newValue} WHERE { OPTIONAL {?s ?p ?oldValue} #derive ?newValue from somewhere } If i want update 3 triples how to use this formats? Can you help me?
Vào 18:39, Th 4, 6 thg 7, 2022 Bartalus Gáspár <bartalus.gas...@codespring.ro.invalid> đã viết: > Hi, > > Most of the updates are DELETE/INSERT queries, i.e > > DELETE {?s ?p ?oldValue} > INSERT {?s ?p ?newValue} > WHERE { > OPTIONAL {?s ?p ?oldValue} > #derive ?newValue from somewhere > } > > We also have some separate DELETE queries and INSERT queries. > > I’ve tried HTTP POST /$/compact/db_name and as a result the files are > getting back to normal size. However, as far as I can tell the old files > are also kept. This is the folder structure I see: > - databases/db_name/Data-0001 - with the old large files > - databases/db_name/Data-0002 - presumably the result of the compact > operation with normal file sizes. > > Is there also some operation (http or cli) that would keep only one (the > latest) data folder, i.e. delete the old files from Data-0001? > > Gaspar > > > On 6 Jul 2022, at 12:52, Lorenz Buehmann < > buehm...@informatik.uni-leipzig.de> wrote: > > > > Ok, interesting > > > > so > > > > we have > > > > - 150k triples, rather small dataset > > > > - loaded into 10MB node table files > > > > - 10 updates every 5s > > > > - which makes up to 24 * 60 * 60 / 5 * 10 ~ 200k updates per day > > > > - and leads to 10GB node table files > > > > > > Can you share the shape of those update queries? > > > > > > After doing a "compact" operation, the files are getting back to > "normal" size? > > > > > > On 06.07.22 10:36, Bartalus Gáspár wrote: > >> Hi Lorenz, > >> > >> Thanks for quick feedback and clarification on lucene indexes. > >> > >> Here are my answers to your questions: > >> - We are uploading 7 ttl files to our dataset, where 1 is larger 6Mb, > the others are below 200Kb. > >> - The overall number of triples after data upload is ~150000. > >> - We have around 10 SPARQL UPDATE queries that are executed on a > regular and frequent basis, i.e. every 5 seconds. We also have 5 such > queries that are executed each minute. But most of the time they do not > produce any outcome, i.e. the dataset is not altered, and when they do, > there are just a couple of triples that are added to the dataset. > >> - These *.dat files start from ~10Mb in size, and after a day or so > some of them grow to ~10Gb. > >> > >> We have ~300 blank nodes, and ~half of the triples have a literal in > the object position, so ~75000. > >> > >> Best regards, > >> Gaspar > >> > >> > >> > >>> On 6 Jul 2022, at 10:55, Lorenz Buehmann < > buehm...@informatik.uni-leipzig.de> wrote: > >>> > >>> Hi and welcome Gaspar. > >>> > >>> > >>> Those files do contain the node tables. > >>> > >>> A Lucene index is never computed by default and would be contained in > Lucene specific index files. > >>> > >>> > >>> Can you give some details about the > >>> > >>> - size of the files > >>> - the number of triples > >>> - the number triples added/removed/changed > >>> - the frequency of updates > >>> - how much the files grow > >>> - what kind of data you insert? Lots of blank nodes? Or literals? > >>> > >>> Also, did you try a compact operation during time? > >>> > >>> Lorenz > >>> > >>> On 06.07.22 09:40, Bartalus Gáspár wrote: > >>>> Hi Jena support team, > >>>> > >>>> We are experiencing an issue with Jena Fuseki databases. In the > databases folder we see some files called SPO.dat, OSP.dat, etc., and the > size of these files are growing quickly. From our understanding these files > are containing the Lucene indexes. We would have two questions: > >>>> > >>>> 1. Why are these files growing rapidly, although the underlying data > (triples) are not being changed, or only slightly changed? > >>>> 2. Can we disable indexing easily, since we are not using full text > searches in our SPARQL queries? > >>>> > >>>> Our usage of Jena Fuseki: > >>>> > >>>> * Start the server with `fuseki-server —port 3030` > >>>> * Create databases with HTTP POST to > `/$/datasets?state=active&dbType=tdb2&dbName=db_name` > >>>> * Upload ttl files with HTTP POST to /db_name/data > >>>> > >>>> Thanks in advance for your feedback, and if you’d require more input > from our side, please let me know. > >>>> > >>>> Best regards, > >>>> Gaspar Bartalus > >>>> > >