This is a generic shape. A real-world example would be: DELETE {?subject rdfs:label ?oldLabel} INSERT {?subject rdfs:label ?newLabel} WHERE { ?subject rdf:type SomeType. ?subject rdfs:label ?oldLabel. FILTER(?oldLabel IN (“oldLabel1”, “oldLabel2”, “oldLabel3)) BIND(CONCAT(?oldLabel, “_updated”) AS ?newLabel) }
> On 6 Jul 2022, at 14:50, Dương Hồ <koero5...@gmail.com> wrote: > > DELETE {?s ?p ?oldValue} > INSERT {?s ?p ?newValue} > WHERE { > OPTIONAL {?s ?p ?oldValue} > #derive ?newValue from somewhere > } > If i want update 3 triples how to use this formats? > Can you help me? > > > Vào 18:39, Th 4, 6 thg 7, 2022 Bartalus Gáspár > <bartalus.gas...@codespring.ro.invalid> đã viết: > >> Hi, >> >> Most of the updates are DELETE/INSERT queries, i.e >> >> DELETE {?s ?p ?oldValue} >> INSERT {?s ?p ?newValue} >> WHERE { >> OPTIONAL {?s ?p ?oldValue} >> #derive ?newValue from somewhere >> } >> >> We also have some separate DELETE queries and INSERT queries. >> >> I’ve tried HTTP POST /$/compact/db_name and as a result the files are >> getting back to normal size. However, as far as I can tell the old files >> are also kept. This is the folder structure I see: >> - databases/db_name/Data-0001 - with the old large files >> - databases/db_name/Data-0002 - presumably the result of the compact >> operation with normal file sizes. >> >> Is there also some operation (http or cli) that would keep only one (the >> latest) data folder, i.e. delete the old files from Data-0001? >> >> Gaspar >> >>> On 6 Jul 2022, at 12:52, Lorenz Buehmann < >> buehm...@informatik.uni-leipzig.de> wrote: >>> >>> Ok, interesting >>> >>> so >>> >>> we have >>> >>> - 150k triples, rather small dataset >>> >>> - loaded into 10MB node table files >>> >>> - 10 updates every 5s >>> >>> - which makes up to 24 * 60 * 60 / 5 * 10 ~ 200k updates per day >>> >>> - and leads to 10GB node table files >>> >>> >>> Can you share the shape of those update queries? >>> >>> >>> After doing a "compact" operation, the files are getting back to >> "normal" size? >>> >>> >>> On 06.07.22 10:36, Bartalus Gáspár wrote: >>>> Hi Lorenz, >>>> >>>> Thanks for quick feedback and clarification on lucene indexes. >>>> >>>> Here are my answers to your questions: >>>> - We are uploading 7 ttl files to our dataset, where 1 is larger 6Mb, >> the others are below 200Kb. >>>> - The overall number of triples after data upload is ~150000. >>>> - We have around 10 SPARQL UPDATE queries that are executed on a >> regular and frequent basis, i.e. every 5 seconds. We also have 5 such >> queries that are executed each minute. But most of the time they do not >> produce any outcome, i.e. the dataset is not altered, and when they do, >> there are just a couple of triples that are added to the dataset. >>>> - These *.dat files start from ~10Mb in size, and after a day or so >> some of them grow to ~10Gb. >>>> >>>> We have ~300 blank nodes, and ~half of the triples have a literal in >> the object position, so ~75000. >>>> >>>> Best regards, >>>> Gaspar >>>> >>>> >>>> >>>>> On 6 Jul 2022, at 10:55, Lorenz Buehmann < >> buehm...@informatik.uni-leipzig.de> wrote: >>>>> >>>>> Hi and welcome Gaspar. >>>>> >>>>> >>>>> Those files do contain the node tables. >>>>> >>>>> A Lucene index is never computed by default and would be contained in >> Lucene specific index files. >>>>> >>>>> >>>>> Can you give some details about the >>>>> >>>>> - size of the files >>>>> - the number of triples >>>>> - the number triples added/removed/changed >>>>> - the frequency of updates >>>>> - how much the files grow >>>>> - what kind of data you insert? Lots of blank nodes? Or literals? >>>>> >>>>> Also, did you try a compact operation during time? >>>>> >>>>> Lorenz >>>>> >>>>> On 06.07.22 09:40, Bartalus Gáspár wrote: >>>>>> Hi Jena support team, >>>>>> >>>>>> We are experiencing an issue with Jena Fuseki databases. In the >> databases folder we see some files called SPO.dat, OSP.dat, etc., and the >> size of these files are growing quickly. From our understanding these files >> are containing the Lucene indexes. We would have two questions: >>>>>> >>>>>> 1. Why are these files growing rapidly, although the underlying data >> (triples) are not being changed, or only slightly changed? >>>>>> 2. Can we disable indexing easily, since we are not using full text >> searches in our SPARQL queries? >>>>>> >>>>>> Our usage of Jena Fuseki: >>>>>> >>>>>> * Start the server with `fuseki-server —port 3030` >>>>>> * Create databases with HTTP POST to >> `/$/datasets?state=active&dbType=tdb2&dbName=db_name` >>>>>> * Upload ttl files with HTTP POST to /db_name/data >>>>>> >>>>>> Thanks in advance for your feedback, and if you’d require more input >> from our side, please let me know. >>>>>> >>>>>> Best regards, >>>>>> Gaspar Bartalus >>>>>> >> >>
smime.p7s
Description: S/MIME cryptographic signature