DELETE {?s ?p ?oldValue}
INSERT {?s ?p ?newValue}
WHERE {
  OPTIONAL {?s ?p ?oldValue}
  #derive ?newValue from somewhere
}
If i want update 3 triples how to use this formats?
Can you help me?


Vào 18:39, Th 4, 6 thg 7, 2022 Bartalus Gáspár
<bartalus.gas...@codespring.ro.invalid> đã viết:

> Hi,
>
> Most of the updates are DELETE/INSERT queries, i.e
>
> DELETE {?s ?p ?oldValue}
> INSERT {?s ?p ?newValue}
> WHERE {
>   OPTIONAL {?s ?p ?oldValue}
>   #derive ?newValue from somewhere
> }
>
> We also have some separate DELETE queries and INSERT queries.
>
> I’ve tried HTTP POST /$/compact/db_name and as a result the files are
> getting back to normal size. However, as far as I can tell the old files
> are also kept. This is the folder structure I see:
> - databases/db_name/Data-0001 - with the old large files
> - databases/db_name/Data-0002 - presumably the result of the compact
> operation with normal file sizes.
>
> Is there also some operation (http or cli) that would keep only one (the
> latest) data folder, i.e. delete the old files from Data-0001?
>
> Gaspar
>
> > On 6 Jul 2022, at 12:52, Lorenz Buehmann <
> buehm...@informatik.uni-leipzig.de> wrote:
> >
> > Ok, interesting
> >
> > so
> >
> > we have
> >
> > - 150k triples, rather small dataset
> >
> > - loaded into 10MB node table files
> >
> > - 10 updates every 5s
> >
> > - which makes up to 24 * 60 * 60 / 5 * 10 ~ 200k updates per day
> >
> > - and leads to 10GB node table files
> >
> >
> > Can you share the shape of those update queries?
> >
> >
> > After doing a "compact" operation, the files are getting back to
> "normal" size?
> >
> >
> > On 06.07.22 10:36, Bartalus Gáspár wrote:
> >> Hi Lorenz,
> >>
> >> Thanks for quick feedback and clarification on lucene indexes.
> >>
> >> Here are my answers to your questions:
> >> - We are uploading 7 ttl files to our dataset, where 1 is larger 6Mb,
> the others are below 200Kb.
> >> - The overall number of triples after data upload is  ~150000.
> >> - We have around 10 SPARQL UPDATE queries that are executed on a
> regular and frequent basis, i.e. every 5 seconds. We also have 5 such
> queries that are executed each minute. But most of the time they do not
> produce any outcome, i.e. the dataset is not altered, and when they do,
> there are just a couple of triples that are added to the dataset.
> >> - These *.dat files start from ~10Mb in size, and after a day or so
> some of them grow to ~10Gb.
> >>
> >> We have ~300 blank nodes, and ~half of the triples have a literal in
> the object position, so ~75000.
> >>
> >> Best regards,
> >> Gaspar
> >>
> >>
> >>
> >>> On 6 Jul 2022, at 10:55, Lorenz Buehmann <
> buehm...@informatik.uni-leipzig.de> wrote:
> >>>
> >>> Hi and welcome Gaspar.
> >>>
> >>>
> >>> Those files do contain the node tables.
> >>>
> >>> A Lucene index is never computed by default and would be contained in
> Lucene specific index files.
> >>>
> >>>
> >>> Can you give some details about the
> >>>
> >>> - size of the files
> >>> - the number of triples
> >>> - the number triples added/removed/changed
> >>> - the frequency of updates
> >>> - how much the files grow
> >>> - what kind of data you insert? Lots of blank nodes? Or literals?
> >>>
> >>> Also, did you try a compact operation during time?
> >>>
> >>> Lorenz
> >>>
> >>> On 06.07.22 09:40, Bartalus Gáspár wrote:
> >>>> Hi Jena support team,
> >>>>
> >>>> We are experiencing an issue with Jena Fuseki databases. In the
> databases folder we see some files called SPO.dat, OSP.dat, etc., and the
> size of these files are growing quickly. From our understanding these files
> are containing the Lucene indexes. We would have two questions:
> >>>>
> >>>> 1. Why are these files growing rapidly, although the underlying data
> (triples) are not being changed, or only slightly changed?
> >>>> 2. Can we disable indexing easily, since we are not using full text
> searches in our SPARQL queries?
> >>>>
> >>>> Our usage of Jena Fuseki:
> >>>>
> >>>> * Start the server with `fuseki-server —port 3030`
> >>>> * Create databases with HTTP POST to
> `/$/datasets?state=active&dbType=tdb2&dbName=db_name`
> >>>> * Upload ttl files with HTTP POST to /db_name/data
> >>>>
> >>>> Thanks in advance for your feedback, and if you’d require more input
> from our side, please let me know.
> >>>>
> >>>> Best regards,
> >>>> Gaspar Bartalus
> >>>>
>
>

Reply via email to