Thank you for the quick response, Laura. I'll try using the parallel
parameter on TDB2 and reply back once the job is complete.

On Tue, 12 Nov, 2019, 7:10 PM Laura Morales, <[email protected]> wrote:

> This (
> http://mail-archives.apache.org/mod_mbox/jena-users/201712.mbox/%3CCAHM9nqQfOnyhBQj=jr-i9ieqhiv7vflfnleoycmhsupdd7n...@mail.gmail.com%3E)
> is an old thread, somewhat related to this question.
>
>
>
>
> > Sent: Tuesday, November 12, 2019 at 1:29 PM
> > From: "Amandeep Srivastava" <[email protected]>
> > To: [email protected]
> > Subject: TDB optimization query
> >
> > Hi,
> >
> > I'm trying to create a TDB database from Wikidata's official RDF dump to
> > read the data using Fuseki service. I need to make a few queries for my
> > personal project, running which the online service times out.
> >
> > I have a 12 core machine with 36 GB memory.
> >
> > Can you please advise on the best way for creating the database? Since
> the
> > dump is huge, I cannot try all the approaches. Besides, I'm not sure if
> the
> > tdbloader function works in a similar way on data of different sizes.
> >
> > Questions:
> >
> > 1. Which one would be better to use - tdb.tdbloader2 (TDB1) or
> > tdb2.tdbloader (TDB2) for creating the database and why? Any specific
> > configurations that I should be aware of?
> >
> > 2. I'm running a job currently using tdb.tdbloader2 but it is using just
> a
> > single core. Also, it's loading speed is decreasing slowly. It started at
> > an avg of 120k tuples and is currently at 80k tuples. Can you advise how
> > can I utilize all the cores of my machine and maintain the loading speed
> at
> > the same time?
> >
> > Regards,
> > Aman
> >
>

Reply via email to