This 
(http://mail-archives.apache.org/mod_mbox/jena-users/201712.mbox/%3CCAHM9nqQfOnyhBQj=jr-i9ieqhiv7vflfnleoycmhsupdd7n...@mail.gmail.com%3E)
 is an old thread, somewhat related to this question.




> Sent: Tuesday, November 12, 2019 at 1:29 PM
> From: "Amandeep Srivastava" <[email protected]>
> To: [email protected]
> Subject: TDB optimization query
>
> Hi,
>
> I'm trying to create a TDB database from Wikidata's official RDF dump to
> read the data using Fuseki service. I need to make a few queries for my
> personal project, running which the online service times out.
>
> I have a 12 core machine with 36 GB memory.
>
> Can you please advise on the best way for creating the database? Since the
> dump is huge, I cannot try all the approaches. Besides, I'm not sure if the
> tdbloader function works in a similar way on data of different sizes.
>
> Questions:
>
> 1. Which one would be better to use - tdb.tdbloader2 (TDB1) or
> tdb2.tdbloader (TDB2) for creating the database and why? Any specific
> configurations that I should be aware of?
>
> 2. I'm running a job currently using tdb.tdbloader2 but it is using just a
> single core. Also, it's loading speed is decreasing slowly. It started at
> an avg of 120k tuples and is currently at 80k tuples. Can you advise how
> can I utilize all the cores of my machine and maintain the loading speed at
> the same time?
>
> Regards,
> Aman
>

Reply via email to