Hi, I'm trying to create a TDB database from Wikidata's official RDF dump to read the data using Fuseki service. I need to make a few queries for my personal project, running which the online service times out.
I have a 12 core machine with 36 GB memory. Can you please advise on the best way for creating the database? Since the dump is huge, I cannot try all the approaches. Besides, I'm not sure if the tdbloader function works in a similar way on data of different sizes. Questions: 1. Which one would be better to use - tdb.tdbloader2 (TDB1) or tdb2.tdbloader (TDB2) for creating the database and why? Any specific configurations that I should be aware of? 2. I'm running a job currently using tdb.tdbloader2 but it is using just a single core. Also, it's loading speed is decreasing slowly. It started at an avg of 120k tuples and is currently at 80k tuples. Can you advise how can I utilize all the cores of my machine and maintain the loading speed at the same time? Regards, Aman
