On Jun 20, 2019, at 08:37 AM, Adam Sanchez wrote:
>
> For your information
>
> ...
> b) It took 43 hours to load the Wikidata RDF dump
> (wikidata-20190610-all-BETA.ttl, 383G) in the dev version of Virtuoso
> 07.20.3230.
> I had to patch Virtuoso because it was given the following error each
> t
On 6/20/19 12:48 PM, hellm...@informatik.uni-leipzig.de wrote:
> Hi Adam,
> the server specs you posted are not so important. What disks did you use?
>
> They should be SSD or 15k RPM SAS to make it faster.
>
> Virtuoso can parse multi thread if you split the files before loading,
> but hdd speed i
Hi Adam,
the server specs you posted are not so important. What disks did you use?
They should be SSD or 15k RPM SAS to make it faster.
Virtuoso can parse multi thread if you split the files before loading, but hdd
speed is still the bottleneck.
Sebastian
On June 20, 2019 2:37:16 PM GMT+02:00
Is that with SSD harddisks? Isn't the bottleneck the io traffic to the
harddisks? (I suppose your are not loading into RAM?) What was you
hardware configuration?
best regards
Finn
http://people.compute.dtu.dk/faan/
On 20/06/2019 14:37, Adam Sanchez wrote:
For your information
a) It took 10.2
For your information
a) It took 10.2 days to load the Wikidata RDF dump
(wikidata-20190513-all-BETA.ttl, 379G) in Blazegraph 2.1.5.
The bigdata.jnl file turned to be 1.3T
Server technical features
Architecture: x86_64
CPU op-mode(s):32-bit, 64-bit
Byte Order:Little E
V4 has 8 cores instead of 6.
But well, it's a server grade config on purpose!
Vito
Il giorno mar 4 giu 2019 alle ore 16:32 Guillaume Lederrey <
gleder...@wikimedia.org> ha scritto:
> On Tue, Jun 4, 2019 at 3:14 PM Vi to wrote:
> >
> > AFAIR it's a double Xeon E5-2620 v3.
> > With modern CPUs f
On Tue, Jun 4, 2019 at 3:14 PM Vi to wrote:
>
> AFAIR it's a double Xeon E5-2620 v3.
> With modern CPUs frequency is not so significant.
Our latest batch of servers are: Intel(R) Xeon(R) CPU E5-2620 v4 @
2.10GHz (so v4 instead of v3, but the difference is probably minimal).
> Vito
>
> Il giorno
AFAIR it's a double Xeon E5-2620 v3.
With modern CPUs frequency is not so significant.
Vito
Il giorno mar 4 giu 2019 alle ore 13:00 Adam Sanchez
ha scritto:
> Thanks Guillaume!
> One question more, what is the CPU frequency (GHz)?
>
> Le mar. 4 juin 2019 à 12:25, Guillaume Lederrey
> a écrit :
Thanks Guillaume!
One question more, what is the CPU frequency (GHz)?
Le mar. 4 juin 2019 à 12:25, Guillaume Lederrey
a écrit :
>
> On Tue, Jun 4, 2019 at 12:18 PM Adam Sanchez wrote:
> >
> > Hello,
> >
> > Does somebody know the minimal hardware requirements (disk size and
> > RAM) for loading
thanks Guillaume. How does that compare to the wikidata footprint of the
wikidata service (SQL) not WDQS. I presume it sits in a MyISAM storage
container?
On Tue, Jun 4, 2019 at 11:25 AM Guillaume Lederrey
wrote:
> On Tue, Jun 4, 2019 at 12:18 PM Adam Sanchez
> wrote:
> >
> > Hello,
> >
> > Do
On Tue, Jun 4, 2019 at 12:18 PM Adam Sanchez wrote:
>
> Hello,
>
> Does somebody know the minimal hardware requirements (disk size and
> RAM) for loading wikidata dump in Blazegraph?
The actual hardware requirements will depend on your use case. But for
comparison, our production servers are:
*
Hello,
Does somebody know the minimal hardware requirements (disk size and
RAM) for loading wikidata dump in Blazegraph?
The downloaded dump file wikidata-20190513-all-BETA.ttl is 379G.
The bigdata.jnl file which stores all the triples data in Blazegraph
is 478G but still growing.
I had 1T disk b
12 matches
Mail list logo