Hannah_Bast added a comment.

  I agree with Kingsley that you don't need a distributed SPARQL engine when 
the knowledge graph fits on a single machine and will do so also in the future. 
Which is clearly the case for Wikidata, since it's even the case for the ten 
times larger UniProt (which at the time of this writing already contains over 
90 billion triples).
  
  In fact, I would consider distributing the knowledge graph over multiple 
machines in such a scenario suboptimal because distributing (as opposed to just 
replicating) the data incurs a significant performance overhead. You distribute 
only if you need to.
  
  As Jerven pointed out, if you have a high query load, you can just replicate 
the knowledge graph on multiple machines (one copy per machine) and distribute 
the queries over these machines. This is simple and effective.

TASK DETAIL
  https://phabricator.wikimedia.org/T206561

EMAIL PREFERENCES
  https://phabricator.wikimedia.org/settings/panel/emailpreferences/

To: Hannah_Bast
Cc: Hannah_Bast, KingsleyIdehen, So9q, TallTed, Sj, Jerven, Base, TomT0m, 
Akuckartz, GreenReaper, Addshore, Lucas_Werkmeister_WMDE, Aklapper, Smalyshev, 
Invadibot, MPhamWMF, maantietaja, CBogen, Nandana, Namenlos314, Lahi, Gq86, 
GoranSMilovanovic, QZanden, EBjune, merbst, LawExplorer, _jensen, rosalieper, 
Scott_WUaS, Jonas, Xmlizer, jkroll, Wikidata-bugs, Jdouglas, aude, Tobias1984, 
Manybubbles, Mbch331
_______________________________________________
Wikidata-bugs mailing list -- wikidata-bugs@lists.wikimedia.org
To unsubscribe send an email to wikidata-bugs-le...@lists.wikimedia.org

Reply via email to