Hannah_Bast added a comment.
I agree with Kingsley that you don't need a distributed SPARQL engine when the knowledge graph fits on a single machine and will do so also in the future. Which is clearly the case for Wikidata, since it's even the case for the ten times larger UniProt (which at the time of this writing already contains over 90 billion triples). In fact, I would consider distributing the knowledge graph over multiple machines in such a scenario suboptimal because distributing (as opposed to just replicating) the data incurs a significant performance overhead. You distribute only if you need to. As Jerven pointed out, if you have a high query load, you can just replicate the knowledge graph on multiple machines (one copy per machine) and distribute the queries over these machines. This is simple and effective. TASK DETAIL https://phabricator.wikimedia.org/T206561 EMAIL PREFERENCES https://phabricator.wikimedia.org/settings/panel/emailpreferences/ To: Hannah_Bast Cc: Hannah_Bast, KingsleyIdehen, So9q, TallTed, Sj, Jerven, Base, TomT0m, Akuckartz, GreenReaper, Addshore, Lucas_Werkmeister_WMDE, Aklapper, Smalyshev, Invadibot, MPhamWMF, maantietaja, CBogen, Nandana, Namenlos314, Lahi, Gq86, GoranSMilovanovic, QZanden, EBjune, merbst, LawExplorer, _jensen, rosalieper, Scott_WUaS, Jonas, Xmlizer, jkroll, Wikidata-bugs, Jdouglas, aude, Tobias1984, Manybubbles, Mbch331
_______________________________________________ Wikidata-bugs mailing list -- [email protected] To unsubscribe send an email to [email protected]
