Hi Joshua, Thanks for your input.
Le jeu. 16 mai 2019 à 17:02, Joshua Shinavier <j...@fortytwo.net> a écrit : > Hi Amirouche, > > The version history and time-travel features sound a lot like the > "integrated versioning system" of Freebase, circa 2009 when they (Metaweb) > presented at WWW. > Reading through [0] it seems freebase only allowed undo, whereas in datae it will be possible to query full history and undo commits. [0] https://www.aaai.org/Papers/AAAI/2007/AAAI07-355.pdf > As Freebase's data was transferred to Wikidata, this sounds a little > circular; I wonder what advantages datae would offer vis-a-vis Freebase. > Freebase data was transfered to wikidata. What I am looking for is replacing wikibase + blazegraph that is both edition and query would happen against the same database. Making it much easier to setup and maintain. > Disclaimer: this is coming from a Wikidata lurker who just happens to like > the Freebase approach to versioning of knowledge graphs, similar to what > you have described. > > Josh > > > On Sat, May 4, 2019 at 5:49 AM Amirouche Boubekki < > amirouche.boube...@gmail.com> wrote: > >> Le sam. 4 mai 2019 à 04:00, Yuri Astrakhan <yuriastrak...@gmail.com> a >> écrit : >> >>> Sounds interesting, is there a github repo? >>> >> >> Thanks for your interest. This is still a work-in-progress. I made a >> prototype that demos >> that the history significance measure allows to do time travelling >> queries in the v0 branch. >> Right now, I am working on getting it all together. >> >> >> https://github.com/awesome-data-distribution/datae/tree/master/docs/SCHEME20XY#abstract >> >> On Fri, May 3, 2019 at 8:19 PM Amirouche Boubekki < >>> amirouche.boube...@gmail.com> wrote: >>> >>>> GerardM post triggered my interest to post to the mailing list. As you >>>> might know I am working on functional quadstore that is quadstore that >>>> keeps around old version of data, like a wiki but in direct-acyclic-graph. >>>> It only stores differences between commits. It rely on snapshot of the >>>> latest version for fast reads. My ultimate goal is to build somekind of >>>> portable knowlege base. That is something like WikiBase + blazegraph but >>>> that you spinup on regular machine with the press of button. >>>> >>>> Enought brag about me. I wont't reply to all the message of the threads >>>> one by one but: >>>> >>>> Here is what SHOULD BE possible: >>>> >>>> - incremental dumps >>>> - time traveling queries >>>> - full dumps >>>> - The federation of wikibase SHOULD BE possible since it stored in a >>>> history like GIT and git pull git push are planned in the ROADMAP >>>> >>>> And online edition of the quadstore. >>>> >>>> Access Control List are not designed yet, I except that this should be >>>> enforced by the application layer. >>>> >>>> I planned start working on Data Management System (something like CKAN) >>>> with search featrure. But I would gadly work with wikimedia instead. >>>> >>>> Also, given it modeled after git, one can do merge-request like >>>> features, ie. exit the massive import that is crippled. >>>> >>>> What I would need is logs possibly with timing of queries (read and >>>> write) to do benchmarks. >>>> >>> >> Request: >> >> - Is it possible to have logs of read queries done against blazegraph >> with timings? >> - Is it possible to have logs of write queries done against mysql >> with timings? >> >> In the best of the worlds, it would be best to have logs to replicate the >> workload of the producion databases. >> >> >>> Maybe I should ask for fund at mediawiki? >>>> >>> >> What about this? Any possibility to have my project funded by the >> foundation somehow? >> >> >>> >>>> FWIW, I got 2 times faster than blazegraph on microbenchmark. >>>> >>>> >>>>> Hoi, >>>>> Wikidata grows like mad. This is something we all experience in the >>>>> really bad response times we are suffering. It is so bad that people are >>>>> asked what kind of updates they are running because it makes a difference >>>>> in the lag times there are. >>>>> >>>>> Given that Wikidata is growing like a weed, it follows that there are >>>>> two issues. Technical - what is the maximum that the current approach >>>>> supports - how long will this last us. Fundamental - what funding is >>>>> available to sustain Wikidata. >>>>> >>>>> For the financial guys, growth like Wikidata is experiencing is not >>>>> something you can reliably forecast. As an organisation we have more money >>>>> than we need to spend, so there is no credible reason to be stingy. >>>>> >>>>> For the technical guys, consider our growth and plan for at least one >>>>> year. When the impression exists that the current architecture will not >>>>> scale beyond two years, start a project to future proof Wikidata. >>>>> >>>>> It will grow and the situation will get worse before it gets better. >>>>> Thanks, >>>>> GerardM >>>>> >>>>> PS I know about phabricator tickets, they do not give the answers to >>>>> the questions we need to address. >>>>> >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> Wikidata mailing list >>>> Wikidata@lists.wikimedia.org >>>> https://lists.wikimedia.org/mailman/listinfo/wikidata >>>> >>> _______________________________________________ >>> Wikidata mailing list >>> Wikidata@lists.wikimedia.org >>> https://lists.wikimedia.org/mailman/listinfo/wikidata >>> >> _______________________________________________ >> Wikidata mailing list >> Wikidata@lists.wikimedia.org >> https://lists.wikimedia.org/mailman/listinfo/wikidata >> > _______________________________________________ > Wikidata mailing list > Wikidata@lists.wikimedia.org > https://lists.wikimedia.org/mailman/listinfo/wikidata >
_______________________________________________ Wikidata mailing list Wikidata@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata