Well, PubSubHubbub is a nice idea. However it clearly depends on two
factors:
1. whether Wikidata sets up such an infrastructure (I need to check
whether we have capacities, I am not sure atm)
2. whether performance is good enough to handle high-volume publishers
Basically, polling to recent changes [1] and then do a http request to
the individual pages should be fine for a start. So I guess this is what
we will implement, if there aren't any better suggestions.
The whole issue is problematic and the DBpedia project would be happy,
if this were discussed and decided right now, so we can plan development.
What is the best practice to get updates from Wikipedia at the moment?
We are still using OAI-PMH...
In DBpedia, we use a simple self-created protocol:
http://wiki.dbpedia.org/DBpediaLive#h156-4
/Publication of changesets/: Upon modifications old triples
are replaced with updated triples. Those added and/or deleted triples
are also written as N-Triples files and then compressed. Any client
application or DBpedia-Live mirror can download those files
and integrate and, hence, update a local copy of DBpedia. This enables
that application to always in synchronization with our DBpedia-Live.
This could also work for Wikidata facts, right?
Other useful links:
- http://www.openarchives.org/rs/0.5/resourcesync
- http://www.sdshare.org/
- http://www.w3.org/community/sdshare/
- http://www.rabbitmq.com/
All the best,
Sebastian
[1]
https://www.wikidata.org/w/index.php?title=Special:RecentChanges&feed=atom
Am 26.04.2013 03:15, schrieb Hady elsahar:
Hello Dimirtis
what do you thing of that ?
shall i write this part as an abstract part in the proposal and wait
for more details ,
or could we have a smiliar plan like the one already implemented in
dbpedia http://wiki.dbpedia.org/DBpediaLive#h156-3
thanks
regards
On Fri, Apr 26, 2013 at 12:50 AM, Jeremy Baron <jer...@tuxmachine.com
<mailto:jer...@tuxmachine.com>> wrote:
On Thu, Apr 25, 2013 at 10:42 PM, Hady elsahar
<hadyelsa...@gmail.com <mailto:hadyelsa...@gmail.com>> wrote:
> 2- is there any design pattern or a brief outline for the
change propagation design , how it would be ? in order that i
could make a rough plan and estimation about how it could be
consumed from the DBpedia side ?
I don't know anything about the plan for this but it seems at first
glance like a good place to use [[w:PubSubHubbub]].
-Jeremy
_______________________________________________
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org <mailto:Wikidata-l@lists.wikimedia.org>
https://lists.wikimedia.org/mailman/listinfo/wikidata-l
--
-------------------------------------------------
Hady El-Sahar
Research Assistant
Center of Informatics Sciences | Nile University
<http://nileuniversity.edu.eg/>
email : hadyelsa...@gmail.com <mailto:hadyelsa...@gmail.com>
Phone : +2-01220887311 <tel:%2B2-01220887311>
http://hadyelsahar.me/
<http://www.linkedin.com/in/hadyelsahar>
_______________________________________________
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l
--
Dipl. Inf. Sebastian Hellmann
Department of Computer Science, University of Leipzig
Projects: http://nlp2rdf.org , http://linguistics.okfn.org ,
http://dbpedia.org/Wiktionary , http://dbpedia.org
Homepage: http://bis.informatik.uni-leipzig.de/SebastianHellmann
Research Group: http://aksw.org
_______________________________________________
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l