Hi Richard; Yes, a distributed project would be probably the best solution, but it is not easy to develop, unless you use a library like bittorrent, or similar and you have many peers. Althought most of the people don't seed the files long time, so sometimes is better to depend on a few committed persons than a big but ephemeral crowd.
Regards, emijrp 2011/6/26 Richard Farmbrough <rich...@farmbrough.co.uk> > ** > It would be useful to have an archive of archives. I have to delete my > old data dumps as time passes, for space reasons, however a team could, > between them, maintain multiple copies of every data dump. This would make a > nice distributed project. > > On 26/06/2011 13:53, emijrp wrote: > > Hi all; > > Can you imagine a day when Wikipedia is added to this list?[1] > > WikiTeam have developed a script[2] to download all the Wikipedia dumps > (and her sister projects) from dumps.wikimedia.org. It sorts in folders > and checks md5sum. It only works on Linux (it uses wget). > > You will need about 100GB to download all the 7z files. > > Save our memory. > > Regards, > emijrp > > [1] http://en.wikipedia.org/wiki/Destruction_of_libraries > [2] > http://code.google.com/p/wikiteam/source/browse/trunk/wikipediadownloader.py > > > _______________________________________________ > Xmldatadumps-l mailing > listXmldatadumps-l@lists.wikimedia.orghttps://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l > > >
_______________________________________________ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l