Hi SJ; You know that that is an old item in our TODO list ; )
I heard that Platonides developed a script for that task long time ago. Platonides, are you there? Regards, emijrp 2011/6/27 Samuel Klein <sjkl...@hcs.harvard.edu> > Thank you, Emijrp! > > What about the dump of Commons images? [for those with 10TB to spare] > > SJ > > On Sun, Jun 26, 2011 at 8:53 AM, emijrp <emi...@gmail.com> wrote: > > Hi all; > > > > Can you imagine a day when Wikipedia is added to this list?[1] > > > > WikiTeam have developed a script[2] to download all the Wikipedia dumps > (and > > her sister projects) from dumps.wikimedia.org. It sorts in folders and > > checks md5sum. It only works on Linux (it uses wget). > > > > You will need about 100GB to download all the 7z files. > > > > Save our memory. > > > > Regards, > > emijrp > > > > [1] http://en.wikipedia.org/wiki/Destruction_of_libraries > > [2] > > > http://code.google.com/p/wikiteam/source/browse/trunk/wikipediadownloader.py > > > > _______________________________________________ > > Wiki-research-l mailing list > > Wiki-research-l@lists.wikimedia.org > > https://lists.wikimedia.org/mailman/listinfo/wiki-research-l > > > > > > > > -- > Samuel Klein identi.ca:sj w:user:sj +1 617 529 > 4266 > > _______________________________________________ > Wiki-research-l mailing list > Wiki-research-l@lists.wikimedia.org > https://lists.wikimedia.org/mailman/listinfo/wiki-research-l >
_______________________________________________ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l