Re: [Wiki-research-l] Wikipedia dumps downloader

2011-06-28 Thread Derrick Coetzee
: Samuel Klein sjkl...@hcs.harvard.edu Subject: Re: [Wiki-research-l] Wikipedia dumps downloader Thank you, Emijrp! What about the dump of Commons images? =A0 [for those with 10TB to spare] SJ On Sun, Jun 26, 2011 at 8:53 AM, emijrp emi...@gmail.com wrote: Hi all; Can you imagine a day when

Re: [Wiki-research-l] Wikipedia dumps downloader

2011-06-28 Thread emijrp
Hi; @Derrick: I don't trust Amazon. Really, I don't trust Wikimedia Foundation either. They can't and/or they don't want to provide image dumps (what is worst?). Community donates images to Commons, community donates money every year, and now community needs to develop a software to extract all

[Wiki-research-l] Wikipedia dumps downloader

2011-06-26 Thread emijrp
Hi all; Can you imagine a day when Wikipedia is added to this list?[1] WikiTeam have developed a script[2] to download all the Wikipedia dumps (and her sister projects) from dumps.wikimedia.org. It sorts in folders and checks md5sum. It only works on Linux (it uses wget). You will need about