: Samuel Klein sjkl...@hcs.harvard.edu
Subject: Re: [Wiki-research-l] Wikipedia dumps downloader
Thank you, Emijrp!
What about the dump of Commons images? =A0 [for those with 10TB to spare]
SJ
On Sun, Jun 26, 2011 at 8:53 AM, emijrp emi...@gmail.com wrote:
Hi all;
Can you imagine a day when
Hi;
@Derrick: I don't trust Amazon. Really, I don't trust Wikimedia Foundation
either. They can't and/or they don't want to provide image dumps (what is
worst?). Community donates images to Commons, community donates money every
year, and now community needs to develop a software to extract all
Hi all;
Can you imagine a day when Wikipedia is added to this list?[1]
WikiTeam have developed a script[2] to download all the Wikipedia dumps (and
her sister projects) from dumps.wikimedia.org. It sorts in folders and
checks md5sum. It only works on Linux (it uses wget).
You will need about