Scott, Nemo is referring to the dumpgenerator.py being broken on MediaWiki versions above 1.20, and it should not actually affect older MediaWiki versions.
You can safely continue with your grab. :) On Sat, Nov 10, 2012 at 12:45 PM, Scott Boyd <scottd...@gmail.com> wrote: > At this link: https://code.google.com/p/wikiteam/issues/detail?id=56 , at > the bottom, there is an entry by project member nemowiki that states: > > Comment 7 <https://code.google.com/p/wikiteam/issues/detail?id=56#c7>by > project member > nemowiki <https://code.google.com/u/101255742639286016490/>, Today (9 > hours ago) > > Fixed by emijrp in r806 > <https://code.google.com/p/wikiteam/source/detail?r=806>. :-) > > *Status:* Fixed > > So does that mean this problem that "It's completely broken" is now fixed? > I'm running a huge download of 64K+ page titles, and am now using the > "r806" version of dumpgenerator.py. The first 35K+ page titles were > downloaded with an older version). Both versions sure seem to be > downloading MORE than 500 pages per namespace, but I'm not sure, since I > don't know how you can tell if you are getting them all... > > So is it fixed or not? > > > On Fri, Nov 9, 2012 at 4:27 AM, Federico Leva (Nemo) > <nemow...@gmail.com>wrote: > >> It's completely broken: https://code.google.com/p/** >> wikiteam/issues/detail?id=56<https://code.google.com/p/wikiteam/issues/detail?id=56> >> It will download only a fraction of the wiki, 500 pages at most per >> namespace. >> >> >> > > -- Regards, Hydriz We've created the greatest collection of shared knowledge in history. Help protect Wikipedia. Donate now: http://donate.wikimedia.org
_______________________________________________ Pywikipedia-l mailing list Pywikipedia-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/pywikipedia-l