Re: [Xmldatadumps-l] Need WIkipedia Dumps

2019-05-27 Thread Aron Bergman
> I need the wikipedia dump of 2002 for my Machine Learning Thesis.
> Could you please provide me the data.

Is this what you are looking for?
https://dumps.wikimedia.org/archive/

Regards
Aron Bergman

___
Xmldatadumps-l mailing list
Xmldatadumps-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l


[Xmldatadumps-l] Questions regarding the index.json file

2019-05-10 Thread Aron Bergman

Hi,
I've recently taken interest in the Wikipedia data dumps. I'd like to 
download a subset of files when they are updated. On the Data dumps 
page[1] a monitoring file[2] is mentioned, but the file doesn't contain 
any data (except the "wiki" object).


I did some research and found the monitor.py script and some info in the 
relevant README [3]. If I've understood it correctly, a server will 
periodically run monitor.py which will create the index.json file.


Is this deployed now? Since the file exists with (very little) content 
I'd guess that monitor.py has been run.


[1] https://meta.wikimedia.org/wiki/Data_dumps#Monitoring_dump_generation
[2] https://dumps.wikimedia.org/index.json
[3] 
https://phabricator.wikimedia.org/source/operations-dumps/browse/master/xmldumps-backup/README


Regards
Aron Bergman

___
Xmldatadumps-l mailing list
Xmldatadumps-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l