> Now, to be even more useful, database dumps should be produced on > *regular* intervals. That way, we can compare various measures > such as article growth, link counts or usage of certain words, > without having to introduce the exact dump time in the count.
On a related note: I noticed that the meta-info dumps like stub-meta-history.xml.gz etc appear to be generated from the full history dump - and thus fail if the full history dump fails, and get delayed if the full history dump gets delayed. There are a lot of things that can be done with the meta-info alone, and it seems that dump should be easy and fast to generate. So I propose to genereate it from the database directly, instead of making it depend on the full history dump, which is slow and the most likely to break. -- daniel _______________________________________________ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l