* emijrp <emi...@gmail.com> [Thu, 11 Nov 2010 16:04:53 +0100]:
> There are some old dumps in Internet Archive,[1] but I guess you are
> interested in the most recent ones.
>
> Also, I have a copy of all the pages-meta-history.xml.7z from August
> 2010 at
> home. But I can't upload them anywhere, they are 100 GB.
>
Why such large dumps are not incremental ones? That could bring:
1. Save a lot of disk space
2. Greatly increase speed of dumping
3. Easier to upload and restore missed parts (something like P2P)

With XML import / export that could be done quite easily. And even with 
SQL you probably can limit entries at least of some tables (revision, 
user, page). Though I am unsure, if it is possible to make incremental 
dumps of all the tables - that would require SQL operations log, instead 
of simple range inserts, probably.

Although the developers are smart people so probably I am wasting the 
time and there is no reason to make incremental backups?
Dmitriy

_______________________________________________
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to