On Tue, Jul 21, 2009 at 12:47 PM, Aryeh Gregor <
simetrical+wikil...@gmail.com <simetrical%2bwikil...@gmail.com>> wrote:

> On Tue, Jul 21, 2009 at 11:22 AM, Chengbin Zheng<chengbinzh...@gmail.com>
> wrote:
> > On a side note, if parsing the XML gets you the static HTML version of
> > Wikipedia, why can't Wikimedia just parse it for us and save a lot of our
> > time (parsing and learning), and use that as the static HTML dump
> version?
>
> I'd assume it was a performance issue to parse all the pages for all
> the dumps so often.  It might have just used too much CPU to be worth
> it at the time.  Parsing some individual pages can take 20 seconds or
> more, and there are millions of them (although most much faster to
> parse than that).  I'm sure it could be reinstituted with some effort,
> though.
>
> _______________________________________________
> Wikitech-l mailing list
> Wikitech-l@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
>

Wouldn't parsing it be faster than actually creating that many HTMLs?
_______________________________________________
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to