On Sun, Jul 26, 2009 at 8:17 PM, Chengbin Zhengchengbinzh...@gmail.com wrote:
Anyone know how long it takes to create a static HTML dump? A month?
It would depend completely on your hardware.
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
On Sun, Jul 26, 2009 at 8:51 PM, K. Peachey p858sn...@yahoo.com.au wrote:
On Mon, Jul 27, 2009 at 10:17 AM, Chengbin Zhengchengbinzh...@gmail.com
wrote:
Anyone know how long it takes to create a static HTML dump? A month?
___
Wikitech-l mailing
* Tei oscar.vi...@gmail.com [Tue, 21 Jul 2009 19:42:45 +0200]:
On Tue, Jul 21, 2009 at 7:17 PM, Chengbin
Zhengchengbinzh...@gmail.com
wrote:
...
No, I know what parsing means. Even if it takes 2 days to parse
them,
wouldn't it be faster than to actually create a static HTML dump the
On Wed, Jul 22, 2009 at 8:15 AM, Dmitriy Sintsov ques...@rambler.ru wrote:
* Tei oscar.vi...@gmail.com [Tue, 21 Jul 2009 19:42:45 +0200]:
On Tue, Jul 21, 2009 at 7:17 PM, Chengbin
Zhengchengbinzh...@gmail.com
wrote:
...
No, I know what parsing means. Even if it takes 2 days to parse
On Wed, Jul 22, 2009 at 5:48 PM, Chengbin Zhengchengbinzh...@gmail.com wrote:
...
Yes, the TombRaider version is exactly the version I want for static
HTML.
Just curious, is
pages-articles.xml.bz2http://download.wikimedia.org/enwiki/20090713/enwiki-20090713-pages-articles.xml.bz2
like
a
On Wed, Jul 22, 2009 at 2:37 PM, Tei oscar.vi...@gmail.com wrote:
On Wed, Jul 22, 2009 at 5:48 PM, Chengbin Zhengchengbinzh...@gmail.com
wrote:
...
Yes, the TombRaider version is exactly the version I want for static
HTML.
Just curious, is
pages-articles.xml.bz2
On Wed, Jul 22, 2009 at 6:37 PM, Teioscar.vi...@gmail.com wrote:
At a point, Brion compressed it to 242 MB.
http://www.mail-archive.com/wikitech-l@lists.wikimedia.org/msg00358.html
It looks like it was Platonides, not Brion, and as far as I can tell,
Gregory Maxwell said his compression
On Wed, Jul 22, 2009 at 6:53 PM, Aryeh Gregor
simetrical+wikil...@gmail.comsimetrical%2bwikil...@gmail.com
wrote:
On Wed, Jul 22, 2009 at 6:37 PM, Teioscar.vi...@gmail.com wrote:
At a point, Brion compressed it to 242 MB.
On Tue, Jul 21, 2009 at 3:33 AM, Kwan Ting Chank...@ktchan.info wrote:
I know you want to avoid using command line, but in this case it's really
much simpler / only feasible choice to search the internet / ask around for
the right commands and issue that on the command line. It's only going to
On Mon, Jul 20, 2009 at 11:33 PM, Kwan Ting Chan k...@ktchan.info wrote:
Chengbin Zheng wrote:
Thank you for dropping by and sharing this information with us Tomasz!
It is good just knowing that it is in the queue. Have you considered
making
a version of static HTML Wikipedia where there
Actually, I do have to learn everything. I know absolutely
nothing about
HTML and all the stuff (Maybe I will when I take the computer
science course
in grade 10). Think of it this way, you have a radioactive
material decay
problem, where you want to find out how much mass is left
On Tue, Jul 21, 2009 at 9:37 AM, Lane, Ryan
ryan.l...@ocean.navo.navy.milwrote:
Actually, I do have to learn everything. I know absolutely
nothing about
HTML and all the stuff (Maybe I will when I take the computer
science course
in grade 10). Think of it this way, you have a
On Tue, Jul 21, 2009 at 11:22 AM, Chengbin Zhengchengbinzh...@gmail.com wrote:
On a side note, if parsing the XML gets you the static HTML version of
Wikipedia, why can't Wikimedia just parse it for us and save a lot of our
time (parsing and learning), and use that as the static HTML dump
On Tue, Jul 21, 2009 at 1:08 PM, Chengbin Zhengchengbinzh...@gmail.com wrote:
Wouldn't parsing it be faster than actually creating that many HTMLs?
Parsing it *is* creating the HTML files. That's what parsing means
in MediaWiki, converting wikitext to HTML. It's kind of a misnomer,
admittedly.
On Tue, Jul 21, 2009 at 1:11 PM, Aryeh Gregor
simetrical+wikil...@gmail.comsimetrical%2bwikil...@gmail.com
wrote:
On Tue, Jul 21, 2009 at 1:08 PM, Chengbin Zhengchengbinzh...@gmail.com
wrote:
Wouldn't parsing it be faster than actually creating that many HTMLs?
Parsing it *is* creating the
On Tue, Jul 21, 2009 at 1:17 PM, Chengbin Zhengchengbinzh...@gmail.com wrote:
No, I know what parsing means. Even if it takes 2 days to parse them,
wouldn't it be faster than to actually create a static HTML dump the
traditional way?
I don't know. I can only speculate. Whatever it is, it
On Tue, Jul 21, 2009 at 7:17 PM, Chengbin Zhengchengbinzh...@gmail.com wrote:
...
No, I know what parsing means. Even if it takes 2 days to parse them,
wouldn't it be faster than to actually create a static HTML dump the
traditional way?
If it is not, then what is the difficulty of making
wouldn't it be faster than to actually create a static HTML dump the
traditional way?
The content is wiki-text. It has to be parsed to be turned into HTML. There
isn't a more traditional way, because there is no other way.
Wouldn't it be possible to dump the parser cache instead of dumping
On Tue, Jul 21, 2009 at 1:42 PM, Teioscar.vi...@gmail.com wrote:
On Tue, Jul 21, 2009 at 7:17 PM, Chengbin Zhengchengbinzh...@gmail.com
wrote:
...
No, I know what parsing means. Even if it takes 2 days to parse them,
wouldn't it be faster than to actually create a static HTML dump the
On Tue, Jul 21, 2009 at 1:49 PM, Chad innocentkil...@gmail.com wrote:
On Tue, Jul 21, 2009 at 1:42 PM, Teioscar.vi...@gmail.com wrote:
On Tue, Jul 21, 2009 at 7:17 PM, Chengbin Zhengchengbinzh...@gmail.com
wrote:
...
No, I know what parsing means. Even if it takes 2 days to parse them,
On Tue, Jul 21, 2009 at 2:20 PM, Chengbin Zheng chengbinzh...@gmail.comwrote:
On Tue, Jul 21, 2009 at 1:49 PM, Chad innocentkil...@gmail.com wrote:
On Tue, Jul 21, 2009 at 1:42 PM, Teioscar.vi...@gmail.com wrote:
On Tue, Jul 21, 2009 at 7:17 PM, Chengbin Zhengchengbinzh...@gmail.com
On Tue, Jul 21, 2009 at 8:20 PM, Chengbin Zhengchengbinzh...@gmail.com wrote:
..
Why would you download Wikipedia? Internet is so readily available, and the
online version has images.
It obviusly don't make much sense for final users.
It has been discused before anyway..
On Mon, Jul 20, 2009 at 10:00 PM, Chengbin Zhengchengbinzh...@gmail.com wrote:
It seems that reply doesn't work. So I'll send a new message.
Since the static HTML Wikipedia is not updating (please update), and XML
updates like everyday, the logical choice is to go with XML. Is there any
way to
. . . I should mention, also, that I believe the one in charge of
dumps is Tomasz Finc. You may want to ask him about whether there are
plans to resume the static HTML dumps.
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
On Mon, Jul 20, 2009 at 6:41 PM, Aryeh Gregor
simetrical+wikil...@gmail.comsimetrical%2bwikil...@gmail.com
wrote:
. . . I should mention, also, that I believe the one in charge of
dumps is Tomasz Finc. You may want to ask him about whether there are
plans to resume the static HTML dumps.
On Mon, Jul 20, 2009 at 8:52 PM, Aryeh Gregor
simetrical+wikil...@gmail.comsimetrical%2bwikil...@gmail.com
wrote:
On Mon, Jul 20, 2009 at 11:08 PM, Chengbin Zhengchengbinzh...@gmail.com
wrote:
I tried through Wikipedia mail, and I can't reach him.
How do you use mediawiki? There are no
Chengbin Zheng wrote:
On Mon, Jul 20, 2009 at 6:41 PM, Aryeh Gregor
simetrical+wikil...@gmail.comsimetrical%2bwikil...@gmail.com
wrote:
. . . I should mention, also, that I believe the one in charge of
dumps is Tomasz Finc. You may want to ask him about whether there are
plans to resume
On Mon, Jul 20, 2009 at 10:21 PM, Tomasz Finc tf...@wikimedia.org wrote:
Chengbin Zheng wrote:
On Mon, Jul 20, 2009 at 6:41 PM, Aryeh Gregor
simetrical+wikil...@gmail.com simetrical%2bwikil...@gmail.com
simetrical%2bwikil...@gmail.com simetrical%252bwikil...@gmail.com
wrote:
. . . I
28 matches
Mail list logo