Hi,
Many sites provide RSS feeds for several reasons, usually to save bandwidth, to
give the users concentrated data and so forth.
Some of the RSS files supplied by sites are created specially for search
engines where each RSS "item" represent a web page in the site.
IMHO the only thing "missi
hello!
Does anybody know why I get the following error running ant on the
revision I have checked out from svn? Maybe its a dumb question but...
Thank you for your help!
compile:
[echo] Compiling plugin: parse-html
[javac] Compiling 5 source files to nutch/trunk/build/parse-html/classes
sorry , i will be careful .thx any way
On 1/31/07, chee wu <[EMAIL PROTECTED]> wrote:
set the two java arguments"-Dhadoop.log.file" and "-Dhadoop.log.dir"
should fix your problem.
btw,not to put much chinese characters in your mail..
- Original Message -
From: "kauu" <[EMAIL PROTECTED
hi ,
thx any way , but i don't think I tell clearly enough.
what i want is nutch just fetch rss seeds for 1 depth. So nutch should
just fetch some xml pages .I don't want to fetch the items' outlink 's
pages, because there r too much spam in those pages.
so , i just need to parse the rss fi
set the two java arguments"-Dhadoop.log.file" and "-Dhadoop.log.dir" should fix
your problem.
btw,not to put much chinese characters in your mail..
- Original Message -
From: "kauu" <[EMAIL PROTECTED]>
To:
Sent: Wednesday, January 31, 2007 1:45 PM
Subject: log4j problem
why when I c