Hi,all I deploed two computer for nutch crawl as 'NutchHadoop Tutorial' tells.Butwhen I run nutch crawl the fetch job dosen't get data at all. This is a part of crawl.log
rootUrlDir = urls threads = 6 depth = 7 Injector: starting Injector: crawlDb: crawled/crawldb Injector: urlDir: urls Injector: Converting injected urls to crawl db entries. Injector: Merging injected urls into crawl db. Injector: done Generator: starting Generator: segment: crawled/segments/20070912141603 Generator: Selecting best-scoring urls due for fetch. Generator: Partitioning selected urls by host, for politeness. Generator: done. Fetcher: starting Fetcher: segment: crawled/segments/20070912141603 Fetcher: done CrawlDb update: starting CrawlDb update: db: crawled/crawldb CrawlDb update: segment: crawled/segments/20070912141603 CrawlDb update: Merging segment data into db. CrawlDb update: done Generator: starting …… The day befor yesterday the fetch job got some data,but job failed quickly at my first fetch.It prints this java.io.IOException: Job failed! at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:357) at org.apache.nutch.fetcher.Fetcher.fetch(Fetcher.java:562) at org.apache.nutch.crawl.Crawl.Crawler(Crawl.java:135) at org.apache.nutch.crawl.Crawl.ReplyPNo1Command(Crawl.java:325) at org.apache.nutch.crawl.Crawl.run(Crawl.java:436) Can you tell me what's wrong with my work.Thanks alot Oh, we use redhat9.0 and nutch0.8.1.