Hi Micah,

what is your configuration? do you have multiple nodes or is it a
single machine? what version of hadoop library are you using?

des

On 7/30/07, Micah Vivion <[EMAIL PROTECTED]> wrote:
> Greetings,
>
> So this one has me stumped a little bit. I am running a fairly simple
> Nutch Crawl on our local intranet site or on our partners intranet
> sites. Every now and then when doing a 'bin/nutch crawl urlfile -dir
> webindex/ -depth 5' I get an exception of:
> Optimizing index.
> Indexer: done
> Dedup: starting
> Dedup: adding indexes in: /home/mvivion/webindex/target.com/indexes
> Exception in thread "main" java.io.IOException: Job failed!
>          at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:
> 604)
>          at org.apache.nutch.indexer.DeleteDuplicates.dedup
> (DeleteDuplicates.java:439)
>          at org.apache.nutch.crawl.Crawl.main(Crawl.java:135)
>
> Has anyone see this before? Any solutions to resolve this crash?
>
> Thanks!!!
>

-------------------------------------------------------------------------
This SF.net email is sponsored by: Splunk Inc.
Still grepping through log files to find problems?  Stop.
Now Search log events and configuration files using AJAX and a browser.
Download your FREE copy of Splunk now >>  http://get.splunk.com/
_______________________________________________
Nutch-general mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to