Hi Markus.
After sending the email, I again went through the instructions on [0] link.
Then I saw the instruction to look at hadoop.log file so looking at the log
file, I found [1]
Are there any native Hadoop libraries that we have to install?

I am on Ubuntu 12.1, JDK 1.7 & Trunk Nutch.

[0] http://wiki.apache.org/nutch/RunNutchInEclipse
[1] attached hadoop.log


On Sun, Dec 2, 2012 at 5:45 PM, Markus Jelsma <markus.jel...@openindex.io>wrote:

> hi - Please provide log output and version number.
>
> -----Original message-----
> > From:Prashant Ladha <prashant.la...@gmail.com>
> > Sent: Sun 02-Dec-2012 23:37
> > To: user@nutch.apache.org
> > Subject: Local Trunk Build - java.io.IOException: Job failed!
> >
> > Hi,
> > Earlier, since I was on Windows7 and seeing some exception that nobody
> saw
> > so I moved to Ubuntu.
> > But here, I am seeing the below error message:
> > Can you help in finding out what I could be doing wrong?
> >
> >
> > SLF4J: Class path contains multiple SLF4J bindings.
> > SLF4J: Found binding in
> >
> [jar:file:/home/prashant/.ivy2/cache/org.slf4j/slf4j-log4j12/jars/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: Found binding in
> >
> [jar:file:/home/prashant/workspaceNutchTrunk/trunk/build/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> > explanation.
> > solrUrl is not set, indexing will be skipped...
> > crawl started in: crawl
> > rootUrlDir = urls
> > threads = 10
> > depth = 3
> > solrUrl=null
> > topN = 50
> > Injector: starting at 2012-12-02 17:25:13
> > Injector: crawlDb: crawl/crawldb
> > Injector: urlDir: urls
> > Injector: Converting injected urls to crawl db entries.
> > Injector: total number of urls rejected by filters: 0
> > Injector: total number of urls injected after normalization and
> filtering: 0
> > Injector: Merging injected urls into crawl db.
> > Injector: finished at 2012-12-02 17:25:28, elapsed: 00:00:14
> > Generator: starting at 2012-12-02 17:25:28
> > Generator: Selecting best-scoring urls due for fetch.
> > Generator: filtering: true
> > Generator: normalizing: true
> > Generator: topN: 50
> > Generator: jobtracker is 'local', generating exactly one partition.
> > Exception in thread "main" java.io.IOException: Job failed!
> > at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1265)
> > at org.apache.nutch.crawl.Generator.generate(Generator.java:551)
> > at org.apache.nutch.crawl.Generator.generate(Generator.java:456)
> > at org.apache.nutch.crawl.Crawl.run(Crawl.java:130)
> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
> > at org.apache.nutch.crawl.Crawl.main(Crawl.java:55)
> >
>

Reply via email to