I found a possible solution. I ended up modifying the nutch-default.xml file to hard code the plugins.folder path.[0] If everyone ends up doing the same thing, then we should add it it in installation guide.[1]
[0] <property> <name>plugin.folders</name> <!-- <value>plugins</value> --> * <value>/home/prashant/workspaceNutchTrunk/trunk/build/plugins</value>* <description>Directories where nutch plugins are located. Each element may be a relative or absolute path. If absolute, it is used as is. If relative, it is searched for on the classpath.</description> </property> This was already discussed in Nutch JIRA portal. https://issues.apache.org/jira/browse/NUTCH-937 [1] http://wiki.apache.org/nutch/RunNutchInEclipse On Sun, Dec 2, 2012 at 5:47 PM, Prashant Ladha <prashant.la...@gmail.com>wrote: > Hi Markus. > After sending the email, I again went through the instructions on [0] > link. > Then I saw the instruction to look at hadoop.log file so looking at the > log file, I found [1] > Are there any native Hadoop libraries that we have to install? > > I am on Ubuntu 12.1, JDK 1.7 & Trunk Nutch. > > [0] http://wiki.apache.org/nutch/RunNutchInEclipse > [1] attached hadoop.log > > > On Sun, Dec 2, 2012 at 5:45 PM, Markus Jelsma > <markus.jel...@openindex.io>wrote: > >> hi - Please provide log output and version number. >> >> -----Original message----- >> > From:Prashant Ladha <prashant.la...@gmail.com> >> > Sent: Sun 02-Dec-2012 23:37 >> > To: user@nutch.apache.org >> > Subject: Local Trunk Build - java.io.IOException: Job failed! >> > >> > Hi, >> > Earlier, since I was on Windows7 and seeing some exception that nobody >> saw >> > so I moved to Ubuntu. >> > But here, I am seeing the below error message: >> > Can you help in finding out what I could be doing wrong? >> > >> > >> > SLF4J: Class path contains multiple SLF4J bindings. >> > SLF4J: Found binding in >> > >> [jar:file:/home/prashant/.ivy2/cache/org.slf4j/slf4j-log4j12/jars/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> > SLF4J: Found binding in >> > >> [jar:file:/home/prashant/workspaceNutchTrunk/trunk/build/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an >> > explanation. >> > solrUrl is not set, indexing will be skipped... >> > crawl started in: crawl >> > rootUrlDir = urls >> > threads = 10 >> > depth = 3 >> > solrUrl=null >> > topN = 50 >> > Injector: starting at 2012-12-02 17:25:13 >> > Injector: crawlDb: crawl/crawldb >> > Injector: urlDir: urls >> > Injector: Converting injected urls to crawl db entries. >> > Injector: total number of urls rejected by filters: 0 >> > Injector: total number of urls injected after normalization and >> filtering: 0 >> > Injector: Merging injected urls into crawl db. >> > Injector: finished at 2012-12-02 17:25:28, elapsed: 00:00:14 >> > Generator: starting at 2012-12-02 17:25:28 >> > Generator: Selecting best-scoring urls due for fetch. >> > Generator: filtering: true >> > Generator: normalizing: true >> > Generator: topN: 50 >> > Generator: jobtracker is 'local', generating exactly one partition. >> > Exception in thread "main" java.io.IOException: Job failed! >> > at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1265) >> > at org.apache.nutch.crawl.Generator.generate(Generator.java:551) >> > at org.apache.nutch.crawl.Generator.generate(Generator.java:456) >> > at org.apache.nutch.crawl.Crawl.run(Crawl.java:130) >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >> > at org.apache.nutch.crawl.Crawl.main(Crawl.java:55) >> > >> > >