I found a possible solution. I ended up modifying the nutch-default.xml
file to hard code the plugins.folder path.[0]
If everyone ends up doing the same thing, then we should add it it in
installation guide.[1]

[0]
<property>
  <name>plugin.folders</name>
  <!-- <value>plugins</value> -->
*  <value>/home/prashant/workspaceNutchTrunk/trunk/build/plugins</value>*
  <description>Directories where nutch plugins are located.  Each
  element may be a relative or absolute path.  If absolute, it is used
  as is.  If relative, it is searched for on the classpath.</description>
</property>

This was already discussed in Nutch JIRA portal.
https://issues.apache.org/jira/browse/NUTCH-937

[1]
http://wiki.apache.org/nutch/RunNutchInEclipse


On Sun, Dec 2, 2012 at 5:47 PM, Prashant Ladha <prashant.la...@gmail.com>wrote:

> Hi Markus.
> After sending the email, I again went through the instructions on [0]
> link.
> Then I saw the instruction to look at hadoop.log file so looking at the
> log file, I found [1]
> Are there any native Hadoop libraries that we have to install?
>
> I am on Ubuntu 12.1, JDK 1.7 & Trunk Nutch.
>
> [0] http://wiki.apache.org/nutch/RunNutchInEclipse
> [1] attached hadoop.log
>
>
> On Sun, Dec 2, 2012 at 5:45 PM, Markus Jelsma 
> <markus.jel...@openindex.io>wrote:
>
>> hi - Please provide log output and version number.
>>
>> -----Original message-----
>> > From:Prashant Ladha <prashant.la...@gmail.com>
>> > Sent: Sun 02-Dec-2012 23:37
>> > To: user@nutch.apache.org
>> > Subject: Local Trunk Build - java.io.IOException: Job failed!
>> >
>> > Hi,
>> > Earlier, since I was on Windows7 and seeing some exception that nobody
>> saw
>> > so I moved to Ubuntu.
>> > But here, I am seeing the below error message:
>> > Can you help in finding out what I could be doing wrong?
>> >
>> >
>> > SLF4J: Class path contains multiple SLF4J bindings.
>> > SLF4J: Found binding in
>> >
>> [jar:file:/home/prashant/.ivy2/cache/org.slf4j/slf4j-log4j12/jars/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > SLF4J: Found binding in
>> >
>> [jar:file:/home/prashant/workspaceNutchTrunk/trunk/build/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>> > explanation.
>> > solrUrl is not set, indexing will be skipped...
>> > crawl started in: crawl
>> > rootUrlDir = urls
>> > threads = 10
>> > depth = 3
>> > solrUrl=null
>> > topN = 50
>> > Injector: starting at 2012-12-02 17:25:13
>> > Injector: crawlDb: crawl/crawldb
>> > Injector: urlDir: urls
>> > Injector: Converting injected urls to crawl db entries.
>> > Injector: total number of urls rejected by filters: 0
>> > Injector: total number of urls injected after normalization and
>> filtering: 0
>> > Injector: Merging injected urls into crawl db.
>> > Injector: finished at 2012-12-02 17:25:28, elapsed: 00:00:14
>> > Generator: starting at 2012-12-02 17:25:28
>> > Generator: Selecting best-scoring urls due for fetch.
>> > Generator: filtering: true
>> > Generator: normalizing: true
>> > Generator: topN: 50
>> > Generator: jobtracker is 'local', generating exactly one partition.
>> > Exception in thread "main" java.io.IOException: Job failed!
>> > at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1265)
>> > at org.apache.nutch.crawl.Generator.generate(Generator.java:551)
>> > at org.apache.nutch.crawl.Generator.generate(Generator.java:456)
>> > at org.apache.nutch.crawl.Crawl.run(Crawl.java:130)
>> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>> > at org.apache.nutch.crawl.Crawl.main(Crawl.java:55)
>> >
>>
>
>

Reply via email to