Hi Lewis:

Thank you for the help. This is the (entire) output after I set the log4j
property to debug.
==============================================================
crawl started in: crawl
rootUrlDir = urls
threads = 10
depth = 2
solrUrl=http://localhost:8983/solr/
topN = 10
Injector: starting at 2012-04-11 16:37:20
Injector: crawlDb: crawl/crawldb
Injector: urlDir: urls
Injector: Converting injected urls to crawl db entries.
Exception in thread "main" java.io.IOException: Job failed!
    at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1252)
    at org.apache.nutch.crawl.Injector.inject(Injector.java:217)
    at org.apache.nutch.crawl.Crawl.run(Crawl.java:127)
    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
    at org.apache.nutch.crawl.Crawl.main(Crawl.java:55)
==============================================================

And btw, The "urls" directory is correct and it does contain a txt file of
a list of urls.

Regards
Andy


On 10 April 2012 22:08, Lewis John Mcgibbney <[email protected]>wrote:

> There is no more log information before the solrUrl stuff, no?
>
> try setting log4j.properties to debug in conf/ rebuild the project and see
> whats going on.
>
> On Tue, Apr 10, 2012 at 1:03 PM, Andy Xue <[email protected]> wrote:
>
> > Lewis:
> > Thanks for the reply.
> > However as far as I know, I don't have to set solrUrl unless I want to
> > index using solr.
> >
> > Correct. My fault. I just assumed that this was required.
>
> Lewis
>

Reply via email to