This seems like a better question for the Nutch list. I see hadoop
in there, so unless you've specifically configured solr to use
the HDFS directory writer factory, this has to be coming from
someplace else. And there are map/reduce tasks in here.

BTW, it would be more helpful if you posted the URL that you
successfully queried Solr with... What is the /2 on the end for?
Do you use that when you query?

Best,
Erick


On Mon, Oct 28, 2013 at 2:37 AM, Bayu Widyasanyata
<bwidyasany...@gmail.com>wrote:

> On Mon, Oct 28, 2013 at 1:26 PM, Raymond Wiker <rwi...@gmail.com> wrote:
>
> > > request: http://localhost:8080/solr/update?wt=javabin&version=2
> >
> > I think this url is incorrect: there should be a core name between "solr"
> > and "update".
> >
>
> I changed th SolrURL on crawl script's option to:
>
> ./bin/crawl urls/seed.txt TestCrawl
> http://localhost:8080/solr/mycollection/2
>
> And the result now is "Bad Request".
> I will look for another misconfiguration things...
>
> =====
>
> org.apache.solr.common.SolrException: Bad Request
>
> Bad Request
>
> request:
> http://localhost:8080/solr/mycollection/update?wt=javabin&version=2
>         at
>
> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:430)
>         at
>
> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:244)
>         at
>
> org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105)
>         at
>
> org.apache.nutch.indexwriter.solr.SolrIndexWriter.close(SolrIndexWriter.java:155)
>         at
> org.apache.nutch.indexer.IndexWriters.close(IndexWriters.java:118)
>         at
>
> org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:44)
>         at
>
> org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.close(ReduceTask.java:467)
>         at
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:535)
>         at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:421)
>         at
> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:398)
> 2013-10-28 13:30:02,804 ERROR indexer.IndexingJob - Indexer:
> java.io.IOException: Job failed!
>         at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357)
>         at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:123)
>         at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:185)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>         at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:195)
>
>
>
> --
> wassalam,
> [bayu]
>

Reply via email to