Ok, the address i used was a proxy, now i get 2012-01-27 15:26:40,165 INFO solr.SolrWriter - Adding 1 documents 2012-01-27 15:26:40,292 WARN mapred.LocalJobRunner - job_local_0009 org.apache.solr.common.SolrException: Not Found
Not Found request: http://127.0.0.1:8080/solr_3-5/searchdkdde_en/update?wt=javabin&version=2 at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:430) at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:244) at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105) at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:49) at org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:93) at org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:48) at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:474) at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:411) at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:216) 2012-01-27 15:26:40,543 ERROR solr.SolrIndexer - java.io.IOException: Job failed! if i use the 'crawl' command. If i used the 'solrindex' command on the data previously fetched with 'crawl' - it does get indexed thought. Why is this so? -- [Entwickler] dkd Internet Service GmbH development // kommunikation // design Kaiserstraße 73 60329 Frankfurt/Main fon: +49 69 2475218-0 fax: +49 69 2475218-99 e-mail: [email protected] twitter: http://twitter.com/dkd_de facebook: http://www.facebook.com/www.dkd.de web: http://www.dkd.de Registergericht: Amtsgericht Frankfurt am Main Registernummer: HRB 45590 Geschäftsführer: Olivier Dobberkau, Søren Schaffstein, Götz Wegenast, Christian Zabanski Aktuelle Projekte: http://www.spielwarenmesse-eg.de – Relaunch & Responsive Design (TYPO3) http://www.horsch.com – Relaunch Website (TYPO3) http://www.dosb.de – Refresh Website (TYPO3) Am 26.01.2012 um 14:28 schrieb Markus Jelsma: > Seems like a connection issue. Does it fail all the time? Or you going > through > a proxy? Keep in mind that SolrWriter sends data via a chunked encoding, some > proxies may not deal with it. > > > > On Thursday 26 January 2012 14:11:14 Denis Sinner wrote: >> Hello, >> i tested Nutch setting up a local Solr server, working fine. >> Now i try to index into Solr running on a different server, i get the >> following error (hadoop.log) >> >> 2012-01-26 13:57:30,957 INFO solr.SolrWriter - Adding 1 documents >> 2012-01-26 13:57:31,057 INFO httpclient.HttpMethodDirector - I/O exception >> (java.net.SocketException) caught when processing request: Connection >> reset 2012-01-26 13:57:31,057 INFO httpclient.HttpMethodDirector - >> Retrying request 2012-01-26 13:57:31,059 INFO solr.SolrWriter - Adding 1 >> documents 2012-01-26 13:57:31,064 INFO httpclient.HttpMethodDirector - >> I/O exception (java.net.SocketException) caught when processing request: >> Connection reset 2012-01-26 13:57:31,064 INFO >> httpclient.HttpMethodDirector - Retrying request 2012-01-26 13:57:31,069 >> WARN mapred.LocalJobRunner - job_local_0009 java.io.IOException >> at >> org.apache.nutch.indexer.solr.SolrWriter.makeIOException(SolrWriter.java:1 >> 03) at org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:98) >> at >> org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.j >> ava:48) at >> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:474) at >> org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:411) at >> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:216) >> Caused by: org.apache.solr.client.solrj.SolrServerException: >> org.apache.commons.httpclient.ProtocolException: Unbuffered entity >> enclosing request can not be repeated. at >> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHtt >> pSolrServer.java:477) at >> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHtt >> pSolrServer.java:246) at >> org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(Abstrac >> tUpdateRequest.java:105) at >> org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:69) at >> org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:54) at >> org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:93) ... 4 >> more >> Caused by: org.apache.commons.httpclient.ProtocolException: Unbuffered >> entity enclosing request can not be repeated. at >> org.apache.commons.httpclient.methods.EntityEnclosingMethod.writeRequestBo >> dy(EntityEnclosingMethod.java:487) at >> org.apache.commons.httpclient.HttpMethodBase.writeRequest(HttpMethodBase.j >> ava:2114) at >> org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1 >> 096) at >> org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMeth >> odDirector.java:398) at >> org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodD >> irector.java:171) at >> org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397 >> ) at >> org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323 >> ) at >> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHtt >> pSolrServer.java:424) ... 9 more >> 2012-01-26 13:57:31,613 ERROR solr.SolrIndexer - java.io.IOException: Job >> failed! 2012-01-26 13:57:31,617 INFO solr.SolrDeleteDuplicates - >> SolrDeleteDuplicates: starting at 2012-01-26 13:57:31 >> >> >> What could be the problem, any hints? >> Would be thankful for any help! >> >> Denis > > -- > Markus Jelsma - CTO - Openindex >

