Hi Julien,

On Wed, 28 May 2014 15:32:28 +0100
Julien Nioche <[email protected]> wrote:

JN> Ok, so you are running it on Hadoop 2 then.

Sorry, I forgot to mention that.

JN> [...]
JN> That file names.txt lives in the elasticsearch jar. The explanation that
JN> comes to mind is that using Nutch in deployed mode means that this file is
JN> not loaded. I have used ES in distributed mode with Hadoop 1.x so it could
JN> be that there is a difference in Hadoop 2 (which we do not yet officially
JN> support BTW) or a difference with my configuration which would explain why
JN> I did not get this issue.

I have created a file `conf/names.txt` with some lines of names in it and 
executed `ant runtime` afterwards. But now I get an NoNodeAvailableException
although elasticsearch is running. :-|

2014-05-29 10:13:40,376 INFO  [main] mapreduce.Job 
(Job.java:printTaskEvents(1424)) - Task Id : 
attempt_1401350880587_0001_r_000000_0, Status : FAILED
Error: org.elasticsearch.client.transport.NoNodeAvailableException: No node 
available
        at 
org.elasticsearch.client.transport.TransportClientNodesService.execute(TransportClientNodesService.java:219)
        at 
org.elasticsearch.client.transport.support.InternalTransportClient.execute(InternalTransportClient.java:106)
        at 
org.elasticsearch.client.support.AbstractClient.bulk(AbstractClient.java:147)
        at 
org.elasticsearch.client.transport.TransportClient.bulk(TransportClient.java:360)
        at 
org.elasticsearch.action.bulk.BulkRequestBuilder.doExecute(BulkRequestBuilder.java:165)
        at 
org.elasticsearch.action.ActionRequestBuilder.execute(ActionRequestBuilder.java:85)
        at 
org.elasticsearch.action.ActionRequestBuilder.execute(ActionRequestBuilder.java:59)
        at 
org.apache.nutch.indexwriter.elastic.ElasticIndexWriter.commit(ElasticIndexWriter.java:211)
        at 
org.apache.nutch.indexwriter.elastic.ElasticIndexWriter.close(ElasticIndexWriter.java:229)
        at org.apache.nutch.indexer.IndexWriters.close(IndexWriters.java:118)
        at 
org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:44)
        at 
org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.close(ReduceTask.java:520)
        at 
org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:474)
        at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408)
        at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
        at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157)

I poked a bit in the source and added log output to verify that the correct
settings are used. That seems to be the case (localhost:9300).

Regards,

Jens

-- 
29. Wonnemond 2014, 10:16
Homepage : http://www.wegtam.com

I don't know anything about music.  In my line you don't have to.
                -- Elvis Presley

Attachment: pgprRt75CQ3uN.pgp
Description: PGP signature

Reply via email to