Hi Jens I have been able to reproduce the 'No node available' issue but did not have the one with names.txt Can you please open an issue on JIRA?
Is your cluster somewhere on the cloud or in house? Thanks Julien On 29 May 2014 10:07, Jens Jahnke <[email protected]> wrote: > Hi Julien, > > On Wed, 28 May 2014 15:32:28 +0100 > Julien Nioche <[email protected]> wrote: > > JN> Ok, so you are running it on Hadoop 2 then. > > Sorry, I forgot to mention that. > > JN> [...] > JN> That file names.txt lives in the elasticsearch jar. The explanation > that > JN> comes to mind is that using Nutch in deployed mode means that this > file is > JN> not loaded. I have used ES in distributed mode with Hadoop 1.x so it > could > JN> be that there is a difference in Hadoop 2 (which we do not yet > officially > JN> support BTW) or a difference with my configuration which would explain > why > JN> I did not get this issue. > > I have created a file `conf/names.txt` with some lines of names in it and > executed `ant runtime` afterwards. But now I get an > NoNodeAvailableException > although elasticsearch is running. :-| > > 2014-05-29 10:13:40,376 INFO [main] mapreduce.Job > (Job.java:printTaskEvents(1424)) - Task Id : > attempt_1401350880587_0001_r_000000_0, Status : FAILED > Error: org.elasticsearch.client.transport.NoNodeAvailableException: No > node available > at > org.elasticsearch.client.transport.TransportClientNodesService.execute(TransportClientNodesService.java:219) > at > org.elasticsearch.client.transport.support.InternalTransportClient.execute(InternalTransportClient.java:106) > at > org.elasticsearch.client.support.AbstractClient.bulk(AbstractClient.java:147) > at > org.elasticsearch.client.transport.TransportClient.bulk(TransportClient.java:360) > at > org.elasticsearch.action.bulk.BulkRequestBuilder.doExecute(BulkRequestBuilder.java:165) > at > org.elasticsearch.action.ActionRequestBuilder.execute(ActionRequestBuilder.java:85) > at > org.elasticsearch.action.ActionRequestBuilder.execute(ActionRequestBuilder.java:59) > at > org.apache.nutch.indexwriter.elastic.ElasticIndexWriter.commit(ElasticIndexWriter.java:211) > at > org.apache.nutch.indexwriter.elastic.ElasticIndexWriter.close(ElasticIndexWriter.java:229) > at > org.apache.nutch.indexer.IndexWriters.close(IndexWriters.java:118) > at > org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:44) > at > org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.close(ReduceTask.java:520) > at > org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:474) > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157) > > I poked a bit in the source and added log output to verify that the correct > settings are used. That seems to be the case (localhost:9300). > > Regards, > > Jens > > -- > 29. Wonnemond 2014, 10:16 > Homepage : http://www.wegtam.com > > I don't know anything about music. In my line you don't have to. > -- Elvis Presley > -- Open Source Solutions for Text Engineering http://digitalpebble.blogspot.com/ http://www.digitalpebble.com http://twitter.com/digitalpebble

