Hello , I used "elasticsearch-hadoop-1.3.0.M2" and it given me Failed Jobs: JobId Alias Feature Message Outputs job_201404142111_0008 weblog_count,weblog_group,weblogs GROUP_BY,COMBINER Message: Job failed! Error - # of failed Reduce Tasks exceeded allowed limit. FailedCount: 1. LastFailedTask: task_201404142111_0008_r_000000 weblogs/logs2,
Input(s): Failed to read data from "/user/hive/warehouse/weblogs" Output(s): Failed to produce result in "weblogs/logs2" Counters: Total records written : 0 Total bytes written : 0 Spillable Memory Manager spill count : 0 Total bags proactively spilled: 0 Total records proactively spilled: 0 I think it s better to know how things works from the beginning,so pls would u like to tell what I have to do (what should I start with) what should I do to configure elasticsearch (head) with Hadoop and how can I work with elasticsearch head Thank you so much ,really all what u say is so helpful .Thank you Le lundi 14 avril 2014 09:33:12 UTC+1, Costin Leau a écrit : > > Since you are not specifying the network configuration for an > elasticsearch node, it will default to localhost:9200. This works as long > as you are running Hadoop (Pig, Hive, Cascading, etc...) on the same > machine as Elasticsearch - based on your exception that is unlikely the > case. > Try specifying the `es.nodes` parameter - see the documentation for more > information. > > Additionally, you seem to be using the wrong jar of es-hadoop - in your > script you are registering es-hadoop-1.2.0.jar (which does not support the > pig/hive/cascading functionality) while the stacktrace indicate you are > using es-hadoop-1.3.X.jar. > > Make sure you are using es-hadoop-1.3.0.M3.jar (which is released and > available in Maven Central) and no other version. I recommend starting with > the examples in the reference docs, which show to easily load and store > data to/from Elasticsearch. > Once that works, consider extending your script. > > Hope this helps, > > > On Mon, Apr 14, 2014 at 11:23 AM, hanine haninne > <[email protected]<javascript:> > > wrote: > >> Hi, >> >> Here is my log and my script Pig >> >> log file : >> Backend error message >> --------------------- >> java.io.IOException: java.io.IOException: Out of nodes and retries; >> caught exception >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.runPipeline(PigGenericMapReduce.java:469) >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.processOnePackageOutput(PigGenericMapReduce.java:432) >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.reduce(PigGenericMapReduce.java:412) >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.reduce(PigGenericMapReduce.java:256) >> at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:177) >> at >> org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:649) >> at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:418) >> at org.apache.hadoop.mapred.Child$4.run(Child.java:255) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:415) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190) >> at org.apache.hadoop.mapred.Child.main(Child.java:249) >> Caused by: java.io.IOException: Out of nodes and retries; caught exception >> at >> org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:81) >> at >> org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:221) >> at >> org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:205) >> at >> org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:209) >> at org.elasticsearch.hadoop.rest.RestClient.get(RestClient.java:103) >> at >> org.elasticsearch.hadoop.rest.RestClient.discoverNodes(RestClient.java:85) >> at >> org.elasticsearch.hadoop.rest.InitializationUtils.discoverNodesIfNeeded(InitializationUtils.java:60) >> at >> org.elasticsearch.hadoop.mr.EsOutputFormat$ESRecordWriter.init(EsOutputFormat.java:165) >> at >> org.elasticsearch.hadoop.mr.EsOutputFormat$ESRecordWriter.write(EsOutputFormat.java:147) >> at org.elasticsearch.hadoop.pig.EsStorage.putNext(EsStorage.java:188) >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat$PigRecordWriter.write(PigOutputFormat.java:139) >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat$PigRecordWriter.write(PigOutputFormat.java:98) >> at >> org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.write(ReduceTask.java:586) >> at >> org.apache.hadoop.mapreduce.TaskInputOutputContext.write(TaskInputOutputContext.java:80) >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.runPipeline(PigGenericMapReduce.java:467) >> ... 11 more >> Caused by: java.net.ConnectException: Connection refused >> at java.net.PlainSocketImpl.socketConnect(Native Method) >> at >> java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339) >> at >> java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200) >> at >> java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182) >> at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) >> at java.net.Socket.connect(Socket.java:579) >> at java.net.Socket.connect(Socket.java:528) >> at java.net.Socket.<init>(Socket.java:425) >> at java.net.Socket.<init>(Socket.java:280) >> at >> org.apache.commons.httpclient.protocol.DefaultProtocolSocketFactory.createSocket(DefaultProtocolSocketFactory.java:79) >> at >> org.apache.commons.httpclient.protocol.DefaultProtocolSocketFactory.createSocket(DefaultProtocolSocketFactory.java:121) >> at >> org.apache.commons.httpclient.HttpConnection.open(HttpConnection.java:706) >> at >> org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:386) >> at >> org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:170) >> at >> org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:396) >> at >> org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:324) >> at >> org.elasticsearch.hadoop.rest.commonshttp.CommonsHttpTransport.execute(CommonsHttpTransport.java:160) >> at >> org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:74) >> ... 25 more >> >> Pig script: >> REGISTER /home/hduser/hadoop/lib/elasticsearch-hadoop-1.2.0.jar; >> weblogs = LOAD '/user/hive/warehouse/weblogs' USING PigStorage('\t') >> AS (client_ip : chararray, >> full_request_date : chararray, >> day : int, >> month : chararray, >> month_num : int, >> year : int, >> hour : int, >> minute : int, >> second : int, >> timezone : chararray, >> http_verb : chararray, >> uri : chararray, >> http_status_code : chararray, >> bytes_returned : chararray, >> referrer : chararray, >> user_agent : chararray >> ); >> weblog_group = GROUP weblogs by (client_ip, year, month_num); >> weblog_count = FOREACH weblog_group GENERATE group.client_ip, group.year, >> group.month_num, COUNT_STAR(weblogs) as pageviews; >> >> STORE weblog_count INTO 'weblogs2/logs2' USING >> org.elasticsearch.hadoop.pig.EsStorage(); >> >> And what ever I put in the LOAD it gives me the same result,even if I put >> the path of me desktop >> >> Thx >> >> Le lundi 14 avril 2014 03:11:23 UTC+1, Costin Leau a écrit : >>> >>> Hi, >>> >>> That isn't a lot of information so it's hard to figure out what's >>> actually work - one can only guess. Can you post your >>> stacktrace/logs and your script pig somewhere - like a gist? >>> >>> One thing that stands out is that you mention you are using Pig yet your >>> path points to a Hive warehouse: >>> > /Failed to read data from "/user/hive/warehouse/books"/ >>> >>> I can infer from this that maybe, the issue, is the fact that you are >>> trying to read a Hive internal file, which Pig >>> can't understand, leading to the error that you see. >>> >>> Cheers, >>> >>> >>> On 4/14/14 1:23 AM, hanine haninne wrote: >>> > Hello , >>> > >>> > I m trying to store data in ES (head) using pig script and it gives me >>> > >>> > /Input(s):/ >>> > /Failed to read data from "/user/hive/warehouse/books"/ >>> > >>> > /Output(s):/ >>> > /Failed to produce result in "books/book"/ >>> > >>> > I ll be so thankful if someone would like help me >>> > >>> > -- >>> > You received this message because you are subscribed to the Google >>> Groups "elasticsearch" group. >>> > To unsubscribe from this group and stop receiving emails from it, send >>> an email to >>> > [email protected] <mailto:elasticsearch+ >>> [email protected]>. >>> > To view this discussion on the web visit >>> > https://groups.google.com/d/msgid/elasticsearch/979f5688- >>> bd53-4b76-a97a-5b0359c8be75%40googlegroups.com >>> > <https://groups.google.com/d/msgid/elasticsearch/979f5688- >>> bd53-4b76-a97a-5b0359c8be75%40googlegroups.com?utm_medium= >>> email&utm_source=footer>. >>> > For more options, visit https://groups.google.com/d/optout. >>> >>> -- >>> Costin >>> >> -- >> You received this message because you are subscribed to the Google Groups >> "elasticsearch" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to [email protected] <javascript:>. >> To view this discussion on the web visit >> https://groups.google.com/d/msgid/elasticsearch/e766ac9d-788e-4614-80ea-04960fdc257f%40googlegroups.com<https://groups.google.com/d/msgid/elasticsearch/e766ac9d-788e-4614-80ea-04960fdc257f%40googlegroups.com?utm_medium=email&utm_source=footer> >> . >> >> For more options, visit https://groups.google.com/d/optout. >> > > -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/72350b3d-dc2f-4f5a-b066-cfd0b2241545%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.
