I had get the same erreur but I don't know what I have to change in my
"/etc/hosts"
thank you for your help
Le mercredi 5 mars 2014 09:39:46 UTC, Yann Barraud a écrit :
>
> Hi,
>
> Is your ES instance known by your Hadoop cluster (/etc/hosts) ?
>
> It does not even seems to read in it.
>
> Cheers,
> Yann
>
> Le mercredi 5 mars 2014 06:32:55 UTC+1, siva mannem a écrit :
>>
>> I installed ES(at the location /usr/lib/elasticsearch/) on our gateway
>> server and i am able to run some basic curl commands like XPUT and XGET to
>> create some indices and retrieve the data in them.
>> i am able to give single line JSON record but i am unable to give JSON
>> file as input to curl XPUT .
>> can anybody give me the syntax for giving JSON file as input for curl
>> XPUT command?
>>
>> my next issue is i copied the following 4 elasticsearch-hadoop jar files
>> elasticsearch-hadoop-1.3.0.M2.jar
>> elasticsearch-hadoop-1.3.0.M2-sources.jar
>> elasticsearch-hadoop-1.3.0.M2-javadoc.jar
>> elasticsearch-hadoop-1.3.0.M2-yarn.jar
>>
>> to /usr/lib/elasticsearch/elasticsearch-0.90.9/lib
>> and /usr/lib/gphd/pig/
>>
>> i have the following json file j.json
>> ++++++
>> {"k1":"v1" , "k2":"v2" , "k3":"v3"}
>> ++++++++
>>
>> in my_hdfs_path.
>>
>> my pig script is write_data_to_es.pig
>> +++++++++++++
>> REGISTER /usr/lib/gphd/pig/elasticsearch-hadoop-1.3.0.M2-yarn.jar;
>> DEFINE ESTOR org.elasticsearch.hadoop.pig.EsStorage('es.resource=usa/ca');
>> A = LOAD '/my_hdfs_path/j.json' using
>> JsonLoader('k1:chararray,k2:chararray,k3:chararray');
>> STORE A into 'usa/ca' USING ESTOR('es.input.json=true');
>> ++++++++++++++
>>
>> when i run my pig script
>> +++++++++
>> pig -x mapreduce write_data_to_es.pig
>> ++++++++++++
>>
>> i am getting following error
>> +++++++++
>> Input(s):
>> Failed to read data from "/my_hdfs_path/j.json"
>>
>> Output(s):
>> Failed to produce result in "usa/ca"
>>
>> Counters:
>> Total records written : 0
>> Total bytes written : 0
>> Spillable Memory Manager spill count : 0
>> Total bags proactively spilled: 0
>> Total records proactively spilled: 0
>>
>> Job DAG:
>> job_1390436301987_0089
>>
>>
>> 2014-03-05 00:26:50,839 [main] INFO
>>
>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>
>> - Failed!
>> 2014-03-05 00:26:50,841 [main] ERROR
>> org.apache.pig.tools.grunt.GruntParser - ERROR 2997: Input(s):
>> Failed to read data from "/elastic_search/es_hadoop_test.json"
>>
>> Output(s):
>> Failed to produce result in "mannem/siva"
>>
>> Counters:
>> Total records written : 0
>> Total bytes written : 0
>> Spillable Memory Manager spill count : 0
>> Total bags proactively spilled: 0
>> Total records proactively spilled: 0
>>
>> Job DAG:
>> job_1390436301987_0089
>>
>> 2014-03-05 00:26:50,839 [main] INFO
>>
>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>
>> - Failed!
>> 2014-03-05 00:26:50,841 [main] ERROR
>> org.apache.pig.tools.grunt.GruntParser - *ERROR 2997: Encountered
>> IOException. Out of nodes and retries; caught exception*
>> Details at logfile:
>> /usr/lib/elasticsearch/elasticsearch-0.90.9/pig_1393997175206.log
>> ++++++++++++
>>
>> i am using pivotal hadoop version (1.0.1) which is basically apache
>> hadoop (hadoop-2.0.2)
>> and pig version is 0.10.1
>> and elastic search version is 0.90.9
>>
>> can anybody help me out here?
>> thank you so much in advance for your help.
>>
>>
--
You received this message because you are subscribed to the Google Groups
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/elasticsearch/1dd8ff7d-ef53-4614-9300-13b5f6ed66fa%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.