I had get the same erreur but I don't know what I have to change in my 
"/etc/hosts"
thank you for your help

Le mercredi 5 mars 2014 09:39:46 UTC, Yann Barraud a écrit :
>
> Hi,
>
> Is your ES instance known by your Hadoop cluster (/etc/hosts) ? 
>
> It does not even seems to read in it.
>
> Cheers,
> Yann
>
> Le mercredi 5 mars 2014 06:32:55 UTC+1, siva mannem a écrit :
>>
>> I installed ES(at the location /usr/lib/elasticsearch/) on our gateway 
>> server and i am able to run some basic curl commands like XPUT and XGET to 
>> create some indices and retrieve the data in them.
>> i am able to give single line JSON record but i am unable to give JSON 
>> file as input to curl XPUT .
>> can anybody give me the syntax for giving JSON file as input for curl 
>> XPUT command?
>>
>> my next issue is i copied  the following 4 elasticsearch-hadoop jar files
>> elasticsearch-hadoop-1.3.0.M2.jar          
>> elasticsearch-hadoop-1.3.0.M2-sources.jar
>> elasticsearch-hadoop-1.3.0.M2-javadoc.jar  
>> elasticsearch-hadoop-1.3.0.M2-yarn.jar
>>
>> to  /usr/lib/elasticsearch/elasticsearch-0.90.9/lib
>> and /usr/lib/gphd/pig/
>>
>> i have the following json file j.json
>> ++++++
>> {"k1":"v1" ,  "k2":"v2" , "k3":"v3"}
>> ++++++++
>>
>> in my_hdfs_path.
>>
>> my pig script is write_data_to_es.pig
>> +++++++++++++
>> REGISTER /usr/lib/gphd/pig/elasticsearch-hadoop-1.3.0.M2-yarn.jar;
>> DEFINE ESTOR org.elasticsearch.hadoop.pig.EsStorage('es.resource=usa/ca');
>> A = LOAD '/my_hdfs_path/j.json' using 
>> JsonLoader('k1:chararray,k2:chararray,k3:chararray');
>> STORE A into 'usa/ca' USING ESTOR('es.input.json=true');
>> ++++++++++++++
>>
>> when i run my pig script 
>> +++++++++
>> pig -x mapreduce  write_data_to_es.pig 
>> ++++++++++++
>>
>> i am getting following error
>> +++++++++
>> Input(s):
>> Failed to read data from "/my_hdfs_path/j.json"
>>
>> Output(s):
>> Failed to produce result in "usa/ca"
>>
>> Counters:
>> Total records written : 0
>> Total bytes written : 0
>> Spillable Memory Manager spill count : 0
>> Total bags proactively spilled: 0
>> Total records proactively spilled: 0
>>
>> Job DAG:
>> job_1390436301987_0089
>>
>>
>> 2014-03-05 00:26:50,839 [main] INFO 
>>  
>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>  
>> - Failed!
>> 2014-03-05 00:26:50,841 [main] ERROR 
>> org.apache.pig.tools.grunt.GruntParser - ERROR 2997: Input(s):
>> Failed to read data from "/elastic_search/es_hadoop_test.json"
>>
>> Output(s):
>> Failed to produce result in "mannem/siva"
>>
>> Counters:
>> Total records written : 0
>> Total bytes written : 0
>> Spillable Memory Manager spill count : 0
>> Total bags proactively spilled: 0
>> Total records proactively spilled: 0
>>
>> Job DAG:
>> job_1390436301987_0089
>>
>> 2014-03-05 00:26:50,839 [main] INFO 
>>  
>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>  
>> - Failed!
>> 2014-03-05 00:26:50,841 [main] ERROR 
>> org.apache.pig.tools.grunt.GruntParser - *ERROR 2997: Encountered 
>> IOException. Out of nodes and retries; caught exception*
>> Details at logfile: 
>> /usr/lib/elasticsearch/elasticsearch-0.90.9/pig_1393997175206.log
>> ++++++++++++
>>
>> i am using pivotal hadoop version (1.0.1)  which is basically apache 
>> hadoop (hadoop-2.0.2)
>> and pig version is 0.10.1
>> and elastic search version is 0.90.9
>>
>> can anybody help me out here?
>> thank you so much in advance for your help.
>>
>>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/1dd8ff7d-ef53-4614-9300-13b5f6ed66fa%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to