And that it is caused by this exception, as I've found out

2013-12-19 13:14:28,237 ERROR [pipe-uplink-handler]
org.apache.hadoop.mapred.pipes.BinaryProtocol: java.io.EOFException
at java.io.DataInputStream.readByte(DataInputStream.java:267)
at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:308)
at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:329)
at
org.apache.hadoop.mapred.pipes.BinaryProtocol$UplinkReaderThread.run(BinaryProtocol.java:125)


On 18 December 2013 10:52, Silvina Caíno Lores <silvi.ca...@gmail.com>wrote:

> I forgot to mention that the wordcount pipes example runs successfully.
>
>
> On 18 December 2013 10:50, Silvina Caíno Lores <silvi.ca...@gmail.com>wrote:
>
>> Hi everyone,
>>
>> I'm working with a single node cluster and a Hadoop Pipes job that throws
>> the following exception on execution:
>>
>> 13/12/18 10:44:55 INFO mapreduce.Job: Running job: job_1387359324416_0002
>> 13/12/18 10:45:03 INFO mapreduce.Job: Job job_1387359324416_0002 running
>> in uber mode : false
>> 13/12/18 10:45:03 INFO mapreduce.Job: map 0% reduce 0%
>> 13/12/18 10:45:08 INFO mapreduce.Job: Task Id :
>> attempt_1387359324416_0002_m_000000_0, Status : FAILED
>> Error: java.io.IOException
>> at
>> org.apache.hadoop.mapred.pipes.OutputHandler.waitForAuthentication(OutputHandler.java:186)
>> at
>> org.apache.hadoop.mapred.pipes.Application.waitForAuthentication(Application.java:195)
>> at org.apache.hadoop.mapred.pipes.Application.<init>(Application.java:150)
>> at
>> org.apache.hadoop.mapred.pipes.PipesMapRunner.run(PipesMapRunner.java:69)
>> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430)
>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:342)
>> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:171)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at javax.security.auth.Subject.doAs(Subject.java:415)
>> at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1515)
>> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:166)
>>
>> /// same exception several times ///
>>
>> 13/12/18 10:45:25 INFO mapreduce.Job: map 93% reduce 0%
>> 13/12/18 10:45:26 INFO mapreduce.Job: map 100% reduce 100%
>> 13/12/18 10:45:26 INFO mapreduce.Job: Job job_1387359324416_0002 failed
>> with state FAILED due to: Task failed task_1387359324416_0002_m_000000
>> Job failed as tasks failed. failedMaps:1 failedReduces:0
>>
>>
>> I don't really get why the job is at 100% if it actually failed by the
>> way.
>>
>> Any ideas? Thanks in advance!
>>
>> Best,
>> Silvina
>>
>>
>>
>>
>>
>>
>

Reply via email to