Hi, I'm using hadoop-0.20.2-cdh3u2. I'm not really sure what is happening. But my Job stopped working at map --> 99,94% and reduce --> 100%. Thats kind of strange.
┏━━━━━━━━┯━━━━━━━━━━━━━┯━━━━━━━━━━━┯━━━━━━━━━┯━━━━━━━━━┯━━━━━━━━━━┯━━━━━━━━┯━━━━━━━━━━━━━━━━┓ ┃ Kind │ % Complete │ Num Tasks │ Pending │ Running │ Complete │ Killed │ Failed/Killed ┃ ┃ │ │ │ │ │ │ │ Task Attempts ┃ ┠────────┼─────────────┼───────────┼─────────┼─────────┼──────────┼────────┼────────────────┨ ┃ │ 99,94% │ │ │ │ │ │ ┃ ┃ map │ ┌────────┬┐ │ 1785 │ 0 │ 0 │ 1785 │ 0 │ 5 / 7 ┃ ┃ │ └────────┴┘ │ │ │ │ │ │ ┃ ┠────────┼─────────────┼───────────┼─────────┼─────────┼──────────┼────────┼────────────────┨ ┃ │ 100,00% │ │ │ │ │ │ ┃ ┃ reduce │ ┌─────────┐ │ 12 │ 0 │ 0 │ 12 │ 0 │ 24 / 3 ┃ ┃ │ └─────────┘ │ │ │ │ │ │ ┃ ┗━━━━━━━━┷━━━━━━━━━━━━━┷━━━━━━━━━━━┷━━━━━━━━━┷━━━━━━━━━┷━━━━━━━━━━┷━━━━━━━━┷━━━━━━━━━━━━━━━━┛ Yeah, i love w3m...but there is some relevant data in it. I have 1785 Tasks and that's the number of executed tasks is equal to this. That should be 100% am I right? There are some Failed/Killed tasks, did hadoop execute them again or is it possible that they are missing? What can I do that they will be executed again? During the process the output gives output like that: 12/02/23 12:54:41 INFO mapred.JobClient: map 99% reduce 79% 12/02/23 12:54:41 INFO mapred.JobClient: Task Id : attempt_201201101557_0519_r_000005_2, Status : FAILED java.io.IOException: Error Recovery for block blk_2222580152515928964_34420 failed because recovery from primary datanode 10.6.0.19:50010 failed 6 times. Pipeline was 10.6.0.19:50010. Aborting... at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.processDatanodeError(DFSClient.java:2833) at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$1600(DFSClient.java:2305) at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2477) attempt_201201101557_0519_r_000005_2: RedStage1: block_width=16 attempt_201201101557_0519_r_000005_2: log4j:WARN No appenders could be found for logger (org.apache.hadoop.hdfs.DFSClient). attempt_201201101557_0519_r_000005_2: log4j:WARN Please initialize the log4j system properly. 12/02/23 12:54:42 INFO mapred.JobClient: map 99% reduce 72% Or in some other cases: 12/02/23 13:22:49 INFO mapred.JobClient: map 99% reduce 89% 12/02/23 13:23:19 INFO mapred.JobClient: Task Id : attempt_201201101557_0519_r_000004_2, Status : FAILED Task attempt_201201101557_0519_r_000004_2 failed to report status for 600 seconds. Killing! 12/02/23 13:24:19 WARN mapred.JobClient: Error reading task outputRead timed out 12/02/23 13:25:19 WARN mapred.JobClient: Error reading task outputRead timed out 12/02/23 13:26:47 INFO mapred.JobClient: map 99% reduce 90% Anybody who can help? Kind regards, Mat