[
https://issues.apache.org/jira/browse/MAPREDUCE-4030?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13251506#comment-13251506
]
Devaraj K commented on MAPREDUCE-4030:
--------------------------------------
Hi Nishan, Before the reducer bailing out, it will check the health and
progress using these below conditions
1. failureCounts.size() >= maxFailedUniqueFetches or failureCounts.size() ==
(totalMaps - doneMaps)
2. reducerHealthy ((totalFailures / (totalFailures + doneMaps)) < 0.5f)
3. reducerProgressed is not enough Enough((doneMaps / totalMaps) < 0.5f ) or
reducerStalled
I think it has run all or most of the maps in the down node and satisfied the
above conditions and then reducer failed without re-running the maps in your
case.
Can you please check and confirm whether the same case happened in your env?
> If the nodemanager on which the maptask is executed is going down before the
> mapoutput is consumed by the reducer,then the job is failing with shuffle
> error
> ------------------------------------------------------------------------------------------------------------------------------------------------------------
>
> Key: MAPREDUCE-4030
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4030
> Project: Hadoop Map/Reduce
> Issue Type: Bug
> Components: mrv2
> Reporter: Nishan Shetty
> Assignee: Devaraj K
>
> My cluster has 2 NM's.
> The value of "mapreduce.job.reduce.slowstart.completedmaps" is set to 1.
> When the job execution is in progress and Mappers has finished about 99%
> completion,one of the NM has gone down.
> The job has failed with the following trace
> "Error: org.apache.hadoop.mapreduce.task.reduce.Shuffle$ShuffleError: error
> in shuffle in fetcher#1 at
> org.apache.hadoop.mapreduce.task.reduce.Shuffle.run(Shuffle.java:123) at
> org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:371) at
> org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:148) at
> java.security.AccessController.doPrivileged(Native Method) at
> javax.security.auth.Subject.doAs(Subject.java:396) at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:143) Caused by:
> java.io.IOException: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out. at
> org.apache.hadoop.mapreduce.task.reduce.ShuffleScheduler.checkReducerHealth(ShuffleScheduler.java:253)
> at
> org.apache.hadoop.mapreduce.task.reduce.ShuffleScheduler.copyFailed(ShuffleScheduler.java:187)
> at
> org.apache.hadoop.mapreduce.task.reduce.Fetcher.copyFromHost(Fetcher.java:240)
> at org.apache.hadoop.mapreduce.task.reduce.Fetcher.run(Fetcher.java:152) "
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators:
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira