[ 
https://issues.apache.org/jira/browse/HADOOP-4246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12635684#action_12635684
 ] 

Jothi Padmanabhan commented on HADOOP-4246:
-------------------------------------------

Having a lower bound of 1 on map-fetch-retries might not be efficient as it 
opens the possibility for map reexecution on transient errors -- three 
different reducers reporting error when the serving task tracker had a 
transient problem could lead to re-execution of the map. We probably should try 
at least twice.

> Reduce task copy errors may not kill it eventually
> --------------------------------------------------
>
>                 Key: HADOOP-4246
>                 URL: https://issues.apache.org/jira/browse/HADOOP-4246
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: mapred
>    Affects Versions: 0.19.0
>            Reporter: Amareshwari Sriramadasu
>            Assignee: Amareshwari Sriramadasu
>            Priority: Blocker
>             Fix For: 0.19.0
>
>         Attachments: patch-4246.txt, patch-4246.txt
>
>
> maxFetchRetriesPerMap in reduce task can be zero some times (when 
> maxMapRunTime is less than 4 seconds or mapred.reduce.copy.backoff is less 
> than 4). This will not count reduce task copy errors to kill it eventually.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to