[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2324?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13065505#comment-13065505
 ] 

Robert Joseph Evans commented on MAPREDUCE-2324:
------------------------------------------------

I don't believe that the fix I submitted is incomplete the issue is that MRv2 
does things so very differently we need to tackle the problem in a different 
way.  I am sure the patch is not perfect and I am very happy to see any better 
ideas/patches.  Also I am getting noise from my customers about this so I would 
like to see a fix in a sustaining release.  It is not a lot of noise but I do 
have to at least try to get a fix in.

I do agree that having different configuration values is an issue that I would 
like to avoid, but currently 0.23 has dropped mapreduce.reduce.input.limit all 
together along with who knows what other configuration values.  I do not see 
any way to maintain mapreduce.reduce.input.limit in MRv2.

I have started looking at the scheduler code in yarn and this is just 
preliminary but it looks like what we want to do is to extend Resource to 
include disk space not just RAM.  The NodeManager can then also report back the 
amount of disk space that it has free, just like the TaskTracker does.  Then 
for Reduce Tasks we teh MR Application Master can request the container based 
off of the estimated reduce input size. We can also put in a more generic 
resource starvation detection mechanism that would work for both RAM and Disk.

> Job should fail if a reduce task can't be scheduled anywhere
> ------------------------------------------------------------
>
>                 Key: MAPREDUCE-2324
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-2324
>             Project: Hadoop Map/Reduce
>          Issue Type: Bug
>    Affects Versions: 0.20.2, 0.20.205.0
>            Reporter: Todd Lipcon
>            Assignee: Robert Joseph Evans
>         Attachments: MR-2324-security-v1.txt
>
>
> If there's a reduce task that needs more disk space than is available on any 
> mapred.local.dir in the cluster, that task will stay pending forever. For 
> example, we produced this in a QA cluster by accidentally running terasort 
> with one reducer - since no mapred.local.dir had 1T free, the job remained in 
> pending state for several days. The reason for the "stuck" task wasn't clear 
> from a user perspective until we looked at the JT logs.
> Probably better to just fail the job if a reduce task goes through all TTs 
> and finds that there isn't enough space.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to