[ 
https://issues.apache.org/jira/browse/HADOOP-4523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12644220#action_12644220
 ] 

Vivek Ratan commented on HADOOP-4523:
-------------------------------------

I'm proposing a couple of improvements: 
# The TT currently monitors each task (and its descendants) to see if that 
task's memory usage goes over a per-task limit. The TT should additionally 
monitor to make sure that the sum of memory used by all tasks should not go 
over a per-node limit (_tasks.maxmemory_). This situation is unlikely t happen 
if schedulers consider memory judiciously when scheduling, but not all 
schedulers may. In addition, the TT should pick tasks that ran last, when 
deciding what tasks to kill. 
# I'm also proposing a way to specify _tasks.maxmemory_. This particular 
discussion is, however, going on in HADOOP-4035, so we can continue discussion 
there. 


> Enhance how memory-intensive user tasks are handled
> ---------------------------------------------------
>
>                 Key: HADOOP-4523
>                 URL: https://issues.apache.org/jira/browse/HADOOP-4523
>             Project: Hadoop Core
>          Issue Type: Improvement
>          Components: mapred
>    Affects Versions: 0.19.0
>            Reporter: Vivek Ratan
>            Assignee: Vinod K V
>
> HADOOP-3581 monitors each Hadoop task to see if its memory usage (which 
> includes usage of any tasks spawned by it and so on) is within a per-task 
> limit. If the task's memory usage goes over its limit, the task is killed. 
> This, by itself, is not enough to prevent badly behaving jobs from bringing 
> down nodes. What is also needed is the ability to make sure that the sum 
> total of VM usage of all Hadoop tasks does not exceed a certain limit.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to