[ https://issues.apache.org/jira/browse/HADOOP-4035?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12644660#action_12644660 ]
Hemanth Yamijala commented on HADOOP-4035: ------------------------------------------ bq. Should the memory-related config values be expressed in MB or GB or KB or just bytes? MB sounds good to me. The other parameter we have in hadoop related to memory is mapred.child.ulimit which is specified in KB. I think expressing these values in KB would keep things consistent. bq. If a job's specified VM or RAM task limit is higher than the max limit, that job shouldn't be allowed to run. Should the JT reject the job when it is submitted, or should the scheduler do it, by failing the job? I think scheduler failing the job is more consistent if the scheduling decisions are being made in the scheduler. bq. Should the Capacity Scheduler use the entire RAM of a TT when making a scheduling decision, or an offset? I am not really sure either way. Given earlier discussions we've had that virtual memory is what really matters, I am guessing we don't need it. Regarding the config variable names, a few concerns/suggestions: - mapred.tasktracker.virtualmemory.reserved: This seems like specifying the amount of memory reserved for Hadoop, whereas it means the opposite. Can we call it mapred.tasktracker.vmem.excluded ? - We are using 'virtualmemory', and 'vm' to represent virtual memory. Should we consistently name it as 'vmem' everywhere ? - Similar to excluded, rename variables to mapred.task.maxvmem.default and mapred.task.maxvmem.limit ? Does this make sense ? > Modify the capacity scheduler (HADOOP-3445) to schedule tasks based on memory > requirements and task trackers free memory > ------------------------------------------------------------------------------------------------------------------------ > > Key: HADOOP-4035 > URL: https://issues.apache.org/jira/browse/HADOOP-4035 > Project: Hadoop Core > Issue Type: Bug > Components: contrib/capacity-sched > Affects Versions: 0.19.0 > Reporter: Hemanth Yamijala > Assignee: Vinod K V > Priority: Blocker > Fix For: 0.20.0 > > Attachments: 4035.1.patch, HADOOP-4035-20080918.1.txt, > HADOOP-4035-20081006.1.txt, HADOOP-4035-20081006.txt, HADOOP-4035-20081008.txt > > > HADOOP-3759 introduced configuration variables that can be used to specify > memory requirements for jobs, and also modified the tasktrackers to report > their free memory. The capacity scheduler in HADOOP-3445 should schedule > tasks based on these parameters. A task that is scheduled on a TT that uses > more than the default amount of memory per slot can be viewed as effectively > using more than one slot, as it would decrease the amount of free memory on > the TT by more than the default amount while it runs. The scheduler should > make the used capacity account for this additional usage while enforcing > limits, etc. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.