[ 
https://issues.apache.org/jira/browse/YARN-193?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13620651#comment-13620651
 ] 

Zhijie Shen commented on YARN-193:
----------------------------------

{quote}
Default value of max-vcores of 32 might be too high.
{quote}

Why 32 is originally used?

In http://hortonworks.com/blog/apache-hadoop-yarn-background-and-an-overview/, 
it is said:

2012 – 16+ cores, 48-96GB of RAM, 12x2TB or 12x3TB of disk.

How about we choosing 16?

{quote}
Why is conf being set 2 times for each value? Same for vcores.
{quote}

I'll fix the bug.
                
> Scheduler.normalizeRequest does not account for allocation requests that 
> exceed maximumAllocation limits 
> ---------------------------------------------------------------------------------------------------------
>
>                 Key: YARN-193
>                 URL: https://issues.apache.org/jira/browse/YARN-193
>             Project: Hadoop YARN
>          Issue Type: Bug
>          Components: resourcemanager
>    Affects Versions: 2.0.2-alpha, 3.0.0
>            Reporter: Hitesh Shah
>            Assignee: Zhijie Shen
>         Attachments: MR-3796.1.patch, MR-3796.2.patch, MR-3796.3.patch, 
> MR-3796.wip.patch, YARN-193.10.patch, YARN-193.11.patch, YARN-193.12.patch, 
> YARN-193.4.patch, YARN-193.5.patch, YARN-193.6.patch, YARN-193.7.patch, 
> YARN-193.8.patch, YARN-193.9.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to