[
https://issues.apache.org/jira/browse/YARN-4678?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15211403#comment-15211403
]
Brahma Reddy Battula commented on YARN-4678:
--------------------------------------------
bq.1) Understand why reserved resource + allocated resource could excess
queue's max capacity, maybe we can add a test to make sure it won't happen
Total= 9216 MB ( 9GB) where each NM is having 3GB ( total 3 NM's)
Configure a queueA with 10% and submit pi job with 2GB map memory and total-5
Containers request. Please check the following for same.
||Used||Usedcapacity||absoluteUsedCapacity||NM-Remaining||
|512 MB ( AM Conatiner Request)|0.55 ( 512 MB/ 921 MB)|0.055 ( 512MB /9216
MB)|{color:blue}NodeA{color}-2560MB|
|2560 MB ( Container allocation with
2GB)|2.7795875|0.2777778|{color:green}NodeB{color}-1024MB|
|4608 MB ( Container Reservation with
2GB)|5.0032573|0.5|{color:green}NodeB{color}-1024MB,{color:red}2048MB
Reserved{color}|
|6656 MB ( Container allocation with 2GB)|7.226927|0.7222222|NodeC-1024MB|
|8704 MB ( Container Reservation with 2GB)|
9.450597|0.9444444|NodeC-1024MB,{color:red}2048MB Reserved{color}||
|10752 MB ( Container allocation with 2GB)|
11.674267|1.1674267|{color:blue}NodeA{color}-512MB|
bq. 2) If we simply deduct reserved resources from used and show on the UI,
user could find cluster utilization is < 100 in most of the time, and it gonna
be hard to explain the reason of why it cannot reach 100%. The ideal solution
is that we can show reserved and allocated resources on the same bar with
different color.
I think, will it make confusion ..? as sunil mentioned we are giving metrics
for both resereved and used.. I feel, subracting from used should be ok.
> Cluster used capacity is > 100 when container reserved
> -------------------------------------------------------
>
> Key: YARN-4678
> URL: https://issues.apache.org/jira/browse/YARN-4678
> Project: Hadoop YARN
> Issue Type: Bug
> Reporter: Brahma Reddy Battula
> Assignee: Sunil G
> Attachments: 0001-YARN-4678.patch, 0002-YARN-4678.patch,
> 0003-YARN-4678.patch, reservedCapInClusterMetrics.png
>
>
> *Scenario:*
> * Start cluster with Three NM's each having 8GB (cluster memory:24GB).
> * Configure queues with elasticity and userlimitfactor=10.
> * disable pre-emption.
> * run two job with different priority in different queue at the same time
> ** yarn jar hadoop-mapreduce-examples-2.7.2.jar pi -Dyarn.app.priority=LOW
> -Dmapreduce.job.queuename=QueueA -Dmapreduce.map.memory.mb=4096
> -Dyarn.app.mapreduce.am.resource.mb=1536
> -Dmapreduce.job.reduce.slowstart.completedmaps=1.0 10 1000000000000
> ** yarn jar hadoop-mapreduce-examples-2.7.2.jar pi -Dyarn.app.priority=HIGH
> -Dmapreduce.job.queuename=QueueB -Dmapreduce.map.memory.mb=4096
> -Dyarn.app.mapreduce.am.resource.mb=1536 3 1000000000000
> * observe the cluster capacity which was used in RM web UI
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)