[ 
https://issues.apache.org/jira/browse/YARN-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17680793#comment-17680793
 ] 

ASF GitHub Bot commented on YARN-10965:
---------------------------------------

9uapaw commented on code in PR #3470:
URL: https://github.com/apache/hadoop/pull/3470#discussion_r1087133894


##########
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacitySchedulerQueueManager.java:
##########
@@ -100,6 +101,7 @@ public CapacitySchedulerQueueManager(Configuration conf,
     this.queueStateManager = new QueueStateManager<>();
     this.appPriorityACLManager = appPriorityACLManager;
     this.configuredNodeLabels = new ConfiguredNodeLabels();
+    this.queueCapacityHandler = new 
CapacitySchedulerQueueCapacityHandler(labelManager);

Review Comment:
   As discussed, it will be in use after YARN-11000.





> Centralize queue resource calculation based on CapacityVectors
> --------------------------------------------------------------
>
>                 Key: YARN-10965
>                 URL: https://issues.apache.org/jira/browse/YARN-10965
>             Project: Hadoop YARN
>          Issue Type: Sub-task
>          Components: capacity scheduler
>            Reporter: Andras Gyori
>            Assignee: Andras Gyori
>            Priority: Major
>              Labels: pull-request-available
>          Time Spent: 6.5h
>  Remaining Estimate: 0h
>
> With the introduction of YARN-10930 it is possible to unify queue resource 
> calculation. In order to narrow down the scope of this patch, the base system 
> is implemented here, without refactoring the existing resource calculation in 
> updateClusterResource (which will be done in YARN-11000).



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to