[ 
https://issues.apache.org/jira/browse/YARN-7681?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16315524#comment-16315524
 ] 

Arun Suresh commented on YARN-7681:
-----------------------------------

[~cheersyang], When the processor is turned on, the SchedulingRequest does not 
flow through the {{AppPlacementAllocator}}. The code paths are somewhat 
separate.
So, if the processor is turned on, we need to perform the check in the 
[attemptAllocationOnNode|https://github.com/apache/hadoop/blob/YARN-6592/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacityScheduler.java#L2504-L2519]
 method.

> Scheduler should double-check placement constraint before actual allocation 
> is made
> -----------------------------------------------------------------------------------
>
>                 Key: YARN-7681
>                 URL: https://issues.apache.org/jira/browse/YARN-7681
>             Project: Hadoop YARN
>          Issue Type: Sub-task
>          Components: RM, scheduler
>            Reporter: Weiwei Yang
>            Assignee: Weiwei Yang
>
> This JIRA is created based on the discussions under YARN-7612, see comments 
> after [this 
> comment|https://issues.apache.org/jira/browse/YARN-7612?focusedCommentId=16303051&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16303051].
>  AllocationTagsManager maintains tags info that helps to make placement 
> decisions at placement phase, however tags are changing along with 
> container's lifecycle, so it is possible that the placement violates the 
> constraints at the scheduling phase. Propose to add an extra check in the 
> scheduler to make sure constraints are still satisfied during the actual 
> allocation.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to