[jira] [Commented] (YARN-4362) Too many preemption activity when nodelabels are non exclusive

2015-11-17 Thread Varun Saxena (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15008774#comment-15008774
 ] 

Varun Saxena commented on YARN-4362:


bq. We should restrict assigning to partition 2.
I think it depends on how to treat non exclusive partition. 
Does it make sense to have a non exclusive partition without it being 
accessible from any queue and assign to it ? Not sure if there is a use case 
for this. If we consider it as free for all kind of partition and its alright 
to assign containers to such a partition, we need to fix the preemption logic. 
Because such assignments will be preempted despite only one app running.
However, if we consider that such a partition has no real meaning until and 
unless its accessible from a queue, you are correct that we should not assign 
to it. Although in this case we can argue that why have a node assigned to a 
partition if its not accessible from any queue. But I guess there can be some 
scenarios where such a situation is possible.

Wangda, thoughts on this ?


> Too many preemption activity when nodelabels are non exclusive
> --
>
> Key: YARN-4362
> URL: https://issues.apache.org/jira/browse/YARN-4362
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Bibin A Chundatt
>Assignee: Varun Saxena
>Priority: Critical
> Attachments: Preemptedpartition.log, ProportionalDefaultQueue.log, 
> ProportionalPolicy.log, capacity-scheduler.xml
>
>
> Steps to reproduce
> ===
> 1.Configure HA cluster with 6 nodes and 3 partition(1,2,3) all non exclusive
> *Partition configuration is as follows*
> 1,2 NM's mapped with Label 1
> NM 3  to  label 2
> 4,5 NM's mapped to Label 3
> NM 6 in DEFAULT partition
> In capacity scheduler the queue are linked only to 1,3 partition.
> The NM 3 with label 2 is a backup node for any partition whenever required 
> will change the label.
> Submit and application/job with 200 containers to default queue.
> All containers that gets assigned to partition 2 gets preempted 
> The application/map task execution is taking more time since 30-40 task gets 
> assigned to partition 2 then gets preempted and all of them needs to be 
> relaunched.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-4362) Too many preemption activity when nodelabels are non exclusive

2015-11-17 Thread Sunil G (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15008812#comment-15008812
 ] 

Sunil G commented on YARN-4362:
---

Ideally at first thought we should fix at preemption logic as more preemption 
is happening there. But in this specific case, some non-used partition has 
resources in it, and preemption logic will always  mark this as over-used 
resources.
I am not very much inclined in having a new non-exclusive partition which makes 
allocation more complex (we have no_label already for general cases). However I 
feel we can come with a default allocation limit/unit per partition even though 
it is not used/configured.
 
Hence in cases like this, guaranteed resources wont be Zero. But when real 
allocation starts happening to that label, this default allocation must be 
removed/shrinked. With this we can get some flexibility in using these free 
nodes w/o violating node label strict labeling concept (only no_label can use 
this limit).

If we do not allocate to these labels, it will be more stricter and these 
resource will be non-usable. This is also acceptable but comes with the cost of 
some non-usable nodes.

> Too many preemption activity when nodelabels are non exclusive
> --
>
> Key: YARN-4362
> URL: https://issues.apache.org/jira/browse/YARN-4362
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Bibin A Chundatt
>Assignee: Varun Saxena
>Priority: Critical
> Attachments: Preemptedpartition.log, ProportionalDefaultQueue.log, 
> ProportionalPolicy.log, capacity-scheduler.xml
>
>
> Steps to reproduce
> ===
> 1.Configure HA cluster with 6 nodes and 3 partition(1,2,3) all non exclusive
> *Partition configuration is as follows*
> 1,2 NM's mapped with Label 1
> NM 3  to  label 2
> 4,5 NM's mapped to Label 3
> NM 6 in DEFAULT partition
> In capacity scheduler the queue are linked only to 1,3 partition.
> The NM 3 with label 2 is a backup node for any partition whenever required 
> will change the label.
> Submit and application/job with 200 containers to default queue.
> All containers that gets assigned to partition 2 gets preempted 
> The application/map task execution is taking more time since 30-40 task gets 
> assigned to partition 2 then gets preempted and all of them needs to be 
> relaunched.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-4362) Too many preemption activity when nodelabels are non exclusive

2015-11-16 Thread Bibin A Chundatt (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15008072#comment-15008072
 ] 

Bibin A Chundatt commented on YARN-4362:


Attached logs and xml.

Looks like the guaranteed resource for partition 2 for queue default will be 
always zero. So any container assigned to partition 2 will get preempted from 
ProportionalCapacityPreemptionPolicy even when no other application is running.

We should restrict assigning to partition 2.
Thoughts?

> Too many preemption activity when nodelabels are non exclusive
> --
>
> Key: YARN-4362
> URL: https://issues.apache.org/jira/browse/YARN-4362
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Bibin A Chundatt
>Assignee: Varun Saxena
>Priority: Critical
> Attachments: Preemptedpartition.log, ProportionalDefaultQueue.log, 
> ProportionalPolicy.log, capacity-scheduler.xml
>
>
> Steps to reproduce
> ===
> 1.Configure HA cluster with 6 nodes and 3 partition(1,2,3) all non exclusive
> *Partition configuration is as follows*
> 1,2 NM's mapped with Label 1
> NM 3  to  label 2
> 4,5 NM's mapped to Label 3
> NM 6 in DEFAULT partition
> In capacity scheduler the queue are linked only to 1,3 partition.
> The NM 3 with label 2 is a backup node for any partition whenever required 
> will change the label.
> Submit and application/job with 200 containers to default queue.
> All containers that gets assigned to partition 2 gets preempted 
> The application/map task execution is taking more time since 30-40 task gets 
> assigned to partition 2 then gets preempted and all of them needs to be 
> relaunched.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)