[ 
https://issues.apache.org/jira/browse/HADOOP-3376?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12596391#action_12596391
 ] 

Hadoop QA commented on HADOOP-3376:
-----------------------------------

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12381947/checklimits.sh
  against trunk revision 655674.

    +1 @author.  The patch does not contain any @author tags.

    -1 tests included.  The patch doesn't appear to include any new or modified 
tests.
                        Please justify why no tests are needed for this patch.

    -1 patch.  The patch command could not apply the patch.

Console output: 
http://hudson.zones.apache.org/hudson/job/Hadoop-Patch/2457/console

This message is automatically generated.

> [HOD] HOD should have a way to detect and deal with clusters that 
> violate/exceed resource manager limits
> --------------------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-3376
>                 URL: https://issues.apache.org/jira/browse/HADOOP-3376
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: contrib/hod
>            Reporter: Vinod Kumar Vavilapalli
>            Assignee: Vinod Kumar Vavilapalli
>         Attachments: checklimits.sh, HADOOP-3376
>
>
> Currently If we set up resource manager/scheduler limits on the jobs 
> submitted, any HOD cluster that exceeds/violates these limits may 1) get 
> blocked/queued indefinitely or 2) blocked till resources occupied by old 
> clusters get freed. HOD should detect these scenarios and deal intelligently, 
> instead of just waiting for a long time/ for ever. This means more and proper 
> information to the submitter.
> (Internal) Use Case:
>      If there are no resource limits, users can flood the resource manager 
> queue preventing other users from using the queue. To avoid this, we could 
> have various types of limits setup in either resource manager or a scheduler 
> - max node limit in torque(per job limit), maxproc limit in maui (per 
> user/class), maxjob limit in maui(per user/class) etc. But there is one 
> problem with the current setup - for e.g if we set up maxproc limit in maui 
> to limit the aggregate number of nodes by any user over all jobs, 1) jobs get 
> queued indefinitely if jobs exceed max limit and 2) blocked if it asks for 
> nodes < max limit, but some of the resources are already used by jobs from 
> the same user. This issue addresses how to deal with scenarios like these.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to