Am 28.10.2013 um 14:05 schrieb Reuti:
> Am 28.10.2013 um 13:59 schrieb Sangmin Park:
>
>> yes, suspending the job when all 12 slots are used on a particular host.
>> This is what I want to.
>> So, I tried to submit job using 12 slots, but it did not work.
>
> Aha, it might be necessary to change the order of rules in your RQS. The
> first matching one will allow or deny the job to be started. I.e. if all
> slots are used the (current) first rules matches and the job is rejected.
In fact, you can try just this:
limit queues !matlab.q hosts {*} to slots=$num_proc
matlab.q needs no limit, as it's already in the queue definition to have an
upper limit for slots.
-- Reuti
> -- Reuti
>
>
>> Still not working..
>>
>> --Sangmin
>>
>>
>> On Mon, Oct 28, 2013 at 9:47 PM, Reuti <[email protected]> wrote:
>> Am 28.10.2013 um 13:45 schrieb Sangmin Park:
>>
>>> This is the RQS
>>>
>>> limit hosts {@parallelhosts} to slots=$num_proc
>>> limit queues !matlab.q hosts {@matlabhosts} to slots=$num_proc
>>> parallelhosts include matlabhosts.
>>>
>>> slots value in the matlab.q means the number of cores per node.
>>>
>>> All hosts is included in parallelhosts, node1 ~ node30.
>>> matlabhosts include node1 ~ node7.
>>> short.q, normal.q and long.q could be used in node1 ~ node7.
>>>
>>> I want to set up when jobs with short.q, normal.q and long.q are running,
>>> if matlab job is submitted,
>>> running job not using matlab.q in node1 ~ node7 is suspended and matlab job
>>> is run.
>>> This is what I want to set up.
>>>
>>> I don't understand why it can not be happened if I setup slots value 12.
>>
>> It will suspend the job when all 12 slots are used on a particular host. You
>> may want to try with 1 instead. As s refinement, you could also look into
>> slotwise subordination.
>>
>> -- Reuti
>>
>>
>>> --Sangmin
>>>
>>>
>>> On Mon, Oct 28, 2013 at 8:58 PM, Reuti <[email protected]> wrote:
>>> Am 28.10.2013 um 12:30 schrieb Sangmin Park:
>>>
>>>> I've edit the negative value in the priority section, short.q is 4,
>>>> normal.q is 6 and long.q is 8, respectively.
>>>> And I configured 72 cores for each queues.
>>>
>>> But you didn't answer the question: How do you limit the overall slot
>>> count? RQS oder definition in the exechost?
>>>
>>>> Below is matlab.q instance details.
>>>> qname matlab.q
>>>> hostlist @matlabhosts
>>>> seq_no 0
>>>> load_thresholds np_load_avg=1.75
>>>> suspend_thresholds NONE
>>>> nsuspend 1
>>>> suspend_interval 00:05:00
>>>> priority 2
>>>> min_cpu_interval 00:05:00
>>>> processors UNDEFINED
>>>> qtype BATCH INTERACTIVE
>>>> ckpt_list NONE
>>>> pe_list fill_up make matlab
>>>> rerun FALSE
>>>> slots 12
>>>> tmpdir /tmp
>>>> shell /bin/bash
>>>> prolog NONE
>>>> epilog NONE
>>>> shell_start_mode posix_compliant
>>>> starter_method NONE
>>>> suspend_method NONE
>>>> resume_method NONE
>>>> terminate_method NONE
>>>> notify 00:00:60
>>>> owner_list NONE
>>>> user_lists octausers onsiteusers
>>>> xuser_lists NONE
>>>> subordinate_list short.q=72, normal.q=72, long.q=72
>>>
>>> This will suspend these tree queues when 72 slots per queue instance in
>>> matlab.q is used. As you have only 12 defined above, this will never happen.
>>>
>>> What behavior would you like to set up?
>>>
>>> -- Reuti
>>>
>>>
>>>> complex_values NONE
>>>> projects NONE
>>>> xprojects NONE
>>>> calendar NONE
>>>> initial_state default
>>>> s_rt INFINITY
>>>> h_rt 168:00:00
>>>> s_cpu INFINITY
>>>> h_cpu INFINITY
>>>> s_fsize INFINITY
>>>> h_fsize INFINITY
>>>> s_data INFINITY
>>>> h_data INFINITY
>>>> s_stack INFINITY
>>>> h_stack INFINITY
>>>> s_core INFINITY
>>>> h_core INFINITY
>>>> s_rss INFINITY
>>>> h_rss INFINITY
>>>> s_vmem INFINITY
>>>> h_vmem INFINITY
>>>>
>>>> thanks,
>>>>
>>>> --Sangmin
>>>>
>>>>
>>>> On Mon, Oct 28, 2013 at 3:51 PM, Reuti <[email protected]> wrote:
>>>> Hi,
>>>>
>>>> Am 28.10.2013 um 06:40 schrieb Sangmin Park:
>>>>
>>>>> Thanks, adam
>>>>>
>>>>> I configured sge queue configuration following second link you said.
>>>>> But, it does not work.
>>>>>
>>>>> I make 4 queues, short.q, normal.q, long.q and matlab.q
>>>>> short.q, normal.q and long.q queue instances are running all computing
>>>>> nodes, node1 ~ node30.
>>>>> matlab.q instance is configured only for a few nodes, node1 ~ node7,
>>>>> called matlabhosts
>>>>>
>>>>> The priorities of each queue is below.
>>>>> [short.q]
>>>>> priority -5
>>>>
>>>> Don't use negative values here. This number is the "nice value" under
>>>> which the Linux kernel will run the process (i.e. the scheduler in the
>>>> kernel, for SGE it doesn't influence the scheduling). User processes
>>>> should be in the range 0..19 [20 on Solaris]. The negative ones are
>>>> reserved for kernel processes.
>>>>
>>>>
>>>>> subordinate_list NONE
>>>>> [normal.q]
>>>>> priority 0
>>>>> subordinate_list NONE
>>>>> [long.q]
>>>>> priority 5
>>>>> subordinate_list NONE
>>>>>
>>>>> and matlab.q is
>>>>> priority -10
>>>>> subordinate_list short.q normal.q long.q
>>>>
>>>> Same here. It's also worth to note, that these values are relative. I.e.
>>>> having the same number of user processes and cores, it doesn't matter
>>>> which values are used as nice values, as each process gets it's own core
>>>> anyway. Only when there are more processes than cores it will have an
>>>> effect. But as these are relative values, it's the same whether (cores+1)
>>>> processes have all 0 or 19 as nice value.
>>>>
>>>>
>>>>> I submited several jobs using normal.q to the matlabhosts
>>>>> and I submited a job using matlab.q that has subordinate_list
>>>>> I expected one of normal.q queue job is suspended and matlab.q queue job
>>>>> is running.
>>>>> But, matlab.q queue job waits in queue with status qw. not submitted.
>>>>>
>>>>> what's the matter with this?
>>>>> please help!!
>>>>
>>>> http://gridengine.org/pipermail/users/2013-October/006820.html
>>>>
>>>> How do you limit the overall slot count?
>>>>
>>>> -- Reuti
>>>>
>>>>
>>>>> Sangmin
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Tue, Oct 15, 2013 at 3:50 PM, Adam Brenner <[email protected]> wrote:
>>>>> Sangmin,
>>>>>
>>>>> I believe the phrase / term you are looking for is Subordinate
>>>>> Queues[1][2]. This should handle what you are looking for.
>>>>>
>>>>> If not ... I am sure Reuti (or someone else) will correct me on this.
>>>>>
>>>>> Enjoy,
>>>>> -Adam
>>>>>
>>>>> [1]: http://docs.oracle.com/cd/E19957-01/820-0698/i998889/index.html
>>>>> [2]:
>>>>> http://grid-gurus.blogspot.com/2011/03/using-grid-engine-subordinate-queues.html
>>>>>
>>>>> --
>>>>> Adam Brenner
>>>>> Computer Science, Undergraduate Student
>>>>> Donald Bren School of Information and Computer Sciences
>>>>>
>>>>> Research Computing Support
>>>>> Office of Information Technology
>>>>> http://www.oit.uci.edu/rcs/
>>>>>
>>>>> University of California, Irvine
>>>>> www.ics.uci.edu/~aebrenne/
>>>>> [email protected]
>>>>>
>>>>>
>>>>> On Mon, Oct 14, 2013 at 11:18 PM, Sangmin Park <[email protected]>
>>>>> wrote:
>>>>>> Howdy,
>>>>>>
>>>>>> For specific purpose in my organization,
>>>>>> I want to configure something to SGE scheduler.
>>>>>>
>>>>>> Imazine.
>>>>>> a job is running, called A-job.
>>>>>> If B-job is submitted during A-job is running,
>>>>>> I want to hold A-job and run B-job first.
>>>>>> And after B-job is finished, restart A-job.
>>>>>>
>>>>>> What do I do for this?
>>>>>>
>>>>>> Sangmin
>>>>>>
>>>>>> --
>>>>>> ===========================
>>>>>> Sangmin Park
>>>>>> Supercomputing Center
>>>>>> Ulsan National Institute of Science and Technology(UNIST)
>>>>>> Ulsan, 689-798, Korea
>>>>>>
>>>>>> phone : +82-52-217-4201
>>>>>> mobile : +82-10-5094-0405
>>>>>> fax : +82-52-217-4209
>>>>>> ===========================
>>>>>>
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> [email protected]
>>>>>> https://gridengine.org/mailman/listinfo/users
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> ===========================
>>>>> Sangmin Park
>>>>> Supercomputing Center
>>>>> Ulsan National Institute of Science and Technology(UNIST)
>>>>> Ulsan, 689-798, Korea
>>>>>
>>>>> phone : +82-52-217-4201
>>>>> mobile : +82-10-5094-0405
>>>>> fax : +82-52-217-4209
>>>>> ===========================
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> [email protected]
>>>>> https://gridengine.org/mailman/listinfo/users
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> ===========================
>>>> Sangmin Park
>>>> Supercomputing Center
>>>> Ulsan National Institute of Science and Technology(UNIST)
>>>> Ulsan, 689-798, Korea
>>>>
>>>> phone : +82-52-217-4201
>>>> mobile : +82-10-5094-0405
>>>> fax : +82-52-217-4209
>>>> ===========================
>>>
>>>
>>>
>>>
>>> --
>>> ===========================
>>> Sangmin Park
>>> Supercomputing Center
>>> Ulsan National Institute of Science and Technology(UNIST)
>>> Ulsan, 689-798, Korea
>>>
>>> phone : +82-52-217-4201
>>> mobile : +82-10-5094-0405
>>> fax : +82-52-217-4209
>>> ===========================
>>
>>
>>
>>
>> --
>> ===========================
>> Sangmin Park
>> Supercomputing Center
>> Ulsan National Institute of Science and Technology(UNIST)
>> Ulsan, 689-798, Korea
>>
>> phone : +82-52-217-4201
>> mobile : +82-10-5094-0405
>> fax : +82-52-217-4209
>> ===========================
>
>
> _______________________________________________
> users mailing list
> [email protected]
> https://gridengine.org/mailman/listinfo/users
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users