> Am 20.12.2016 um 02:45 schrieb John_Tai <john_...@smics.com>:
> 
> I spoke too soon. I can request PE and virtual_free separately, but I cannot 
> request both:
> 
> 
> 
> # qsub -V -b y -cwd -now n -pe cores 7 -l mem=10G -q all.q@ibm037 xclock

Above you request "mem" (which is a snapshot of the actual usage and may vary 
over the runtime of other jobs [unless they request the total amount already at 
the beginning of the job and stay with it]).

> Your job 180 ("xclock") has been submitted
> # qstat
> job-ID  prior   name       user         state submit/start at     queue       
>                    slots ja-task-ID
> -----------------------------------------------------------------------------------------------------------------
>    180 0.55500 xclock     johnt        qw    12/20/2016 09:43:41              
>                       7
> # qstat -j 180
> ==============================================================
> job_number:                 180
> exec_file:                  job_scripts/180
> submission_time:            Tue Dec 20 09:43:41 2016
> owner:                      johnt
> uid:                        162
> group:                      sa
> gid:                        4563
> sge_o_home:                 /home/johnt
> sge_o_log_name:             johnt
> sge_o_path:                 
> /home/sge/sge8.1.9-1.el5/bin:/home/sge/sge8.1.9-1.el5/bin/lx-amd64:/bin:/usr/bin:/usr/local/bin:/usr/X11R6/bin:/home/johnt/bin:.
> sge_o_shell:                /bin/tcsh
> sge_o_workdir:              /home/johnt/sge8
> sge_o_host:                 ibm005
> account:                    sge
> cwd:                        /home/johnt/sge8
> hard resource_list:         virtual_free=10G

10G times 7 = 70 GB

The node has this amount of memory installed and it is defined this way in 
`qconf -me ibm037`?

-- Reuti


> mail_list:                  johnt@ibm005
> notify:                     FALSE
> job_name:                   xclock
> jobshare:                   0
> hard_queue_list:            all.q@ibm037
> env_list:                   TERM=xterm,DISPLAY=dsls11:3. [..]
> script_file:                xclock
> parallel environment:  cores range: 7
> binding:                    NONE
> job_type:                   binary
> scheduling info:            cannot run in queue "sim.q" because it is not 
> contained in its hard queue list (-q)
>                            cannot run in queue "pc.q" because it is not 
> contained in its hard queue list (-q)
>                            cannot run in PE "cores" because it only offers 0 
> slots
> 
> 
> 
> 
> 
> -----Original Message-----
> From: Reuti [mailto:re...@staff.uni-marburg.de]
> Sent: Saturday, December 17, 2016 10:16
> To: Reuti
> Cc: John_Tai; users@gridengine.org; Coleman, Marcus [JRDUS Non-J&J]
> Subject: Re: [gridengine users] John's cores pe (Was: users Digest...)
> 
> 
> Am 17.12.2016 um 11:34 schrieb Reuti:
> 
>> 
>> Am 17.12.2016 um 02:01 schrieb John_Tai:
>> 
>>> It is working!! Thank you to all that replied to me and helped me figure 
>>> this out.
>>> 
>>> I meant to set the default to 2G so that was my mistake. I changed it to:
>>> 
>>> virtual_free        mem        MEMORY    <=    YES         YES        2G    
>>>    0
>> 
>> That's strange. A plain "2" was for me always two bytes. A "h_vmem" of 2 
>> bytes would crash the job instantly when it got scheduled, but for 
>> "virtual_free" (which is only a guidance for SGE how to distribute jobs) it 
>> shouldn't hinder the scheduling at all.
>> 
>> `man sge_types` also lists:
>> 
>>      If no multiplier is present, the value is  just  counted  in bytes.
> 
> We have set "-w e" in /usr/sge/default/common/sge_request, and then I even 
> face an "Unable to run job: error: no suitable queues." This happens whether 
> the low 2 byte value is specified in the complex definition `qconf -mc` or on 
> the command line as "-l virutal_free=2".
> 
> It turns out, that the minimum value which is being accepted is: 33.
> 
> -- Reuti
> 
> 
>> 
>>> And it's working now. Although I'm not sure why it affected the PE.
>>> 
>>> Also I didn't set a global one, what is the purpose of the global one? 
>>> Should I set it?
>> 
>> No, it was only one place I would have checked too. The global complexes 
>> therein can for example be used for a limit in the number of licenses of an 
>> application you have and which can be used floating in the cluster (one 
>> could prefer to put such a limit in an RQS though).
>> 
>> If you would have set it up there, it would have been the "overall limit of 
>> memory which can be used in the complete cluster at the same time".
>> 
>> -- Reuti
>> 
>> 
>>> # qconf -se global
>>> hostname              global
>>> load_scaling          NONE
>>> complex_values        NONE
>>> load_values           NONE
>>> processors            0
>>> user_lists            NONE
>>> xuser_lists           NONE
>>> projects              NONE
>>> xprojects             NONE
>>> usage_scaling         NONE
>>> report_variables      NONE
>>> 
>>> 
>>> -----Original Message-----
>>> From: Reuti [mailto:re...@staff.uni-marburg.de]
>>> Sent: Friday, December 16, 2016 7:36
>>> To: John_Tai
>>> Cc: Christopher Heiny; users@gridengine.org; Coleman, Marcus [JRDUS Non-J&J]
>>> Subject: Re: [gridengine users] John's cores pe (Was: users Digest...)
>>> 
>>> 
>>>> Am 16.12.2016 um 09:53 schrieb John_Tai <john_...@smics.com>:
>>>> 
>>>> virtual_free        mem        MEMORY    <=    YES         YES        2    
>>>>     0
>>> 
>>> This would mean, that the default consumption is 2 bytes. I already feared 
>>> that a high values was programmed here. More suitable would be a default of 
>>> 1G or so.
>>> 
>>> Is there any virtual_free complex defined on a global level: qconf -se 
>>> global
>>> 
>>> -- Reuti
>>> ________________________________
>>> 
>>> This email (including its attachments, if any) may be confidential and 
>>> proprietary information of SMIC, and intended only for the use of the named 
>>> recipient(s) above. Any unauthorized use or disclosure of this email is 
>>> strictly prohibited. If you are not the intended recipient(s), please 
>>> notify the sender immediately and delete this email from your computer.
>>> 
>> 
>> 
>> _______________________________________________
>> users mailing list
>> users@gridengine.org
>> https://gridengine.org/mailman/listinfo/users
> 
> ________________________________
> 
> This email (including its attachments, if any) may be confidential and 
> proprietary information of SMIC, and intended only for the use of the named 
> recipient(s) above. Any unauthorized use or disclosure of this email is 
> strictly prohibited. If you are not the intended recipient(s), please notify 
> the sender immediately and delete this email from your computer.
> 


_______________________________________________
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users

Reply via email to