Hi Brett, Try setting consumable to YES instead of JOB.
Ian On Wed, Dec 19, 2012 at 8:05 AM, Brett Taylor <[email protected]> wrote: > Hello,**** > > ** ** > > I seem to be having some issues with the h_vmem setting and proper > allocation. I was under the impression that this would function much like > the CPU slots, and once all of the h_vmem had been allocated, that host > would not accept more jobs. This however does not seem to be the case. I > have two queues running on the same nodes. **** > > ** ** > > My hosts are defined:**** > > ** ** > > complex_values slots=48,h_vmem=95G**** > > ** ** > > complex config:**** > > ** ** > > #name shortcut type relop requestable consumable > default urgency**** > > > #---------------------------------------------------------------------------------------- > **** > > h_vmem h_vmem MEMORY <= YES JOB > 4G 0**** > > ** ** > > queue config, in both queues:**** > > ** ** > > slots 24**** > > h_vmem INFINITY**** > > ** ** > > So if I submit one job to A.q, using –l h_vmem=95G, then `qstat -f -u "*" > -F h_vmem` shows:**** > > ** ** > > > --------------------------------------------------------------------------------- > **** > > [email protected] BIP 0/24/24 1.02 lx26-amd64**** > > hc:h_vmem=0.000**** > > 3970 0.60500 bt_script. btaylor r 12/17/2012 11:26:45 24*** > * > > > --------------------------------------------------------------------------------- > **** > > [email protected] BP 0/0/24 1.02 lx26-amd64**** > > hc:h_vmem=0.000**** > > ** ** > > but if I then submit to the 2nd queue, the h_vmem will go negative and > allow both jobs to run at the same time:**** > > ** ** > > > --------------------------------------------------------------------------------- > **** > > [email protected] BIP 0/24/24 1.01 lx26-amd64**** > > hc:h_vmem=-95.000G**** > > 3970 0.60500 bt_script. btaylor r 12/17/2012 11:26:45 24*** > * > > > --------------------------------------------------------------------------------- > **** > > [email protected] BP 0/24/24 1.01 lx26-amd64**** > > hc:h_vmem=-95.000G**** > > 4012 0.60500 bt_script. btaylor r 12/19/2012 11:03:04 24*** > * > > ** ** > > ** ** > > Is this not supposed to act like the cpu slots? Any ideas on how I might > be able to treat available vmem the same as the cpu slots?**** > > ** ** > > Thanks,**** > > Brett**** > > ** ** > > ** ** > > ** ** > > Brett Taylor**** > > Systems Administrator**** > > Center for Systems and Computational Biology**** > > > The Wistar Institute**** > > 3601 Spruce St.**** > > Room 214**** > > Philadelphia PA 19104**** > > Tel: 215-495-6914**** > > Sending me a large file? Use my secure dropbox:**** > > https://cscb-filetransfer.wistar.upenn.edu/dropbox/[email protected]**** > > ** ** > > _______________________________________________ > users mailing list > [email protected] > https://gridengine.org/mailman/listinfo/users > > -- Ian Kaufman Research Systems Administrator UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
_______________________________________________ users mailing list [email protected] https://gridengine.org/mailman/listinfo/users
