Hello,
I've been testing out the h_vmem settings for a while now, and currently I have
this setup:
Exec host
complex_values slots=36,h_vmem=142G
high_priority.q
h_vmem INFINITY
slots 24
priority 0
low_priority.q
h_vmem INFINITY
priority 18
slots 12
qconf -sc
h_vmem h_vmem MEMORY <= YES YES
3.95G 0
I know that there has been discussion of a bug with respect to setting the
complex to JOB, which is why I settled on this configuration a few months ago
in order to have two queues without oversubscribing the memory. However, this
doesn't seem to actually limit the memory usage during run time, like I have
seen GE do before.
I have one script that I have been using to benchmark my cluster and figure out
the queue stats. It runs tophat and bowtie and my metrics for knowing if the
memory is being limited are the "Max vmem:" and "Wall clock time:" stats. If
the memory isn't limited, then if I submit the job using 24 cores, I'll see
"Max vmem: 35.342G" and a wall clock time around 2:20:00:00. When I was able
to limit the vmem, I saw stats more like " Wallclock Time = 19:51:49... Max
vmem = 3.932G". As you can see, 19 hours is a lot quicker than 2 days.
I don't have definitive proof, but I think changing to JOB and setting a limit
in the queue definition, instead of INFINITY, might restore the actual runtime
limit. But, then I wouldn't be able to have two queues in the way I have them
now. I'd like to test this myself but my tiny cluster is full at the moment.
Can anyone confirm these settings for me?
Thanks,
Brett
Brett Taylor
Systems Administrator
Center for Systems and Computational Biology
The Wistar Institute
3601 Spruce St.
Room 214
Philadelphia PA 19104
Tel: 215-495-6914
Sending me a large file? Use my secure dropbox:
https://cscb-filetransfer.wistar.upenn.edu/dropbox/[email protected]
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users