You may consider specification of licenses in slurm.conf and for jobs. For example license=storage1*3 and #SBATCH -l storage1.
If you cannot rely on users you can use submit plugins and namespaces to umount storage for jobs without specific license. cheers, Marcin 2017-09-28 17:12 GMT+02:00 E V <eliven...@gmail.com>: > > Looking through the man page on slurm.conf I don't see a way to set a > limit for a partition on the max number of jobs to start up at a time. > Is there such a thing? > > What I'm trying to accomplish is essentially IO throttling. I have > different storage systems hooked up to a set of compute nodes. Let's > say storage x can support 1 job at a time well, i.e. that job is now > CPU bound. Running more then 1 becomes disk bound and they thrash and > throughput goes down. Storage y can support 3 jobs at a time before > throughput starts dropping. So I'd like to allow only 1 job to run at > a time using x, but up to 3 using y across any of the nodes in the > partition. > > So the 2 ways I thought of to accomplish this are either create > separate partitions for x & y and limit the total number of jobs that > can be run from that partition at a time(which I can't find out how to > do,) or use a single partition and have Gres handle the job counting > with an x & y gres. However, Gres doesn't appear to have the concept > of an overall count for the partition or cluster, only a per node > count. So I'm stumped. Am I missing something, or is there another way > of accomplishing this? >