You may consider specification of licenses in slurm.conf and for jobs. For
example license=storage1*3 and #SBATCH -l storage1.

If you cannot rely on users you can use submit plugins and namespaces to
umount storage for jobs without specific license.

cheers,
Marcin

2017-09-28 17:12 GMT+02:00 E V <eliven...@gmail.com>:

>
> Looking through the man page on slurm.conf I don't see a way to set a
> limit for a partition on the max number of jobs to start up at a time.
> Is there such a thing?
>
> What I'm trying to accomplish is essentially IO throttling. I have
> different storage systems hooked up to a set of compute nodes. Let's
> say storage x can support 1 job at a time well, i.e. that job is now
> CPU bound. Running more then 1 becomes disk bound and they thrash and
> throughput goes down. Storage y can support 3 jobs at a time before
> throughput starts dropping. So I'd like to allow only 1 job to run at
> a time using x, but up to 3 using y across any of the nodes in the
> partition.
>
> So the 2 ways I thought of to accomplish this are either create
> separate partitions for x & y and limit the total number of jobs that
> can be run from that partition at a time(which I can't find out how to
> do,) or use a single partition and have Gres handle the job counting
> with an x & y gres. However, Gres doesn't appear to have the concept
> of an overall count for the partition or cluster, only a per node
> count. So I'm stumped. Am I missing something, or is there another way
> of accomplishing this?
>

Reply via email to