You can also use sbatch --gres=bandwidth ....
On Fri, 2014-05-30 at 10:34 -0700, [email protected] wrote: > Define a GRES called "bandwidth" and create a job_submit plugin that > sets a GRES value of "bandwidth:1" for jobs. It's relatively simple. > See: > http://slurm.schedmd.com/gres.html > http://slurm.schedmd.com/job_submit_plugins.html > > Moe Jette > SchedMD > > > Quoting Brian Baughman <[email protected]>: > > > Greetings, > > > > We run multiple types of jobs on our cluster and many of our nodes > > have 48 cores or more. We have found that some jobs are idle on such > > nodes when are accessing data in aggregate over the available > > bandwidth. I was thinking of trying to create a “high-bandwidth” > > queue which would only allow say 10 such processes to run on each > > node so the bandwidth wouldn’t become a problem. Is such a thing > > possible with the slurm scheduler? If not any suggestions on how to > > solve such a problem? > > > > Regards, > > Brian >
