Setting SLURM_EXCLUSIVE environment variable will make this the
default behavior for srun commands. You could just set this for
this particular user if desired.
Moe
Quoting [email protected]:
Ideally I'd like a way of setting this as the default for (s)batch as
it provides an avenue for users to overload the nodes thinking that
they're running faster (don't get me started...). I can insist the
user runs their job steps serially but it's hard to enforce without a
lot of effort.
Jeff
On Fri, Apr 29, 2011 at 2:05 PM, Auble, Danny <[email protected]> wrote:
Hey Jeff, Thanks for the accolades ;).
Have you tried the srun '--exclusive' option? That should keep
things separate.
Let us know if that doesn't work,
Danny
-----Original Message-----
From: [email protected]
[mailto:[email protected]] On Behalf Of
[email protected]
Sent: Friday, April 29, 2011 2:01 PM
To: [email protected]
Subject: [slurm-dev] Stupid Sbatch Question
I have a user (in the same sense that I have an ingrown toenail) who
runs a job like this...
She starts with 'sbatch -p whatever -n 145 scriptname'
The script contains something like:
srun foo &
srun bar
What's happening is that each core is doubly allocated e.g. 16 procs
running on an 8 core node. Why isn't the second srun constrained by
the resources consumed by the 1st?
This is running under (the very excellent) Slurm 2.2.3.
Jeff Katcher
FHCRC Cluster Monkey