Ideally I'd like a way of setting this as the default for (s)batch as it provides an avenue for users to overload the nodes thinking that they're running faster (don't get me started...). I can insist the user runs their job steps serially but it's hard to enforce without a lot of effort.
Jeff On Fri, Apr 29, 2011 at 2:05 PM, Auble, Danny <[email protected]> wrote: > Hey Jeff, Thanks for the accolades ;). > > Have you tried the srun '--exclusive' option? That should keep things > separate. > > Let us know if that doesn't work, > Danny > > >> -----Original Message----- >> From: [email protected] [mailto:[email protected]] >> On Behalf Of >> [email protected] >> Sent: Friday, April 29, 2011 2:01 PM >> To: [email protected] >> Subject: [slurm-dev] Stupid Sbatch Question >> >> I have a user (in the same sense that I have an ingrown toenail) who >> runs a job like this... >> >> She starts with 'sbatch -p whatever -n 145 scriptname' >> The script contains something like: >> srun foo & >> srun bar >> >> What's happening is that each core is doubly allocated e.g. 16 procs >> running on an 8 core node. Why isn't the second srun constrained by >> the resources consumed by the 1st? >> >> This is running under (the very excellent) Slurm 2.2.3. >> >> Jeff Katcher >> FHCRC Cluster Monkey > >
