Not really related to the question, but the documentation you are referencing is years old. You should probably reference the current documentation at SchedMD
http://slurm.schedmd.com In this case http://slurm.schedmd.com/cons_res_share.html On January 29, 2016 6:42:24 AM PST, Benjamin Redling <[email protected]> wrote: > >Am 29.01.2016 um 15:31 schrieb Dennis Mungai: >> Add SHARE=FORCE to your partition settings for each partition entry >in >> the configuration file. > >https://computing.llnl.gov/linux/slurm/cons_res_share.html > >selection setting was: >SelectType=select/cons_res >SelectTypeParameters=CR_Core_Memory > >Shared=FORCE as you recommend leads to: >" >Cores are allocated to jobs. A core may run more than one job. >" > >What does that have to do with the problem? >Can you elaborate on that? > >/Benjamin > > >> On Jan 29, 2016 5:08 PM, David Roman <[email protected]> wrote: >> Hello, >> >> I'm a newbies with SLURM. Perhaps could you help me to understand my >> mistake. >> >> I have 2 nodes (2 sockets with 4 core per socket = 8 CPUs per node) I >> created 3 partitions >> >> DEV with node2 >> OP with node1 >> LOW with node1 and node2 >> >> I created 2 jobs >> Job_A uses 8 CPUS in partion DEV >> Job_B uses 16 CPUS in partion LOW >> >> If I start Job_A before Job_B, all is ok. Job_A is in RUNNING state >and >> Job_B is in PENDING state >> >> BUT, If I start Job_B before Job_A. The both jobs are in RUNNING >state. >> >> Thanks for your help, >> >> David. >> >> >> Here my slurm.conf without comments >> >> ClusterName=Noveltits >> ControlMachine=slurm >> SlurmUser=slurm >> SlurmctldPort=6817 >> SlurmdPort=6818 >> AuthType=auth/munge >> StateSaveLocation=/tmp >> SlurmdSpoolDir=/tmp/slurmd >> SwitchType=switch/none >> MpiDefault=none >> SlurmctldPidFile=/var/run/slurmctld.pid >> SlurmdPidFile=/var/run/slurmd.pid >> ProctrackType=proctrack/pgid >> CacheGroups=0 >> ReturnToService=0 >> SlurmctldTimeout=300 >> SlurmdTimeout=300 >> InactiveLimit=0 >> MinJobAge=300 >> KillWait=30 >> Waittime=0 >> SchedulerType=sched/backfill >> SelectType=select/cons_res >> SelectTypeParameters=CR_CORE_Memory >> FastSchedule=0 >> SlurmctldDebug=3 >> SlurmdDebug=3 >> JobCompType=jobcomp/none >> >> PreemptMode=SUSPEND,GANG >> PreemptType=preempt/partition_prio >> >> >> NodeName=slurm_node[1-2] CPUs=8 SocketsPerBoard=2 CoresPerSocket=4 >> ThreadsPerCore=1 >> PartitionName=op Nodes=slurm_node1 Priority=100 Default=No >> MaxTime=INFINITE State=UP >> PartitionName=dev Nodes=slurm_node2 Priority=1 Default=yes >> MaxTime=INFINITE State=UP PreemptMode=OFF >> PartitionName=low Nodes=slurm_node[1-2] Priority=1 Default=No >> MaxTime=INFINITE State=UP >> >> >> >______________________________________________________________________ >> >> This e-mail contains information which is confidential. It is >intended >> only for the use of the named recipient. If you have received this >> e-mail in error, please let us know by replying to the sender, and >> immediately delete it from your system. Please note, that in these >> circumstances, the use, disclosure, distribution or copying of this >> information is strictly prohibited. KEMRI-Wellcome Trust Programme >> cannot accept any responsibility for the accuracy or completeness of >> this message as it has been transmitted over a public network. >Although >> the Programme has taken reasonable precautions to ensure no viruses >are >> present in emails, it cannot accept responsibility for any loss or >> damage arising from the use of the email or attachments. Any views >> expressed in this message are those of the individual sender, except >> where the sender specifically states them to be the views of >> KEMRI-Wellcome Trust Programme. >> >______________________________________________________________________ > >-- >FSU Jena | JULIELab.de/Staff/Benjamin+Redling.html >vox: +49 3641 9 44323 | fax: +49 3641 9 44321
