It would be helpful to see how you submitted the job. And the output from "scontrol show job 20".
Regards, Uwe Am 30.03.2015 um 19:49 schrieb Carl E. Fields: > Hello, > > I have installed slurm version version 14.11.4 on a RHEL server with the > following specs: > > > Architecture: x86_64 > > CPU op-mode(s): 32-bit, 64-bit > > Byte Order: Little Endian > > CPU(s): 2 > > On-line CPU(s) list: 0,1 > > Thread(s) per core: 1 > > Core(s) per socket: 2 > > Socket(s): 1 > > NUMA node(s): 1 > > Vendor ID: GenuineIntel > > CPU family: 6 > > Model: 23 > > Stepping: 6 > > CPU MHz: 2300.000 > > BogoMIPS: 4600.00 > > Hypervisor vendor: VMware > > Virtualization type: full > > L1d cache: 32K > > L1i cache: 32K > > L2 cache: 256K > > L3 cache: 15360K > > NUMA node0 CPU(s): 0,1 > > > > I wish to designate one core as the controller. And another core as available > for job submissions which require 1 core. > > I have configured everything however, I believe I have an error in my > slurm.conf file because when I submit a job, it > sits in the queue with node reason as below: > > JOBID PARTITION NAME USER STATE TIME TIME_LIMI > NODES NODELIST(REASON) > > 20 compute calculat SlurmUse PENDING 0:00 10:00 > 1 (Resources) > > > > I believe I am not properly configuring the resources in my file but I am > unsure of wherein the issue lies. I hope one > can assist me in properly configuring my server. Thank you in advance > > My current slurm.conf file: > > > > [SlurmUser@sod264 etc]$ cat slurm.conf > > # slurm.conf file generated by configurator easy.html. > > # Put this file on all nodes of your cluster. > > # See the slurm.conf man page for more information. > > # > > ControlMachine=sod264 > > ControlAddr=129.XXX > > # > > #MailProg=/bin/mail > > MpiDefault=none > > #MpiParams=ports=#-# > > ProctrackType=proctrack/pgid > > ReturnToService=0 > > SlurmctldPidFile=/var/run/slurmctld.pid > > #SlurmctldPort=6817 > > SlurmdPidFile=/var/run/slurmd.pid > > #SlurmdPort=6818 > > SlurmdSpoolDir=/var/spool/slurmd > > SlurmUser=SlurmUser > > SlurmdUser=SlurmUser > > StateSaveLocation=/var/spool/statesave > > SwitchType=switch/none > > TaskPlugin=task/none > > # > > # > > # TIMERS > > #KillWait=30 > > #MinJobAge=300 > > #SlurmctldTimeout=120 > > #SlurmdTimeout=300 > > # > > # > > # SCHEDULING > > FastSchedule=1 > > SchedulerType=sched/backfill > > #SchedulerPort=7321 > > #SelectType=select/serial > > SelectType=select/cons_res > > SelectTypeParameters=CR_CORE > > # > > # > > # LOGGING AND ACCOUNTING > > AccountingStorageType=accounting_storage/none > > ClusterName=MESA-Web > > #JobAcctGatherFrequency=30 > > JobAcctGatherType=jobacct_gather/none > > SlurmctldDebug=3 > > SlurmctldLogFile=/var/log/slurm/slurmctld.log > > SlurmdDebug=3 > > SlurmdLogFile=/var/log/slurm/slurmd.log > > # > > > # > > # COMPUTE NODES > > NodeName=sod264 Sockets=1 CoresPerSocket=2 ThreadsPerCore=1 RealMemory=128940 > TmpDisk=19895 > > > PartitionName=compute Nodes=sod264 Default=YES STATE=UP > > > > > Kind Regards, > > Carl > > >