It would be helpful to see how you submitted the job. And the output from 
"scontrol show job 20".

Regards,

        Uwe

Am 30.03.2015 um 19:49 schrieb Carl E. Fields:
> Hello,
> 
> I have installed slurm version version 14.11.4 on a RHEL server with the 
> following specs:
> 
> 
> Architecture:          x86_64
> 
> CPU op-mode(s):        32-bit, 64-bit
> 
> Byte Order:            Little Endian
> 
> CPU(s):                2
> 
> On-line CPU(s) list:   0,1
> 
> Thread(s) per core:    1
> 
> Core(s) per socket:    2
> 
> Socket(s):             1
> 
> NUMA node(s):          1
> 
> Vendor ID:             GenuineIntel
> 
> CPU family:            6
> 
> Model:                 23
> 
> Stepping:              6
> 
> CPU MHz:               2300.000
> 
> BogoMIPS:              4600.00
> 
> Hypervisor vendor:     VMware
> 
> Virtualization type:   full
> 
> L1d cache:             32K
> 
> L1i cache:             32K
> 
> L2 cache:              256K
> 
> L3 cache:              15360K
> 
> NUMA node0 CPU(s):     0,1
> 
> 
> 
> I wish to designate one core as the controller. And another core as available 
> for job submissions which require 1 core. 
> 
> I have configured everything however, I believe I have an error in my 
> slurm.conf file because when I submit a job, it
> sits in the queue with node reason as below:
> 
>              JOBID PARTITION     NAME     USER    STATE       TIME TIME_LIMI  
> NODES NODELIST(REASON)
> 
>                 20   compute calculat SlurmUse  PENDING       0:00     10:00  
>     1 (Resources)
> 
> 
> 
> I believe I am not properly configuring the resources in my file but I am 
> unsure of wherein the issue lies. I hope one
> can assist me in properly configuring my server. Thank you in advance
> 
> My current slurm.conf file:
> 
> 
> 
> [SlurmUser@sod264 etc]$ cat slurm.conf 
> 
> # slurm.conf file generated by configurator easy.html.
> 
> # Put this file on all nodes of your cluster.
> 
> # See the slurm.conf man page for more information.
> 
> #
> 
> ControlMachine=sod264
> 
> ControlAddr=129.XXX
> 
> # 
> 
> #MailProg=/bin/mail 
> 
> MpiDefault=none
> 
> #MpiParams=ports=#-# 
> 
> ProctrackType=proctrack/pgid
> 
> ReturnToService=0
> 
> SlurmctldPidFile=/var/run/slurmctld.pid
> 
> #SlurmctldPort=6817 
> 
> SlurmdPidFile=/var/run/slurmd.pid
> 
> #SlurmdPort=6818 
> 
> SlurmdSpoolDir=/var/spool/slurmd
> 
> SlurmUser=SlurmUser
> 
> SlurmdUser=SlurmUser 
> 
> StateSaveLocation=/var/spool/statesave
> 
> SwitchType=switch/none
> 
> TaskPlugin=task/none
> 
> # 
> 
> # 
> 
> # TIMERS 
> 
> #KillWait=30 
> 
> #MinJobAge=300 
> 
> #SlurmctldTimeout=120 
> 
> #SlurmdTimeout=300 
> 
> # 
> 
> # 
> 
> # SCHEDULING 
> 
> FastSchedule=1
> 
> SchedulerType=sched/backfill
> 
> #SchedulerPort=7321 
> 
> #SelectType=select/serial
> 
> SelectType=select/cons_res
> 
> SelectTypeParameters=CR_CORE
> 
> # 
> 
> # 
> 
> # LOGGING AND ACCOUNTING 
> 
> AccountingStorageType=accounting_storage/none
> 
> ClusterName=MESA-Web
> 
> #JobAcctGatherFrequency=30 
> 
> JobAcctGatherType=jobacct_gather/none
> 
> SlurmctldDebug=3
> 
> SlurmctldLogFile=/var/log/slurm/slurmctld.log
> 
> SlurmdDebug=3
> 
> SlurmdLogFile=/var/log/slurm/slurmd.log
> 
> #
> 
> 
> # 
> 
> # COMPUTE NODES 
> 
> NodeName=sod264 Sockets=1 CoresPerSocket=2 ThreadsPerCore=1 RealMemory=128940 
> TmpDisk=19895  
> 
> 
> PartitionName=compute Nodes=sod264 Default=YES STATE=UP
> 
> 
> 
> 
> Kind Regards,
> 
> Carl
> 
> 
> 

Reply via email to