Hi all, I want to be able to submit 16 serial jobs on my compute nodes at the same time since each node is 2 sockets, 4 core, hyperthreading. We see a speedup when saturating the node with 16 different serial jobs (launched manually) so I want to take advantage of this with slurm.
I tough it would be easy... Unfortunately, I always get at most 8 jobs running on nodes. Here is the relevant (I think) part of /etc/slurm.conf: # SCHEDULING #DefMemPerCPU=0 FastSchedule=1 #MaxMemPerCPU=0 #SchedulerRootFilter=1 #SchedulerTimeSlice=30 SchedulerType=sched/backfill SchedulerPort=7321 SelectType=select/cons_res NodeName=node[69-71] RealMemory=23000 Sockets=2 CoresPerSocket=4 ThreadsPerCore=2 State=UNKNOWN PartitionName=test Nodes=node[69-71] MaxTime=INFINITE State=UP The logs don't show anything interesting. For example, setting ThreadsPerCore to 1 will print a warning for the compute nodes that the number of hardward cpu is not the same as the config's. So the compute nodes are correctly detecting the number of threads possible. How can I achieve this? Thanks!
