Hi all,

I want to be able to submit 16 serial jobs on my compute nodes at the
same time since each node is 2 sockets, 4 core, hyperthreading. We see
a speedup when saturating the node with 16 different serial jobs
(launched manually) so I want to take advantage of this with slurm.

I tough it would be easy...

Unfortunately, I always get at most 8 jobs running on nodes.

Here is the relevant (I think) part of /etc/slurm.conf:
# SCHEDULING
#DefMemPerCPU=0
FastSchedule=1
#MaxMemPerCPU=0
#SchedulerRootFilter=1
#SchedulerTimeSlice=30
SchedulerType=sched/backfill
SchedulerPort=7321
SelectType=select/cons_res
NodeName=node[69-71] RealMemory=23000 Sockets=2 CoresPerSocket=4
ThreadsPerCore=2 State=UNKNOWN
PartitionName=test         Nodes=node[69-71]
MaxTime=INFINITE State=UP

The logs don't show anything interesting. For example, setting
ThreadsPerCore to 1 will print a warning for the compute nodes that
the number of hardward cpu is not the same as the config's. So the
compute nodes are correctly detecting the number of threads possible.

How can I achieve this?

Thanks!

Reply via email to