Dear all,
We would like to use slurm to start managing the compute resources of a small cluster of nodes, O(10). Please, let me check with you what is the correct way to setup a cluster where users shall login to a number of separate nodes to submit their jobs. Ideally, those login nodes do not take part of the compute pool. Additionally, gres is used to control gpu resources. The hardware of the compute and login pools is different (e.g. lack of GPUs on the login part). We have got a working configuration where all compute nodes share the same slurm.conf file. In the login node, slurm.conf is slightly modified because Gres fails to load in absence of the GPU dev files. That is, login nodes have the option GresTypes commented. I have seen the option NO_CONF_HASH to prevent from logging conf. related error messages. I'm not sure this is desired though. Could you please shed some light into this? I have also seen the --enable-front-end option, but I'm not sure it applies to this case, does it? Thank you! Carlos
