Am Fri, 28 Oct 2016 00:21:12 -0700
schrieb Manuel Rodríguez Pascual <manuel.rodriguez.pasc...@gmail.com>:

> Altogether, I think the condition should be rewritten to something like:
>                     if ((opt.ntasks_per_node != NO_VAL) &&
>                             if (opt.ntasks < opt.ntasks_per_node
> *opt.min_nodes )
> 
> but with opt.ntasks being the value introduced by the user, not the one
> internally considered by Slurm at this point.  I don't know how to do
> correct this, but I hope this helps to point towards the problem.

Is there any comment on this by Slurm developers? We tried upgrading to
Slurm 16.05.6 here and hit this bug, mpiexec/srun complaining about
requested tasks per node that do not match what the user requested. We
downgraded to 15.08.12 and that seems to work fine.

I am astonished that there is no reaction at all to this … also not to
the bug tracker report under

        https://bugs.schedmd.com/show_bug.cgi?id=3032

. It looks to me like Slurm 16.05 clearly broke somethinh. Well … it is
broken for at least three people. Are we facing a strange fringe use
case?

Is Intel MPI just buggy? I am not sure about OpenMPI, I figured I have
to downgrade Slurm when Intel MPI does not work properly anymore. Even
if this is ultimately Intel MPI's fault, this would be a strong reason
for us to keep Slurm at the older version for the whole lifetime of our
cluster in order to support the existing binaries.


Alrighty then,

Thomas

-- 
Dr. Thomas Orgis
Universität Hamburg
RRZ / Basis-Infrastruktur / HPC
Schlüterstr. 70
20146 Hamburg
Tel.: 040/42838 8826
Fax: 040/428 38 6270

Attachment: smime.p7s
Description: S/MIME cryptographic signature

Reply via email to