> Couple of comments. Your original cmd line:
> 
> >>   srun -n 2 mpirun MPI-hellow
> 
> tells srun to launch two copies of mpirun, each of which is to run as many 
> processes as there are slots assigned to the allocation. srun will get an 
> allocation of two slots, and so you?ll get two concurrent MPI jobs, each 
> consisting of two procs.

Exactly. Hence, launching a job via this "minimal command line" and a certain 
number x after the -n, will always occupy x^2 processors.
Just out of curiocity: In which situation is this reasonable? IMHO the number 
of concurrent MPI jobs on one hand and the number of procs allocated by each of 
them on the other, these are quite independent parameters.


> Your other cmd line:
> 
> >>        srun -c 2 mpirun -np 2 MPI-hellow
> 
> told srun to get two slots but only run one copy (the default value of the -n 
> option) of mpirun, and you told mpirun to launch two procs. So you got one 
> job consisting of two procs.

Exactly. Sadly, it bails out with Slurm 16.05.9 + OpenMPI 2.0.2 (while it used 
to work with Slurm 14.03.9 + OpenMPI 1.6.5). But most probabably I'm barking up 
the wrong tree here: I've just checked that the handling of SLURM_CPUS_PER_TASK 
hasn't changed in OpenMPI's ras_slurm_module.c.


> What you probably want to do is what Gilles advised. However, Slurm 16.05 
> only supports PMIx v1,

Actually, --mpi=list doesn't show any PMIx at all.

Well, I'll probably pass to newer versions of Slurm AND OpenMPI.

Thanks a lot to both of you
        Lothar
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to