Seems rather odd - since this is managed by Moab, you shouldn't be seeing SLURM 
envars at all. What you should see are PBS_* envars, including a PBS_NODEFILE 
that actually contains the allocation.


On Feb 12, 2014, at 4:42 AM, Adrian Reber <adr...@lisas.de> wrote:

> I tried the nightly snapshot (openmpi-1.7.5a1r30692.tar.gz) on a system
> with slurm and moab. I requested an interactive session using:
> 
> msub -I -l nodes=3:ppn=8
> 
> and started a simple test case which fails:
> 
> $ mpirun -np 2 ./mpi-test 1
> --------------------------------------------------------------------------
> There are not enough slots available in the system to satisfy the 2 slots 
> that were requested by the application:
>  ./mpi-test
> 
> Either request fewer slots for your application, or make more slots available
> for use.
> --------------------------------------------------------------------------
> srun: error: xxxx108: task 1: Exited with exit code 1
> srun: Terminating job step 131823.4
> srun: error: xxxx107: task 0: Exited with exit code 1
> srun: Job step aborted
> slurmd[xxxx108]: *** STEP 131823.4 KILLED AT 2014-02-12T13:30:32 WITH SIGNAL 
> 9 ***
> 
> 
> requesting only one core works:
> 
> $ mpirun  ./mpi-test 1
> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 0 on xxxx106 out of 1: 0.000000
> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 0 on xxxx106 out of 1: 0.000000
> 
> 
> using openmpi-1.6.5 works with multiple cores:
> 
> $ mpirun -np 24 ./mpi-test 2
> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 0 on xxxx106 out of 24: 0.000000
> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 12 on xxxx106 out of 24: 12.000000
> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 11 on xxxx108 out of 24: 11.000000
> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 18 on xxxx106 out of 24: 18.000000
> 
> $ echo $SLURM_JOB_CPUS_PER_NODE 
> 8(x3)
> 
> I never used slurm before so this could also be a user error on my side.
> But as 1.6.5 works it seems something has changed and wanted to let
> you know in case it was not intentionally.
> 
>               Adrian
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel

Reply via email to