I have a compiled binary that will run with srun but not with mpirun.
The attempts to run with mpirun all result in failures to initialize.
I have tried this on one node, and on two nodes, with firewall turned
on and with it off.

Am I missing some command line option for mpirun?

OMPI built from this configure command

  $ ./configure --prefix=/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b
--mandir=/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/share/man
--with-pmix=/opt/pmix/2.0.2 --with-libevent=external
--with-hwloc=external --with-slurm --disable-dlopen CC=gcc CXX=g++
FC=gfortran

All tests from `make check` passed, see below.

[bennet@cavium-hpc ~]$ mpicc --show
gcc -I/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/include -pthread
-L/opt/pmix/2.0.2/lib -Wl,-rpath -Wl,/opt/pmix/2.0.2/lib -Wl,-rpath
-Wl,/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/lib
-Wl,--enable-new-dtags
-L/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/lib -lmpi

The test_mpi was compiled with

$ gcc -o test_mpi test_mpi.c -lm

This is the runtime library path

[bennet@cavium-hpc ~]$ echo $LD_LIBRARY_PATH
/opt/slurm/lib64:/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/lib:/opt/arm/gcc-7.1.0_Generic-AArch64_RHEL-7_aarch64-linux/lib64:/opt/arm/gcc-7.1.0_Generic-AArch64_RHEL-7_aarch64-linux/lib:/opt/slurm/lib64:/opt/pmix/2.0.2/lib:/sw/arcts/centos7/hpc-utils/lib


These commands are given in exact sequence in which they were entered
at a console.

[bennet@cavium-hpc ~]$ salloc -N 1 --ntasks-per-node=24
salloc: Pending job allocation 156
salloc: job 156 queued and waiting for resources
salloc: job 156 has been allocated resources
salloc: Granted job allocation 156

[bennet@cavium-hpc ~]$ mpirun ./test_mpi
--------------------------------------------------------------------------
An ORTE daemon has unexpectedly failed after launch and before
communicating back to mpirun. This could be caused by a number
of factors, including an inability to create a connection back
to mpirun due to a lack of common network interfaces and/or no
route found between them. Please check network connectivity
(including firewalls and network routing requirements).
--------------------------------------------------------------------------

[bennet@cavium-hpc ~]$ srun ./test_mpi
The sum = 0.866386
Elapsed time is:  5.425439
The sum = 0.866386
Elapsed time is:  5.427427
The sum = 0.866386
Elapsed time is:  5.422579
The sum = 0.866386
Elapsed time is:  5.424168
The sum = 0.866386
Elapsed time is:  5.423951
The sum = 0.866386
Elapsed time is:  5.422414
The sum = 0.866386
Elapsed time is:  5.427156
The sum = 0.866386
Elapsed time is:  5.424834
The sum = 0.866386
Elapsed time is:  5.425103
The sum = 0.866386
Elapsed time is:  5.422415
The sum = 0.866386
Elapsed time is:  5.422948
Total time is:  59.668622

Thanks,    -- bennet


make check results
----------------------------------------------

make  check-TESTS
make[3]: Entering directory `/tmp/build/openmpi-3.1.0/ompi/debuggers'
make[4]: Entering directory `/tmp/build/openmpi-3.1.0/ompi/debuggers'
PASS: predefined_gap_test
PASS: predefined_pad_test
SKIP: dlopen_test
============================================================================
Testsuite summary for Open MPI 3.1.0
============================================================================
# TOTAL: 3
# PASS:  2
# SKIP:  1
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0
============================================================================
[ elided ]
PASS: atomic_cmpset_noinline
    - 5 threads: Passed
PASS: atomic_cmpset_noinline
    - 8 threads: Passed
============================================================================
Testsuite summary for Open MPI 3.1.0
============================================================================
# TOTAL: 8
# PASS:  8
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0
============================================================================
[ elided ]
make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/class'
PASS: ompi_rb_tree
PASS: opal_bitmap
PASS: opal_hash_table
PASS: opal_proc_table
PASS: opal_tree
PASS: opal_list
PASS: opal_value_array
PASS: opal_pointer_array
PASS: opal_lifo
PASS: opal_fifo
============================================================================
Testsuite summary for Open MPI 3.1.0
============================================================================
# TOTAL: 10
# PASS:  10
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0
============================================================================
[ elided ]
make  opal_thread opal_condition
make[3]: Entering directory `/tmp/build/openmpi-3.1.0/test/threads'
  CC       opal_thread.o
  CCLD     opal_thread
  CC       opal_condition.o
  CCLD     opal_condition
make[3]: Leaving directory `/tmp/build/openmpi-3.1.0/test/threads'
make  check-TESTS
make[3]: Entering directory `/tmp/build/openmpi-3.1.0/test/threads'
make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/threads'
============================================================================
Testsuite summary for Open MPI 3.1.0
============================================================================
# TOTAL: 0
# PASS:  0
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0
============================================================================
[ elided ]
make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/datatype'
PASS: opal_datatype_test
PASS: unpack_hetero
PASS: checksum
PASS: position
PASS: position_noncontig
PASS: ddt_test
PASS: ddt_raw
PASS: unpack_ooo
PASS: ddt_pack
PASS: external32
============================================================================
Testsuite summary for Open MPI 3.1.0
============================================================================
# TOTAL: 10
# PASS:  10
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0
============================================================================
[ elided ]
make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/util'
PASS: opal_bit_ops
PASS: opal_path_nfs
PASS: bipartite_graph
============================================================================
Testsuite summary for Open MPI 3.1.0
============================================================================
# TOTAL: 3
# PASS:  3
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0
============================================================================
[ elided ]
make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/dss'
PASS: dss_buffer
PASS: dss_cmp
PASS: dss_payload
PASS: dss_print
============================================================================
Testsuite summary for Open MPI 3.1.0
============================================================================
# TOTAL: 4
# PASS:  4
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0
============================================================================
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to