Simon -- You don't currently have another Open MPI installation in your PATH / LD_LIBRARY_PATH, do you?
I have seen dependency library loads cause "make check" to get confused, and instead of loading the libraries from the build tree, actually load some -- but not all -- of the required OMPI/ORTE/OPAL/etc. libraries from an installation tree. Hilarity ensues (to include symptoms such as running forever). Can you double check that you have no Open MPI libraries in your LD_LIBRARY_PATH before running "make check" on the build tree? > On Jun 30, 2018, at 3:18 PM, Hammond, Simon David via users > <users@lists.open-mpi.org> wrote: > > Nathan, > > Same issue with OpenMPI 3.1.1 on POWER9 with GCC 7.2.0 and CUDA9.2. > > S. > > -- > Si Hammond > Scalable Computer Architectures > Sandia National Laboratories, NM, USA > [Sent from remote connection, excuse typos] > > > On 6/16/18, 10:10 PM, "Nathan Hjelm" <hje...@me.com> wrote: > > Try the latest nightly tarball for v3.1.x. Should be fixed. > >> On Jun 16, 2018, at 5:48 PM, Hammond, Simon David via users >> <users@lists.open-mpi.org> wrote: >> >> The output from the test in question is: >> >> Single thread test. Time: 0 s 10182 us 10 nsec/poppush >> Atomics thread finished. Time: 0 s 169028 us 169 nsec/poppush >> <then runs forever> >> >> S. >> >> -- >> Si Hammond >> Scalable Computer Architectures >> Sandia National Laboratories, NM, USA >> [Sent from remote connection, excuse typos] >> >> >> On 6/16/18, 5:45 PM, "Hammond, Simon David" <sdha...@sandia.gov> wrote: >> >> Hi OpenMPI Team, >> >> We have recently updated an install of OpenMPI on POWER9 system >> (configuration details below). We migrated from OpenMPI 2.1 to OpenMPI 3.1. >> We seem to have a symptom where code than ran before is now locking up and >> making no progress, getting stuck in wait-all operations. While I think it's >> prudent for us to root cause this a little more, I have gone back and >> rebuilt MPI and re-run the "make check" tests. The opal_fifo test appears to >> hang forever. I am not sure if this is the cause of our issue but wanted to >> report that we are seeing this on our system. >> >> OpenMPI 3.1.0 Configuration: >> >> ./configure >> --prefix=/home/projects/ppc64le-pwr9-nvidia/openmpi/3.1.0-nomxm/gcc/7.2.0/cuda/9.2.88 >> --with-cuda=$CUDA_ROOT --enable-mpi-java --enable-java >> --with-lsf=/opt/lsf/10.1 >> --with-lsf-libdir=/opt/lsf/10.1/linux3.10-glibc2.17-ppc64le/lib --with-verbs >> >> GCC versions are 7.2.0, built by our team. CUDA is 9.2.88 from NVIDIA for >> POWER9 (standard download from their website). We enable IBM's JDK 8.0.0. >> RedHat: Red Hat Enterprise Linux Server release 7.5 (Maipo) >> >> Output: >> >> make[3]: Entering directory >> `/home/sdhammo/openmpi/openmpi-3.1.0/test/class' >> make[4]: Entering directory >> `/home/sdhammo/openmpi/openmpi-3.1.0/test/class' >> PASS: ompi_rb_tree >> PASS: opal_bitmap >> PASS: opal_hash_table >> PASS: opal_proc_table >> PASS: opal_tree >> PASS: opal_list >> PASS: opal_value_array >> PASS: opal_pointer_array >> PASS: opal_lifo >> <runs forever> >> >> Output from Top: >> >> 20 0 73280 4224 2560 S 800.0 0.0 17:22.94 lt-opal_fifo >> >> -- >> Si Hammond >> Scalable Computer Architectures >> Sandia National Laboratories, NM, USA >> [Sent from remote connection, excuse typos] >> >> >> >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://lists.open-mpi.org/mailman/listinfo/users > > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users -- Jeff Squyres jsquy...@cisco.com _______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users