Jeff and I talked and believe the patch in https://github.com/open-mpi/ompi/pull/5836 should fix the problem.
> On Oct 2, 2018, at 2:50 PM, Jeff Squyres (jsquyres) via users > <users@lists.open-mpi.org> wrote: > > (Ralph sent me Siegmar's pmix config.log, which Siegmar sent to him off-list) > > It looks like Siegmar passed --with-hwloc=internal. > > Open MPI's configure understood this and did the appropriate things. > PMIX's configure didn't. > > I think we need to add an adjustment into the PMIx configure.m4 in OMPI... > > >> On Oct 2, 2018, at 5:25 PM, Ralph H Castain <r...@open-mpi.org> wrote: >> >> Hi Siegmar >> >> I honestly have no idea - for some reason, the PMIx component isn’t seeing >> the internal hwloc code in your environment. >> >> Jeff, Brice - any ideas? >> >> >>> On Oct 2, 2018, at 1:18 PM, Siegmar Gross >>> <siegmar.gr...@informatik.hs-fulda.de> wrote: >>> >>> Hi Ralph, >>> >>> how can I confirm that HWLOC built? Some hwloc files are available >>> in the built directory. >>> >>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 111 find . >>> -name '*hwloc*' >>> ./opal/mca/btl/usnic/.deps/btl_usnic_hwloc.Plo >>> ./opal/mca/hwloc >>> ./opal/mca/hwloc/external/.deps/hwloc_external_component.Plo >>> ./opal/mca/hwloc/base/hwloc_base_frame.lo >>> ./opal/mca/hwloc/base/.deps/hwloc_base_dt.Plo >>> ./opal/mca/hwloc/base/.deps/hwloc_base_maffinity.Plo >>> ./opal/mca/hwloc/base/.deps/hwloc_base_frame.Plo >>> ./opal/mca/hwloc/base/.deps/hwloc_base_util.Plo >>> ./opal/mca/hwloc/base/hwloc_base_dt.lo >>> ./opal/mca/hwloc/base/hwloc_base_util.lo >>> ./opal/mca/hwloc/base/hwloc_base_maffinity.lo >>> ./opal/mca/hwloc/base/.libs/hwloc_base_util.o >>> ./opal/mca/hwloc/base/.libs/hwloc_base_dt.o >>> ./opal/mca/hwloc/base/.libs/hwloc_base_maffinity.o >>> ./opal/mca/hwloc/base/.libs/hwloc_base_frame.o >>> ./opal/mca/hwloc/.libs/libmca_hwloc.la >>> ./opal/mca/hwloc/.libs/libmca_hwloc.a >>> ./opal/mca/hwloc/libmca_hwloc.la >>> ./opal/mca/hwloc/hwloc201 >>> ./opal/mca/hwloc/hwloc201/.deps/hwloc201_component.Plo >>> ./opal/mca/hwloc/hwloc201/hwloc201_component.lo >>> ./opal/mca/hwloc/hwloc201/hwloc >>> ./opal/mca/hwloc/hwloc201/hwloc/include/hwloc >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/libhwloc_embedded.la >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_pci_la-topology-pci.Plo >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_gl_la-topology-gl.Plo >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_cuda_la-topology-cuda.Plo >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_xml_libxml_la-topology-xml-libxml.Plo >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_opencl_la-topology-opencl.Plo >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_nvml_la-topology-nvml.Plo >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.la >>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.a >>> ./opal/mca/hwloc/hwloc201/.libs/hwloc201_component.o >>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.la >>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.a >>> ./opal/mca/hwloc/hwloc201/libmca_hwloc_hwloc201.la >>> ./orte/mca/rtc/hwloc >>> ./orte/mca/rtc/hwloc/rtc_hwloc.lo >>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc.Plo >>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc_component.Plo >>> ./orte/mca/rtc/hwloc/mca_rtc_hwloc.la >>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.so >>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.la >>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc.o >>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc_component.o >>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.soT >>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.lai >>> ./orte/mca/rtc/hwloc/rtc_hwloc_component.lo >>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 112 >>> >>> And some files are available in the install directory. >>> >>> loki openmpi-master_64_gcc 116 find . -name '*hwloc*' >>> ./share/openmpi/help-orte-rtc-hwloc.txt >>> ./share/openmpi/help-opal-hwloc-base.txt >>> ./lib64/openmpi/mca_rtc_hwloc.so >>> ./lib64/openmpi/mca_rtc_hwloc.la >>> loki openmpi-master_64_gcc 117 >>> >>> I don't see any unavailable libraries so that the only available >>> hwloc library should work. >>> >>> loki openmpi 126 ldd -v mca_rtc_hwloc.so >>> linux-vdso.so.1 (0x00007ffd2df5b000) >>> libopen-rte.so.0 => >>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0 (0x00007f082b7fb000) >>> libopen-pal.so.0 => >>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0 (0x00007f082b493000) >>> libdl.so.2 => /lib64/libdl.so.2 (0x00007f082b28f000) >>> libudev.so.1 => /usr/lib64/libudev.so.1 (0x00007f082b06e000) >>> libpciaccess.so.0 => /usr/lib64/libpciaccess.so.0 (0x00007f082ae64000) >>> librt.so.1 => /lib64/librt.so.1 (0x00007f082ac5c000) >>> libm.so.6 => /lib64/libm.so.6 (0x00007f082a95f000) >>> libutil.so.1 => /lib64/libutil.so.1 (0x00007f082a75c000) >>> libz.so.1 => /lib64/libz.so.1 (0x00007f082a546000) >>> libpthread.so.0 => /lib64/libpthread.so.0 (0x00007f082a329000) >>> libc.so.6 => /lib64/libc.so.6 (0x00007f0829f84000) >>> libgcc_s.so.1 => /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>> (0x00007f0829d6c000) >>> /lib64/ld-linux-x86-64.so.2 (0x00007f082bd24000) >>> libselinux.so.1 => /lib64/libselinux.so.1 (0x00007f0829b46000) >>> libcap.so.2 => /lib64/libcap.so.2 (0x00007f0829941000) >>> libresolv.so.2 => /lib64/libresolv.so.2 (0x00007f082972a000) >>> libpcre.so.1 => /usr/lib64/libpcre.so.1 (0x00007f08294bb000) >>> >>> Version information: >>> ./mca_rtc_hwloc.so: >>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0: >>> libz.so.1 (ZLIB_1.2.0) => /lib64/libz.so.1 >>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0: >>> librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1 >>> libgcc_s.so.1 (GCC_3.0) => >>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>> libgcc_s.so.1 (GCC_3.3.1) => >>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>> libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2 >>> libutil.so.1 (GLIBC_2.2.5) => /lib64/libutil.so.1 >>> libudev.so.1 (LIBUDEV_183) => /usr/lib64/libudev.so.1 >>> libm.so.6 (GLIBC_2.2.5) => /lib64/libm.so.6 >>> libpthread.so.0 (GLIBC_2.3.4) => /lib64/libpthread.so.0 >>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>> libc.so.6 (GLIBC_2.6) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /lib64/libdl.so.2: >>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>> /lib64/ld-linux-x86-64.so.2 >>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /usr/lib64/libudev.so.1: >>> librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1 >>> ld-linux-x86-64.so.2 (GLIBC_2.3) => /lib64/ld-linux-x86-64.so.2 >>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.9) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.16) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /usr/lib64/libpciaccess.so.0: >>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>> /lib64/librt.so.1: >>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>> libpthread.so.0 (GLIBC_PRIVATE) => /lib64/libpthread.so.0 >>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>> /lib64/libm.so.6: >>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /lib64/libutil.so.1: >>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /lib64/libz.so.1: >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>> /lib64/libpthread.so.0: >>> ld-linux-x86-64.so.2 (GLIBC_2.2.5) => >>> /lib64/ld-linux-x86-64.so.2 >>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>> /lib64/ld-linux-x86-64.so.2 >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>> /lib64/libc.so.6: >>> ld-linux-x86-64.so.2 (GLIBC_2.3) => /lib64/ld-linux-x86-64.so.2 >>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>> /lib64/ld-linux-x86-64.so.2 >>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1: >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /lib64/libselinux.so.1: >>> libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2 >>> ld-linux-x86-64.so.2 (GLIBC_2.3) => /lib64/ld-linux-x86-64.so.2 >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>> /lib64/libcap.so.2: >>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> /lib64/libresolv.so.2: >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>> /usr/lib64/libpcre.so.1: >>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>> loki openmpi 127 >>> >>> Hopefully that helps to find the problem. I will answer your emails >>> tommorrow if you need anything else. >>> >>> >>> Best regards >>> >>> Siegmar >>> >>> >>> Am 02.10.2018 um 19:48 schrieb Ralph H Castain: >>>> So the problem is here when configuring the internal PMIx code: >>>> configure:3383: === HWLOC >>>> configure:36189: checking for hwloc in >>>> configure:36201: result: Could not find internal/lib or internal/lib64 >>>> configure:36203: error: Can not continue >>>> Can you confirm that HWLOC built? I believe we require it, but perhaps >>>> something is different about this environment. >>>>> On Oct 2, 2018, at 6:36 AM, Ralph H Castain <r...@open-mpi.org> wrote: >>>>> >>>>> Looks like PMIx failed to build - can you send the config.log? >>>>> >>>>>> On Oct 2, 2018, at 12:00 AM, Siegmar Gross >>>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote: >>>>>> >>>>>> Hi, >>>>>> >>>>>> yesterday I've installed openmpi-v4.0.x-201809290241-a7e275c and >>>>>> openmpi-master-201805080348-b39bbfb on my "SUSE Linux Enterprise Server >>>>>> 12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and Portland >>>>>> Group pgcc 18.4-0. Unfortunately, I get the following error for all seven >>>>>> installed versions (Sun C couldn't built master as I mentioned in another >>>>>> email). >>>>>> >>>>>> >>>>>> loki hello_1 118 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi >>>>>> [loki:11423] [[45859,0],0] ORTE_ERROR_LOG: Not found in file >>>>>> ../../../../../openmpi-v4.0.x-201809290241-a7e275c/orte/mca/ess/hnp/ess_hnp_module.c >>>>>> at line 321 >>>>>> -------------------------------------------------------------------------- >>>>>> It looks like orte_init failed for some reason; your parallel process is >>>>>> likely to abort. There are many reasons that a parallel process can >>>>>> fail during orte_init; some of which are due to configuration or >>>>>> environment problems. This failure appears to be an internal failure; >>>>>> here's some additional information (which may only be relevant to an >>>>>> Open MPI developer): >>>>>> >>>>>> opal_pmix_base_select failed >>>>>> --> Returned value Not found (-13) instead of ORTE_SUCCESS >>>>>> -------------------------------------------------------------------------- >>>>>> loki hello_1 119 >>>>>> >>>>>> >>>>>> >>>>>> I would be grateful, if somebody can fix the problem. Do you need >>>>>> anything >>>>>> else? Thank you very much for any help in advance. >>>>>> >>>>>> >>>>>> Kind regards >>>>>> >>>>>> Siegmar >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org >>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.open-mpi.org >>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://lists.open-mpi.org/mailman/listinfo/users >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://lists.open-mpi.org/mailman/listinfo/users > > > -- > Jeff Squyres > jsquy...@cisco.com > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users _______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users