Hi Sigmar, I observed this problem yesterday myself and should have a fix in to master later today.
Howard Am Fr., 4. Jan. 2019 um 05:30 Uhr schrieb Siegmar Gross < siegmar.gr...@informatik.hs-fulda.de>: > Hi, > > I've installed (tried to install) openmpi-master-201901030305-ee26ed9 on > my "SUSE Linux Enterprise Server 12.3 (x86_64)" with gcc-7.3.0, > icc-19.0.1.144 > pgcc-18.4-0, and Sun C 5.15 (Oracle Developer Studio 12.6). Unfortunately, > I > still cannot build it with Sun C and I get a segmentation fault for one of > my small programs for the other compilers. > > I get the following error for Sun C that I reported some time ago. > https://www.mail-archive.com/users@lists.open-mpi.org/msg32816.html > > > The program runs as expected if I only use my local machine "loki" and it > breaks if I add a remote machine (even if I only use the remote machine > without "loki"). > > loki hello_1 114 ompi_info | grep -e "Open MPI repo revision" -e"Configure > command line" > Open MPI repo revision: v2.x-dev-6601-gee26ed9 > Configure command line: '--prefix=/usr/local/openmpi-master_64_gcc' > '--libdir=/usr/local/openmpi-master_64_gcc/lib64' > '--with-jdk-bindir=/usr/local/jdk-11/bin' > '--with-jdk-headers=/usr/local/jdk-11/include' > 'JAVA_HOME=/usr/local/jdk-11' > 'LDFLAGS=-m64 -L/usr/local/cuda/lib64' 'CC=gcc' 'CXX=g++' 'FC=gfortran' > 'CFLAGS=-m64 -I/usr/local/cuda/include' 'CXXFLAGS=-m64 > -I/usr/local/cuda/include' 'FCFLAGS=-m64' 'CPP=cpp > -I/usr/local/cuda/include' > 'CXXCPP=cpp -I/usr/local/cuda/include' '--enable-mpi-cxx' > '--enable-cxx-exceptions' '--enable-mpi-java' > '--with-cuda=/usr/local/cuda' > '--with-valgrind=/usr/local/valgrind' '--with-hwloc=internal' > '--without-verbs' > '--with-wrapper-cflags=-std=c11 -m64' '--with-wrapper-cxxflags=-m64' > '--with-wrapper-fcflags=-m64' '--enable-debug' > > > loki hello_1 115 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi > Process 0 of 4 running on loki > Process 1 of 4 running on loki > Process 2 of 4 running on nfs2 > Process 3 of 4 running on nfs2 > > Now 3 slave tasks are sending greetings. > > Greetings from task 1: > message type: 3 > msg length: 132 characters > ... (complete output of my program) > > [nfs2:01336] *** Process received signal *** > [nfs2:01336] Signal: Segmentation fault (11) > [nfs2:01336] Signal code: Address not mapped (1) > [nfs2:01336] Failing at address: 0x7feea4849268 > [nfs2:01336] [ 0] /lib64/libpthread.so.0(+0x10c10)[0x7feeacbbec10] > [nfs2:01336] [ 1] > > /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0(+0x7cd34)[0x7feeadd94d34] > [nfs2:01336] [ 2] > > /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0(+0x78673)[0x7feeadd90673] > [nfs2:01336] [ 3] > > /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0(+0x7ac2c)[0x7feeadd92c2c] > [nfs2:01336] [ 4] > > /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0(opal_finalize_cleanup_domain+0x3e)[0x7feeadd56507] > [nfs2:01336] [ 5] > > /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0(opal_finalize_util+0x56)[0x7feeadd56667] > [nfs2:01336] [ 6] > > /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0(opal_finalize+0xd3)[0x7feeadd567de] > [nfs2:01336] [ 7] > > /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0(orte_finalize+0x1ba)[0x7feeae09d7ea] > [nfs2:01336] [ 8] > > /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0(orte_daemon+0x3ddd)[0x7feeae0cf55d] > [nfs2:01336] [ 9] orted[0x40086d] > [nfs2:01336] [10] /lib64/libc.so.6(__libc_start_main+0xf5)[0x7feeac829725] > [nfs2:01336] [11] orted[0x400739] > [nfs2:01336] *** End of error message *** > Segmentation fault (core dumped) > loki hello_1 116 > > > I would be grateful, if somebody can fix the problem. Do you need anything > else? Thank you very much for any help in advance. > > > Kind regards > > Siegmar > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users