Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure
Hi, Last time, I installed the latest package 4.x.x, mpirun didn’t run across machines. I had to set the option --enable-orterun-prefix-by-default and compile. I’ll manually run it later and let you know. The problem disappeared when I turn off the nfs server. I was able to install it again. Thanks. From: users On Behalf Of Gilles Gouaillardet via users Sent: Monday, 22 April 2024 12:29 PM To: Open MPI Users Cc: Gilles Gouaillardet Subject: Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure Hi, Is there any reason why you do not build the latest 5.0.2 package? Anyway, the issue could be related to an unknown filesystem. Do you get a meaningful error if you manually run /.../test/util/opal_path_nfs? If not, can you share the output of mount | cut -f3,5 -d' ' Cheers, Gilles On Sun, Apr 21, 2024 at 10:04 PM Kook Jin Noh via users mailto:users@lists.open-mpi.org>> wrote: After updating prrte and I tried to install openmpi and the problem arises again. Please help. Thanks. From: users mailto:users-boun...@lists.open-mpi.org>> On Behalf Of Kook Jin Noh via users Sent: Sunday, 21 April 2024 9:27 PM To: Open MPI Users mailto:users@lists.open-mpi.org>> Cc: Kook Jin Noh mailto:vorl...@outlook.com>> Subject: Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure I makepkg -s again and it installed. The problem solved but would be nice if someone can explain the situation. Thanks. From: users mailto:users-boun...@lists.open-mpi.org>> On Behalf Of Kook Jin Noh via users Sent: Sunday, 21 April 2024 8:18 PM To: users@lists.open-mpi.org<mailto:users@lists.open-mpi.org> Cc: Kook Jin Noh mailto:vorl...@outlook.com>> Subject: [OMPI users] Another OpenMPI 5.0.1. Installation Failure Hi, I’m installting OpenMPI on Archlinux 6.7.0. Everything goes well till: Testsuite summary for Open MPI 5.0.1 # TOTAL: 13 # PASS: 13 # SKIP: 0 # XFAIL: 0 # FAIL: 0 # XPASS: 0 # ERROR: 0 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' Making check in util make[2]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make opal_bit_ops opal_path_nfs bipartite_graph opal_sha256 make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' CCLD opal_bit_ops CCLD opal_path_nfs CCLD bipartite_graph CCLD opal_sha256 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make check-TESTS make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[4]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' PASS: opal_bit_ops FAIL: opal_path_nfs PASS: bipartite_graph PASS: opal_sha256 Testsuite summary for Open MPI 5.0.1 # TOTAL: 4 # PASS: 3 # SKIP: 0 # XFAIL: 0 # FAIL: 1 # XPASS: 0 # ERROR: 0 See test/util/test-suite.log Please report to https://www.open-mpi.org/community/help/ make[4]: *** [Makefile:1838: test-suite.log] Error 1 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[3]: *** [Makefile:1946: check-TESTS] Error 2 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[2]: *** [Makefile:2040: check-am] Error 2 make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[1]: *** [Makefile:1416: check-recursive] Error 1 make[1]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test' make: *** [Makefile:1533: check-recursive] Error 1 make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1' ==> ERROR: A failure occurred in check(). Aborting... [vorlket@miniserver openmpi-ucx]$ Could you please help me understanding what’s going on? Thanks.
Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure
After updating prrte and I tried to install openmpi and the problem arises again. Please help. Thanks. From: users On Behalf Of Kook Jin Noh via users Sent: Sunday, 21 April 2024 9:27 PM To: Open MPI Users Cc: Kook Jin Noh Subject: Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure I makepkg -s again and it installed. The problem solved but would be nice if someone can explain the situation. Thanks. From: users mailto:users-boun...@lists.open-mpi.org>> On Behalf Of Kook Jin Noh via users Sent: Sunday, 21 April 2024 8:18 PM To: users@lists.open-mpi.org<mailto:users@lists.open-mpi.org> Cc: Kook Jin Noh mailto:vorl...@outlook.com>> Subject: [OMPI users] Another OpenMPI 5.0.1. Installation Failure Hi, I'm installting OpenMPI on Archlinux 6.7.0. Everything goes well till: Testsuite summary for Open MPI 5.0.1 # TOTAL: 13 # PASS: 13 # SKIP: 0 # XFAIL: 0 # FAIL: 0 # XPASS: 0 # ERROR: 0 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' Making check in util make[2]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make opal_bit_ops opal_path_nfs bipartite_graph opal_sha256 make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' CCLD opal_bit_ops CCLD opal_path_nfs CCLD bipartite_graph CCLD opal_sha256 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make check-TESTS make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[4]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' PASS: opal_bit_ops FAIL: opal_path_nfs PASS: bipartite_graph PASS: opal_sha256 Testsuite summary for Open MPI 5.0.1 # TOTAL: 4 # PASS: 3 # SKIP: 0 # XFAIL: 0 # FAIL: 1 # XPASS: 0 # ERROR: 0 See test/util/test-suite.log Please report to https://www.open-mpi.org/community/help/ make[4]: *** [Makefile:1838: test-suite.log] Error 1 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[3]: *** [Makefile:1946: check-TESTS] Error 2 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[2]: *** [Makefile:2040: check-am] Error 2 make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[1]: *** [Makefile:1416: check-recursive] Error 1 make[1]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test' make: *** [Makefile:1533: check-recursive] Error 1 make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1' ==> ERROR: A failure occurred in check(). Aborting... [vorlket@miniserver openmpi-ucx]$ Could you please help me understanding what's going on? Thanks.
Re: [OMPI users] PRTE mismatch
pacman -R prrte, pacman -Syu prrte solved it. From: users On Behalf Of Kook Jin Noh via users Sent: Sunday, 21 April 2024 9:25 PM To: users@lists.open-mpi.org Cc: Kook Jin Noh Subject: [OMPI users] PRTE mismatch Hi, while trying to run a MPI program, I received the following: [vorlket@server ~]$ mpirun -host server:4,midiserver:4,miniserver:1 -np 9 /home/vorlket/sharedfolder/mpi-prime -- PRTE detected a mismatch in versions between two processes. This typically means that you executed "mpirun" (or "mpiexec") from one version of PRTE on on node, but your default path on one of the other nodes upon which you launched found a different version of Open MPI. PRTE only supports running exactly the same version between all processes in a single job. This will almost certainly cause unpredictable behavior, and may end up aborting your job. Local host: server Local process name: [prterun-server-1089@0,0] Local PRTE version: 3.0.3 Peer host: Unknown Peer process name: [prterun-server-1089@0,1] Peer PRTE version: 3.0.5 -- -- PRTE has lost communication with a remote daemon. HNP daemon : [prterun-server-1089@0,0] on node server Remote daemon: [prterun-server-1089@0,1] on node midiserver This is usually due to either a failure of the TCP network connection to the node, or possibly an internal failure of the daemon itself. We cannot recover from this failure, and therefore will terminate the job. Could you please tell me what I can do to either have the same PRTE version across the servers or something else to make the program run? Thanks.
Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure
I makepkg -s again and it installed. The problem solved but would be nice if someone can explain the situation. Thanks. From: users On Behalf Of Kook Jin Noh via users Sent: Sunday, 21 April 2024 8:18 PM To: users@lists.open-mpi.org Cc: Kook Jin Noh Subject: [OMPI users] Another OpenMPI 5.0.1. Installation Failure Hi, I'm installting OpenMPI on Archlinux 6.7.0. Everything goes well till: Testsuite summary for Open MPI 5.0.1 # TOTAL: 13 # PASS: 13 # SKIP: 0 # XFAIL: 0 # FAIL: 0 # XPASS: 0 # ERROR: 0 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' Making check in util make[2]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make opal_bit_ops opal_path_nfs bipartite_graph opal_sha256 make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' CCLD opal_bit_ops CCLD opal_path_nfs CCLD bipartite_graph CCLD opal_sha256 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make check-TESTS make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[4]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' PASS: opal_bit_ops FAIL: opal_path_nfs PASS: bipartite_graph PASS: opal_sha256 Testsuite summary for Open MPI 5.0.1 # TOTAL: 4 # PASS: 3 # SKIP: 0 # XFAIL: 0 # FAIL: 1 # XPASS: 0 # ERROR: 0 See test/util/test-suite.log Please report to https://www.open-mpi.org/community/help/ make[4]: *** [Makefile:1838: test-suite.log] Error 1 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[3]: *** [Makefile:1946: check-TESTS] Error 2 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[2]: *** [Makefile:2040: check-am] Error 2 make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[1]: *** [Makefile:1416: check-recursive] Error 1 make[1]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test' make: *** [Makefile:1533: check-recursive] Error 1 make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1' ==> ERROR: A failure occurred in check(). Aborting... [vorlket@miniserver openmpi-ucx]$ Could you please help me understanding what's going on? Thanks.
[OMPI users] PRTE mismatch
Hi, while trying to run a MPI program, I received the following: [vorlket@server ~]$ mpirun -host server:4,midiserver:4,miniserver:1 -np 9 /home/vorlket/sharedfolder/mpi-prime -- PRTE detected a mismatch in versions between two processes. This typically means that you executed "mpirun" (or "mpiexec") from one version of PRTE on on node, but your default path on one of the other nodes upon which you launched found a different version of Open MPI. PRTE only supports running exactly the same version between all processes in a single job. This will almost certainly cause unpredictable behavior, and may end up aborting your job. Local host: server Local process name: [prterun-server-1089@0,0] Local PRTE version: 3.0.3 Peer host: Unknown Peer process name: [prterun-server-1089@0,1] Peer PRTE version: 3.0.5 -- -- PRTE has lost communication with a remote daemon. HNP daemon : [prterun-server-1089@0,0] on node server Remote daemon: [prterun-server-1089@0,1] on node midiserver This is usually due to either a failure of the TCP network connection to the node, or possibly an internal failure of the daemon itself. We cannot recover from this failure, and therefore will terminate the job. Could you please tell me what I can do to either have the same PRTE version across the servers or something else to make the program run? Thanks.
[OMPI users] Another OpenMPI 5.0.1. Installation Failure
Hi, I'm installting OpenMPI on Archlinux 6.7.0. Everything goes well till: Testsuite summary for Open MPI 5.0.1 # TOTAL: 13 # PASS: 13 # SKIP: 0 # XFAIL: 0 # FAIL: 0 # XPASS: 0 # ERROR: 0 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' Making check in util make[2]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make opal_bit_ops opal_path_nfs bipartite_graph opal_sha256 make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' CCLD opal_bit_ops CCLD opal_path_nfs CCLD bipartite_graph CCLD opal_sha256 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make check-TESTS make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[4]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' PASS: opal_bit_ops FAIL: opal_path_nfs PASS: bipartite_graph PASS: opal_sha256 Testsuite summary for Open MPI 5.0.1 # TOTAL: 4 # PASS: 3 # SKIP: 0 # XFAIL: 0 # FAIL: 1 # XPASS: 0 # ERROR: 0 See test/util/test-suite.log Please report to https://www.open-mpi.org/community/help/ make[4]: *** [Makefile:1838: test-suite.log] Error 1 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[3]: *** [Makefile:1946: check-TESTS] Error 2 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[2]: *** [Makefile:2040: check-am] Error 2 make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util' make[1]: *** [Makefile:1416: check-recursive] Error 1 make[1]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test' make: *** [Makefile:1533: check-recursive] Error 1 make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1' ==> ERROR: A failure occurred in check(). Aborting... [vorlket@miniserver openmpi-ucx]$ Could you please help me understanding what's going on? Thanks.
[OMPI users] OpenMPI 5.0.1 Installation Failure
Hi, I'm installing OpenMPI 5.0.1 on Archlinux 6.7.1. Everything goes well till: Making check in datatype make[2]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make opal_datatype_test unpack_hetero checksum position position_noncontig ddt_test ddt_raw ddt_raw2 unpack_ooo ddt_pack external32 large_data partial to_self reduce_local make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' CCLD opal_datatype_test CCLD unpack_hetero CCLD checksum CCLD position CCLD position_noncontig CCLD ddt_test CCLD ddt_raw CCLD ddt_raw2 CCLD unpack_ooo CCLD ddt_pack CCLD external32 CCLD large_data CCLD partial CCLD to_self CCLD reduce_local make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make check-TESTS make[3]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[4]: Entering directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' ../../config/test-driver: line 112: 1380808 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: opal_datatype_test PASS: unpack_hetero ../../config/test-driver: line 112: 1380857 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: checksum ../../config/test-driver: line 112: 1380884 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: position ../../config/test-driver: line 112: 1380916 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: position_noncontig ../../config/test-driver: line 112: 1380944 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: ddt_test ../../config/test-driver: line 112: 1380975 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: ddt_raw PASS: ddt_raw2 PASS: unpack_ooo ../../config/test-driver: line 112: 1381044 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: ddt_pack ../../config/test-driver: line 112: 1381070 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: external32 PASS: large_data ../../config/test-driver: line 112: 1381120 Segmentation fault (core dumped) "$@" >> "$log_file" 2>&1 FAIL: partial Testsuite summary for Open MPI 5.0.1 # TOTAL: 13 # PASS: 4 # SKIP: 0 # XFAIL: 0 # FAIL: 9 # XPASS: 0 # ERROR: 0 See test/datatype/test-suite.log Please report to https://www.open-mpi.org/community/help/ make[4]: *** [Makefile:2012: test-suite.log] Error 1 make[4]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[3]: *** [Makefile:2120: check-TESTS] Error 2 make[3]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[2]: *** [Makefile:2277: check-am] Error 2 make[2]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype' make[1]: *** [Makefile:1416: check-recursive] Error 1 make[1]: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test' make: *** [Makefile:1533: check-recursive] Error 1 make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1' ==> ERROR: A failure occurred in check(). Aborting...
[OMPI users] ORTE was unable to reliably start one or more daemons.
Hi, I tried to run the following and got an error (the compiled program (with mpicc) on the sharedfolder runs ok on each individual machine): [vorlket@server ~]$ mpirun -host server,miniserver -np 4 /home/vorlket/sharedfolder/mpi-prime -- ORTE was unable to reliably start one or more daemons. This usually is caused by: * not finding the required libraries and/or binaries on one or more nodes. Please check your PATH and LD_LIBRARY_PATH settings, or configure OMPI with --enable-orterun-prefix-by-default * lack of authority to execute on one or more specified nodes. Please verify your allocation and authorities. * the inability to write startup files into /tmp (--tmpdir/orte_tmpdir_base). Please check with your sys admin to determine the correct location to use. * compilation of the orted with dynamic libraries when static are required (e.g., on Cray). Please check your configure cmd line and consider using one of the contrib/platform definitions for your system type. * an inability to create a connection back to mpirun due to a lack of common network interfaces and/or no route found between them. Please check network connectivity (including firewalls and network routing requirements). -- Hi I have the following installations on archlinux: server: sudo pacman -S openmpi cd ~/.ssh ssh-keygen -t rsa cp id_rsa.pub authorized_keys chmod 600 authorized_key sssh-copy-id miniserver sudo vi /etc/ssh/sshd_config: Add the following lines PubkeyAuthentication yes RSAAuthentication yes sudo pacman -S ntp sudo vi /etc/ntp.conf: Add the following lines server 0.arch.pool.ntp.org iburst server 1.arch.pool.ntp.org iburst server 2.arch.pool.ntp.org iburst server 3.arch.pool.ntp.org iburst tos orphan 15 logfile /var/log/ntp.log sudo ntpd -u ntp:ntp sudo systemctl start ntpd.service sudo systemctl enable ntpd.service sudo pacman -S nfs-utils mkdir ~/sharedfolder chmod 777 ~/sharedfolder sudo vi /etc/exports: /home/vorlket/sharedfolder miniserver(rw,sync) sudo exportfs -arv sudo systemctl start nfs-server.service sudo systemctl enable nfs-server.service miniserver: sudo pacman -S openmpi cd ~/.ssh ssh-keygen -t rsa cp id_rsa.pub authorized_keys chmod 600 authorized_key sssh-copy-id server sudo vi /etc/ssh/sshd_config: PubkeyAuthentication yes RSAAuthentication yes sudo pacman -S ntp sudo vi /etc/ntp.conf: Add the following lines server 192.168.1.3 sudo systemctl start ntpd.service sudo systemctl enable ntpd.service sudo pacman -S nfs-utils mkdir ~/sharedfolder chmod 777 ~/sharedfolder sudo vi /etc/fstab: server:/home/vorlket/sharedfolder /home/vorlket/sharedfolder nfs _netdev,noauto,x-systemd.automount,x-systemd.mount-timeout=10,timeo=14,x-systemd.idle-timeout=1min 0 0 sudo systemctl daemon-reload sudo systemctl restart remote-fs.target I tried to run the following and got an error: [vorlket@server ~]$ mpirun -host server,miniserver -np 4 /home/vorlket/sharedfolder/mpi-prime -- ORTE was unable to reliably start one or more daemons. This usually is caused by: * not finding the required libraries and/or binaries on one or more nodes. Please check your PATH and LD_LIBRARY_PATH settings, or configure OMPI with --enable-orterun-prefix-by-default * lack of authority to execute on one or more specified nodes. Please verify your allocation and authorities. * the inability to write startup files into /tmp (--tmpdir/orte_tmpdir_base). Please check with your sys admin to determine the correct location to use. * compilation of the orted with dynamic libraries when static are required (e.g., on Cray). Please check your configure cmd line and consider using one of the contrib/platform definitions for your system type. * an inability to create a connection back to mpirun due to a lack of common network interfaces and/or no route found between them. Please check network connectivity (including firewalls and network routing requirements). -- mpi-prime.c: [vorlket@server ~]$ cat /home/vorlket/mpiprac/mpi-prime.c # include # include # include # include # include int main ( int argc, char *argv[] ); int prime_number ( int n, int id, int p ); void timestamp ( ); /**/ int main ( int argc, char *argv[] ) /**/ /* Purpose: MAIN is the main program for PRIME_MPI. Discussion: This program calls a version of PRIME_NUMBER that includes MPI calls for parallel processing. Licensing: