Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

2024-04-22 Thread Kook Jin Noh via users
Hi,

Last time, I installed the latest package 4.x.x, mpirun didn’t run across 
machines. I had to set the option --enable-orterun-prefix-by-default and 
compile.

I’ll manually run it later and let you know. The problem disappeared when I 
turn off the nfs server. I was able to install it again.

Thanks.

From: users  On Behalf Of Gilles Gouaillardet 
via users
Sent: Monday, 22 April 2024 12:29 PM
To: Open MPI Users 
Cc: Gilles Gouaillardet 
Subject: Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

Hi,

Is there any reason why you do not build the latest 5.0.2 package?

Anyway, the issue could be related to an unknown filesystem.
Do you get a meaningful error if you manually run /.../test/util/opal_path_nfs?
If not, can you share the output of

mount | cut -f3,5 -d' '

Cheers,

Gilles

On Sun, Apr 21, 2024 at 10:04 PM Kook Jin Noh via users 
mailto:users@lists.open-mpi.org>> wrote:
After updating prrte and I tried to install openmpi and the problem arises 
again. Please help. Thanks.

From: users 
mailto:users-boun...@lists.open-mpi.org>> On 
Behalf Of Kook Jin Noh via users
Sent: Sunday, 21 April 2024 9:27 PM
To: Open MPI Users mailto:users@lists.open-mpi.org>>
Cc: Kook Jin Noh mailto:vorl...@outlook.com>>
Subject: Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

I makepkg -s again and it installed. The problem solved but would be nice if 
someone can explain the situation. Thanks.

From: users 
mailto:users-boun...@lists.open-mpi.org>> On 
Behalf Of Kook Jin Noh via users
Sent: Sunday, 21 April 2024 8:18 PM
To: users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>
Cc: Kook Jin Noh mailto:vorl...@outlook.com>>
Subject: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

Hi,

I’m installting OpenMPI on Archlinux 6.7.0. Everything goes well till:


Testsuite summary for Open MPI 5.0.1

# TOTAL: 13
# PASS:  13
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0

make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
Making check in util
make[2]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make  opal_bit_ops opal_path_nfs bipartite_graph opal_sha256
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
  CCLD opal_bit_ops
  CCLD opal_path_nfs
  CCLD bipartite_graph
  CCLD opal_sha256
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make  check-TESTS
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[4]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
PASS: opal_bit_ops
FAIL: opal_path_nfs
PASS: bipartite_graph
PASS: opal_sha256

Testsuite summary for Open MPI 5.0.1

# TOTAL: 4
# PASS:  3
# SKIP:  0
# XFAIL: 0
# FAIL:  1
# XPASS: 0
# ERROR: 0

See test/util/test-suite.log
Please report to https://www.open-mpi.org/community/help/

make[4]: *** [Makefile:1838: test-suite.log] Error 1
make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[3]: *** [Makefile:1946: check-TESTS] Error 2
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[2]: *** [Makefile:2040: check-am] Error 2
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[1]: *** [Makefile:1416: check-recursive] Error 1
make[1]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test'
make: *** [Makefile:1533: check-recursive] Error 1
make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1'
==> ERROR: A failure occurred in check().
Aborting...
[vorlket@miniserver openmpi-ucx]$

Could you please help me understanding what’s going on? Thanks.


Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

2024-04-21 Thread Kook Jin Noh via users
After updating prrte and I tried to install openmpi and the problem arises 
again. Please help. Thanks.

From: users  On Behalf Of Kook Jin Noh via 
users
Sent: Sunday, 21 April 2024 9:27 PM
To: Open MPI Users 
Cc: Kook Jin Noh 
Subject: Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

I makepkg -s again and it installed. The problem solved but would be nice if 
someone can explain the situation. Thanks.

From: users 
mailto:users-boun...@lists.open-mpi.org>> On 
Behalf Of Kook Jin Noh via users
Sent: Sunday, 21 April 2024 8:18 PM
To: users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>
Cc: Kook Jin Noh mailto:vorl...@outlook.com>>
Subject: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

Hi,

I'm installting OpenMPI on Archlinux 6.7.0. Everything goes well till:


Testsuite summary for Open MPI 5.0.1

# TOTAL: 13
# PASS:  13
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0

make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
Making check in util
make[2]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make  opal_bit_ops opal_path_nfs bipartite_graph opal_sha256
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
  CCLD opal_bit_ops
  CCLD opal_path_nfs
  CCLD bipartite_graph
  CCLD opal_sha256
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make  check-TESTS
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[4]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
PASS: opal_bit_ops
FAIL: opal_path_nfs
PASS: bipartite_graph
PASS: opal_sha256

Testsuite summary for Open MPI 5.0.1

# TOTAL: 4
# PASS:  3
# SKIP:  0
# XFAIL: 0
# FAIL:  1
# XPASS: 0
# ERROR: 0

See test/util/test-suite.log
Please report to https://www.open-mpi.org/community/help/

make[4]: *** [Makefile:1838: test-suite.log] Error 1
make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[3]: *** [Makefile:1946: check-TESTS] Error 2
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[2]: *** [Makefile:2040: check-am] Error 2
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[1]: *** [Makefile:1416: check-recursive] Error 1
make[1]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test'
make: *** [Makefile:1533: check-recursive] Error 1
make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1'
==> ERROR: A failure occurred in check().
Aborting...
[vorlket@miniserver openmpi-ucx]$

Could you please help me understanding what's going on? Thanks.


Re: [OMPI users] PRTE mismatch

2024-04-21 Thread Kook Jin Noh via users
pacman -R prrte, pacman -Syu prrte solved it.

From: users  On Behalf Of Kook Jin Noh via 
users
Sent: Sunday, 21 April 2024 9:25 PM
To: users@lists.open-mpi.org
Cc: Kook Jin Noh 
Subject: [OMPI users] PRTE mismatch

Hi, while trying to run a MPI program, I received the following:

[vorlket@server ~]$ mpirun -host server:4,midiserver:4,miniserver:1 -np 9 
/home/vorlket/sharedfolder/mpi-prime
--
PRTE detected a mismatch in versions between two processes.  This
typically means that you executed "mpirun" (or "mpiexec") from one
version of PRTE on on node, but your default path on one of the
other nodes upon which you launched found a different version of Open
MPI.

PRTE only supports running exactly the same version between all
processes in a single job.

This will almost certainly cause unpredictable behavior, and may end
up aborting your job.

  Local host: server
  Local process name: [prterun-server-1089@0,0]
  Local PRTE version: 3.0.3
  Peer host:  Unknown
  Peer process name:  [prterun-server-1089@0,1]
  Peer PRTE version:  3.0.5
--
--
PRTE has lost communication with a remote daemon.

  HNP daemon   : [prterun-server-1089@0,0] on node server
  Remote daemon: [prterun-server-1089@0,1] on node midiserver

This is usually due to either a failure of the TCP network
connection to the node, or possibly an internal failure of
the daemon itself. We cannot recover from this failure, and
therefore will terminate the job.

Could you please tell me what I can do to either have the same PRTE version 
across the servers or something else to make the program run?

Thanks.


Re: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

2024-04-21 Thread Kook Jin Noh via users
I makepkg -s again and it installed. The problem solved but would be nice if 
someone can explain the situation. Thanks.

From: users  On Behalf Of Kook Jin Noh via 
users
Sent: Sunday, 21 April 2024 8:18 PM
To: users@lists.open-mpi.org
Cc: Kook Jin Noh 
Subject: [OMPI users] Another OpenMPI 5.0.1. Installation Failure

Hi,

I'm installting OpenMPI on Archlinux 6.7.0. Everything goes well till:


Testsuite summary for Open MPI 5.0.1

# TOTAL: 13
# PASS:  13
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0

make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
Making check in util
make[2]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make  opal_bit_ops opal_path_nfs bipartite_graph opal_sha256
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
  CCLD opal_bit_ops
  CCLD opal_path_nfs
  CCLD bipartite_graph
  CCLD opal_sha256
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make  check-TESTS
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[4]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
PASS: opal_bit_ops
FAIL: opal_path_nfs
PASS: bipartite_graph
PASS: opal_sha256

Testsuite summary for Open MPI 5.0.1

# TOTAL: 4
# PASS:  3
# SKIP:  0
# XFAIL: 0
# FAIL:  1
# XPASS: 0
# ERROR: 0

See test/util/test-suite.log
Please report to https://www.open-mpi.org/community/help/

make[4]: *** [Makefile:1838: test-suite.log] Error 1
make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[3]: *** [Makefile:1946: check-TESTS] Error 2
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[2]: *** [Makefile:2040: check-am] Error 2
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[1]: *** [Makefile:1416: check-recursive] Error 1
make[1]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test'
make: *** [Makefile:1533: check-recursive] Error 1
make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1'
==> ERROR: A failure occurred in check().
Aborting...
[vorlket@miniserver openmpi-ucx]$

Could you please help me understanding what's going on? Thanks.


[OMPI users] PRTE mismatch

2024-04-21 Thread Kook Jin Noh via users
Hi, while trying to run a MPI program, I received the following:

[vorlket@server ~]$ mpirun -host server:4,midiserver:4,miniserver:1 -np 9 
/home/vorlket/sharedfolder/mpi-prime
--
PRTE detected a mismatch in versions between two processes.  This
typically means that you executed "mpirun" (or "mpiexec") from one
version of PRTE on on node, but your default path on one of the
other nodes upon which you launched found a different version of Open
MPI.

PRTE only supports running exactly the same version between all
processes in a single job.

This will almost certainly cause unpredictable behavior, and may end
up aborting your job.

  Local host: server
  Local process name: [prterun-server-1089@0,0]
  Local PRTE version: 3.0.3
  Peer host:  Unknown
  Peer process name:  [prterun-server-1089@0,1]
  Peer PRTE version:  3.0.5
--
--
PRTE has lost communication with a remote daemon.

  HNP daemon   : [prterun-server-1089@0,0] on node server
  Remote daemon: [prterun-server-1089@0,1] on node midiserver

This is usually due to either a failure of the TCP network
connection to the node, or possibly an internal failure of
the daemon itself. We cannot recover from this failure, and
therefore will terminate the job.

Could you please tell me what I can do to either have the same PRTE version 
across the servers or something else to make the program run?

Thanks.


[OMPI users] Another OpenMPI 5.0.1. Installation Failure

2024-04-21 Thread Kook Jin Noh via users
Hi,

I'm installting OpenMPI on Archlinux 6.7.0. Everything goes well till:


Testsuite summary for Open MPI 5.0.1

# TOTAL: 13
# PASS:  13
# SKIP:  0
# XFAIL: 0
# FAIL:  0
# XPASS: 0
# ERROR: 0

make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
Making check in util
make[2]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make  opal_bit_ops opal_path_nfs bipartite_graph opal_sha256
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
  CCLD opal_bit_ops
  CCLD opal_path_nfs
  CCLD bipartite_graph
  CCLD opal_sha256
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make  check-TESTS
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[4]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
PASS: opal_bit_ops
FAIL: opal_path_nfs
PASS: bipartite_graph
PASS: opal_sha256

Testsuite summary for Open MPI 5.0.1

# TOTAL: 4
# PASS:  3
# SKIP:  0
# XFAIL: 0
# FAIL:  1
# XPASS: 0
# ERROR: 0

See test/util/test-suite.log
Please report to https://www.open-mpi.org/community/help/

make[4]: *** [Makefile:1838: test-suite.log] Error 1
make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[3]: *** [Makefile:1946: check-TESTS] Error 2
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[2]: *** [Makefile:2040: check-am] Error 2
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/util'
make[1]: *** [Makefile:1416: check-recursive] Error 1
make[1]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test'
make: *** [Makefile:1533: check-recursive] Error 1
make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1'
==> ERROR: A failure occurred in check().
Aborting...
[vorlket@miniserver openmpi-ucx]$

Could you please help me understanding what's going on? Thanks.


[OMPI users] OpenMPI 5.0.1 Installation Failure

2024-01-26 Thread Kook Jin Noh via users
Hi,

I'm installing OpenMPI 5.0.1 on Archlinux 6.7.1. Everything goes well till:

Making check in datatype
make[2]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make  opal_datatype_test unpack_hetero checksum position position_noncontig 
ddt_test ddt_raw ddt_raw2 unpack_ooo ddt_pack external32 large_data partial 
to_self reduce_local
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
  CCLD opal_datatype_test
  CCLD unpack_hetero
  CCLD checksum
  CCLD position
  CCLD position_noncontig
  CCLD ddt_test
  CCLD ddt_raw
  CCLD ddt_raw2
  CCLD unpack_ooo
  CCLD ddt_pack
  CCLD external32
  CCLD large_data
  CCLD partial
  CCLD to_self
  CCLD reduce_local
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make  check-TESTS
make[3]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[4]: Entering directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
../../config/test-driver: line 112: 1380808 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: opal_datatype_test
PASS: unpack_hetero
../../config/test-driver: line 112: 1380857 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: checksum
../../config/test-driver: line 112: 1380884 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: position
../../config/test-driver: line 112: 1380916 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: position_noncontig
../../config/test-driver: line 112: 1380944 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: ddt_test
../../config/test-driver: line 112: 1380975 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: ddt_raw
PASS: ddt_raw2
PASS: unpack_ooo
../../config/test-driver: line 112: 1381044 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: ddt_pack
../../config/test-driver: line 112: 1381070 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: external32
PASS: large_data
../../config/test-driver: line 112: 1381120 Segmentation fault  (core 
dumped) "$@" >> "$log_file" 2>&1
FAIL: partial

Testsuite summary for Open MPI 5.0.1

# TOTAL: 13
# PASS:  4
# SKIP:  0
# XFAIL: 0
# FAIL:  9
# XPASS: 0
# ERROR: 0

See test/datatype/test-suite.log
Please report to https://www.open-mpi.org/community/help/

make[4]: *** [Makefile:2012: test-suite.log] Error 1
make[4]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[3]: *** [Makefile:2120: check-TESTS] Error 2
make[3]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[2]: *** [Makefile:2277: check-am] Error 2
make[2]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test/datatype'
make[1]: *** [Makefile:1416: check-recursive] Error 1
make[1]: Leaving directory 
'/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1/test'
make: *** [Makefile:1533: check-recursive] Error 1
make: Leaving directory '/home/vorlket/build/openmpi-ucx/src/openmpi-5.0.1'
==> ERROR: A failure occurred in check().
Aborting...


[OMPI users] ORTE was unable to reliably start one or more daemons.

2024-01-11 Thread Kook Jin Noh via users
Hi, I tried to run the following and got an error (the compiled program (with 
mpicc) on the sharedfolder runs ok on each individual machine):

[vorlket@server ~]$ mpirun -host server,miniserver -np 4 
/home/vorlket/sharedfolder/mpi-prime
--
ORTE was unable to reliably start one or more daemons.
This usually is caused by:

* not finding the required libraries and/or binaries on
  one or more nodes. Please check your PATH and LD_LIBRARY_PATH
  settings, or configure OMPI with --enable-orterun-prefix-by-default

* lack of authority to execute on one or more specified nodes.
  Please verify your allocation and authorities.

* the inability to write startup files into /tmp (--tmpdir/orte_tmpdir_base).
  Please check with your sys admin to determine the correct location to use.

*  compilation of the orted with dynamic libraries when static are required
  (e.g., on Cray). Please check your configure cmd line and consider using
  one of the contrib/platform definitions for your system type.

* an inability to create a connection back to mpirun due to a
  lack of common network interfaces and/or no route found between
  them. Please check network connectivity (including firewalls
  and network routing requirements).
--


Hi I have the following installations on archlinux:

server:

sudo pacman -S openmpi

cd ~/.ssh

ssh-keygen -t rsa

cp id_rsa.pub authorized_keys

chmod 600 authorized_key

sssh-copy-id miniserver

sudo vi /etc/ssh/sshd_config: Add the following lines

PubkeyAuthentication yes

RSAAuthentication yes

sudo pacman -S ntp

sudo vi /etc/ntp.conf: Add the following lines

server 0.arch.pool.ntp.org iburst

server 1.arch.pool.ntp.org iburst

server 2.arch.pool.ntp.org iburst

server 3.arch.pool.ntp.org iburst

tos orphan 15

logfile /var/log/ntp.log

sudo ntpd -u ntp:ntp

sudo systemctl start ntpd.service

sudo systemctl enable ntpd.service

sudo pacman -S nfs-utils

mkdir ~/sharedfolder

chmod 777 ~/sharedfolder

sudo vi /etc/exports:

/home/vorlket/sharedfolder  miniserver(rw,sync)

sudo exportfs -arv

sudo systemctl start nfs-server.service

sudo systemctl enable nfs-server.service

miniserver:

sudo pacman -S openmpi

cd ~/.ssh

ssh-keygen -t rsa

cp id_rsa.pub authorized_keys

chmod 600 authorized_key

sssh-copy-id server

sudo vi /etc/ssh/sshd_config:

PubkeyAuthentication yes

RSAAuthentication yes

sudo pacman -S ntp

sudo vi /etc/ntp.conf: Add the following lines

server 192.168.1.3

sudo systemctl start ntpd.service

sudo systemctl enable ntpd.service

sudo pacman -S nfs-utils

mkdir ~/sharedfolder

chmod 777 ~/sharedfolder

sudo vi /etc/fstab:

server:/home/vorlket/sharedfolder   /home/vorlket/sharedfolder  nfs 
_netdev,noauto,x-systemd.automount,x-systemd.mount-timeout=10,timeo=14,x-systemd.idle-timeout=1min
 0 0

sudo systemctl daemon-reload

sudo systemctl restart remote-fs.target

I tried to run the following and got an error:

[vorlket@server ~]$ mpirun -host server,miniserver -np 4 
/home/vorlket/sharedfolder/mpi-prime

--

ORTE was unable to reliably start one or more daemons.

This usually is caused by:



* not finding the required libraries and/or binaries on

  one or more nodes. Please check your PATH and LD_LIBRARY_PATH

  settings, or configure OMPI with --enable-orterun-prefix-by-default



* lack of authority to execute on one or more specified nodes.

  Please verify your allocation and authorities.



* the inability to write startup files into /tmp (--tmpdir/orte_tmpdir_base).

  Please check with your sys admin to determine the correct location to use.



*  compilation of the orted with dynamic libraries when static are required

  (e.g., on Cray). Please check your configure cmd line and consider using

  one of the contrib/platform definitions for your system type.



* an inability to create a connection back to mpirun due to a

  lack of common network interfaces and/or no route found between

  them. Please check network connectivity (including firewalls

  and network routing requirements).

--

mpi-prime.c:

[vorlket@server ~]$ cat /home/vorlket/mpiprac/mpi-prime.c

# include 

# include 

# include 

# include 

# include 



int main ( int argc, char *argv[] );

int prime_number ( int n, int id, int p );

void timestamp ( );



/**/



int main ( int argc, char *argv[] )



/**/

/*

  Purpose:



MAIN is the main program for PRIME_MPI.



  Discussion:



This program calls a version of PRIME_NUMBER that includes

MPI calls for parallel processing.



  Licensing: