Hello I posted this topic in last week. But Information about this problem was few. And I post again with more information.
I build bewulf type PC Cluster (Cent OS release 6.4). And I studing about MPI.(Open MPI Ver.1.6.4) I tried following sample which using MPI_REDUCE (FORTRAN). Then, following Error occured. -------------------------------------- [bwslv01:30793] *** An error occurred in MPI_Reduce: the reduction operation MPI_SUM is not defined on the MPI_INTEGER datatype [bwslv01:30793] *** on communicator MPI_COMM_WORLD [bwslv01:30793] *** MPI_ERR_OP: invalid reduce operation [bwslv01:30793] *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort-------------------------------------- All informaion is showd in attached file err.log. and source file is attached as main.f
This cluster system consist of one head node and 2 slave nodes. And sharing home directory in head node by NFS. so Open MPI is installed each nodes. When I test this program on only head node, program is run correctly. and output result. But When I test this program on only slave node, same error occured. Please tell me, good idea or advise. Other information is included attached file. Following is construction of directories . ompiReport ├── head │ ├── config.log // Item 3 on help page │ ├── ifocnfig.txt // Item 8 on help page │ ├── lstopo.txt // Item 5 on help page | |---- PATH.txt // Item 7 on help page | |---- LD_LIBRARY_PATH.txt // Item 7 on help page │ └── ompi_info_all.txt // Item 4 on help page ├── ompi_info_full.txt // Item 6 on help page |---- main.f // source file |---- err.log // error message ├── slv01 │ ├── config.log // Item 3 on help page │ ├── ifconfig.txt // Item 8 on help page │ ├── lstopo.txt // Item 5 on help page | |---- PATH.txt // Item 7 on help page | |---- LD_LIBRARY_PATH.txt // Item 7 on help page │ └── ompi_info_all.txt // Item 4 on help page └── slv02 ├── config.log // Item 3 on help page ├── ifconfig.txt // Item 8 on help page ├── lstopo.txt // Item 5 on help page |---- PATH.txt // Item 7 on help page |---- LD_LIBRARY_PATH.txt // Item 7 on help page └── ompi_info_all.txt // Item 4 on help page 3 directories, 13 files Best regards (2013/05/16 23:24), Jeff Squyres (jsquyres) wrote:
(OFF LIST) Let's figure this out off-list and post the final resolution back to the list. This is quite odd. You launched this mpirun from a single node, right? I'm trying to make sure that you're doing non-interactive logins on the remote nodes to find the ompi_info's, because sometimes there's a difference between paths that are set for interactive and non-interactive logins. Can you send all the information listed here: http://www.open-mpi.org/community/help/ On May 16, 2013, at 9:53 AM, Hayato KUNIIE <kuni...@oita.email.ne.jp> wrote:Following is result of mpirun ompi_info on three_nodes. three nodes version is same. Package: Open MPI root@bwhead.clnet Distribution Open MPI root@bwslv01 Distribution Open MPI root@bwslv02 Distribution Open MPI: 1.6.4 1.6.4 1.6.4 Open MPI SVN revision: r28081 r28081 r28081 Open MPI release date: Feb 19, 2013 Feb 19, 2013 Feb 19, 2013 Open RTE: 1.6.4 1.6.4 1.6.4 Open RTE SVN revision: r28081 r28081 r28081 Open RTE release date: Feb 19, 2013 Feb 19, 2013 Feb 19, 2013 OPAL: 1.6.4 1.6.4 1.6.4 OPAL SVN revision: r28081 r28081 r28081 OPAL release date: Feb 19, 2013 Feb 19, 2013 Feb 19, 2013 MPI API: 2.1 2.1 2.1 Ident string: 1.6.4 1.6.4 1.6.4 Prefix: /usr/local /usr/local /usr/local Configured architecture: x86_64-unknown-linux-gnu x86_64-unknown-linux-gnu x86_64-unknown-linux-gnu Configure host: bwhead.clnet bwslv01 bwslv02 Configured by: root root root Configured on: Wed May 8 20:38:14 JST 2013 45 JST 2013 29 JST 2013 Configure host: bwhead.clnet bwslv01 bwslv02 Built by: root root root Built on: Wed May 8 20:48:44 JST 2013 43 JST 2013 38 JST 2013 Built host: bwhead.clnet bwslv01 bwslv02 C bindings: yes yes yes C++ bindings: yes yes yes Fortran77 bindings: yes (all) yes (all) yes (all) Fortran90 bindings: yes yes yes Fortran90 bindings size: small small small C compiler: gcc gcc gcc C compiler absolute: /usr/bin/gcc /usr/bin/gcc /usr/bin/gcc C compiler family name: GNU GNU GNU C compiler version: 4.4.7 4.4.7 4.4.7 C++ compiler: g++ g++ g++ C++ compiler absolute: /usr/bin/g++ /usr/bin/g++ /usr/bin/g++ Fortran77 compiler: gfortran gfortran gfortran Fortran77 compiler abs: /usr/bin/gfortran /usr/bin/gfortran /usr/bin/gfortran Fortran90 compiler: gfortran gfortran gfortran Fortran90 compiler abs: /usr/bin/gfortran /usr/bin/gfortran /usr/bin/gfortran C profiling: yes yes yes C++ profiling: yes yes yes Fortran77 profiling: yes yes yes Fortran90 profiling: yes yes yes C++ exceptions: no no no Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no) no) no) Sparse Groups: no no no Internal debug support: no no no MPI interface warnings: no no no MPI parameter check: runtime runtime runtime Memory profiling support: no no no Memory debugging support: no no no libltdl support: yes yes yes Heterogeneous support: no no no mpirun default --prefix: no no no MPI I/O support: yes yes yes MPI_WTIME support: gettimeofday gettimeofday gettimeofday Symbol vis. support: yes yes yes Host topology support: yes yes yes MPI extensions: affinity example affinity example affinity example FT Checkpoint support: no (checkpoint thread: no) no) no) VampirTrace support: yes yes yes MPI_MAX_PROCESSOR_NAME: 256 256 256 MPI_MAX_ERROR_STRING: 256 256 256 MPI_MAX_OBJECT_NAME: 64 64 64 MPI_MAX_INFO_KEY: 36 36 36 MPI_MAX_INFO_VAL: 256 256 256 MPI_MAX_PORT_NAME: 1024 1024 1024 MPI_MAX_DATAREP_STRING: 128 128 128 Package: Open MPI root@bwslv01 Distribution execinfo (MCA v2.0, API v2.0, Component v1.6.4) execinfo (MCA v2.0, API v2.0, Component v1.6.4) Open MPI: 1.6.4 linux (MCA v2.0, API v2.0, Component v1.6.4) linux (MCA v2.0, API v2.0, Component v1.6.4) Open MPI SVN revision: r28081 hwloc (MCA v2.0, API v2.0, Component v1.6.4) hwloc (MCA v2.0, API v2.0, Component v1.6.4) Open MPI release date: Feb 19, 2013 auto_detect (MCA v2.0, API v2.0, Component v1.6.4) auto_detect (MCA v2.0, API v2.0, Component v1.6.4) Open RTE: 1.6.4 file (MCA v2.0, API v2.0, Component v1.6.4) file (MCA v2.0, API v2.0, Component v1.6.4) Open RTE SVN revision: r28081 mmap (MCA v2.0, API v2.0, Component v1.6.4) mmap (MCA v2.0, API v2.0, Component v1.6.4) Open RTE release date: Feb 19, 2013 posix (MCA v2.0, API v2.0, Component v1.6.4) posix (MCA v2.0, API v2.0, Component v1.6.4) OPAL: 1.6.4 sysv (MCA v2.0, API v2.0, Component v1.6.4) sysv (MCA v2.0, API v2.0, Component v1.6.4) OPAL SVN revision: r28081 first_use (MCA v2.0, API v2.0, Component v1.6.4) first_use (MCA v2.0, API v2.0, Component v1.6.4) OPAL release date: Feb 19, 2013 hwloc (MCA v2.0, API v2.0, Component v1.6.4) hwloc (MCA v2.0, API v2.0, Component v1.6.4) MPI API: 2.1 linux (MCA v2.0, API v2.0, Component v1.6.4) linux (MCA v2.0, API v2.0, Component v1.6.4) Ident string: 1.6.4 env (MCA v2.0, API v2.0, Component v1.6.4) env (MCA v2.0, API v2.0, Component v1.6.4) Prefix: /usr/local config (MCA v2.0, API v2.0, Component v1.6.4) config (MCA v2.0, API v2.0, Component v1.6.4) Configured architecture: x86_64-unknown-linux-gnu linux (MCA v2.0, API v2.0, Component v1.6.4) linux (MCA v2.0, API v2.0, Component v1.6.4) Configure host: bwslv01 hwloc132 (MCA v2.0, API v2.0, Component v1.6.4) hwloc132 (MCA v2.0, API v2.0, Component v1.6.4) Configured by: root orte (MCA v2.0, API v2.0, Component v1.6.4) orte (MCA v2.0, API v2.0, Component v1.6.4) Configured on: Wed May 8 20:56:45 JST 2013 orte (MCA v2.0, API v2.0, Component v1.6.4) orte (MCA v2.0, API v2.0, Component v1.6.4) Configure host: bwslv01 basic (MCA v2.0, API v2.0, Component v1.6.4) basic (MCA v2.0, API v2.0, Component v1.6.4) Built by: root bucket (MCA v2.0, API v2.0, Component v1.6.4) bucket (MCA v2.0, API v2.0, Component v1.6.4) Built on: Wed May 8 21:05:43 JST 2013 basic (MCA v2.0, API v2.0, Component v1.6.4) basic (MCA v2.0, API v2.0, Component v1.6.4) Built host: bwslv01 hierarch (MCA v2.0, API v2.0, Component v1.6.4) hierarch (MCA v2.0, API v2.0, Component v1.6.4) C bindings: yes inter (MCA v2.0, API v2.0, Component v1.6.4) inter (MCA v2.0, API v2.0, Component v1.6.4) C++ bindings: yes self (MCA v2.0, API v2.0, Component v1.6.4) self (MCA v2.0, API v2.0, Component v1.6.4) Fortran77 bindings: yes (all) sm (MCA v2.0, API v2.0, Component v1.6.4) sm (MCA v2.0, API v2.0, Component v1.6.4) Fortran90 bindings: yes sync (MCA v2.0, API v2.0, Component v1.6.4) sync (MCA v2.0, API v2.0, Component v1.6.4) Fortran90 bindings size: small tuned (MCA v2.0, API v2.0, Component v1.6.4) tuned (MCA v2.0, API v2.0, Component v1.6.4) C compiler: gcc romio (MCA v2.0, API v2.0, Component v1.6.4) romio (MCA v2.0, API v2.0, Component v1.6.4) C compiler absolute: /usr/bin/gcc fake (MCA v2.0, API v2.0, Component v1.6.4) fake (MCA v2.0, API v2.0, Component v1.6.4) C compiler family name: GNU rdma (MCA v2.0, API v2.0, Component v1.6.4) rdma (MCA v2.0, API v2.0, Component v1.6.4) C compiler version: 4.4.7 sm (MCA v2.0, API v2.0, Component v1.6.4) sm (MCA v2.0, API v2.0, Component v1.6.4) C++ compiler: g++ bfo (MCA v2.0, API v2.0, Component v1.6.4) bfo (MCA v2.0, API v2.0, Component v1.6.4) C++ compiler absolute: /usr/bin/g++ csum (MCA v2.0, API v2.0, Component v1.6.4) csum (MCA v2.0, API v2.0, Component v1.6.4) Fortran77 compiler: gfortran ob1 (MCA v2.0, API v2.0, Component v1.6.4) ob1 (MCA v2.0, API v2.0, Component v1.6.4) Fortran77 compiler abs: /usr/bin/gfortran v (MCA v2.0, API v2.0, Component v1.6.4) v (MCA v2.0, API v2.0, Component v1.6.4) Fortran90 compiler: gfortran r2 (MCA v2.0, API v2.0, Component v1.6.4) r2 (MCA v2.0, API v2.0, Component v1.6.4) Fortran90 compiler abs: /usr/bin/gfortran vma (MCA v2.0, API v2.0, Component v1.6.4) vma (MCA v2.0, API v2.0, Component v1.6.4) C profiling: yes self (MCA v2.0, API v2.0, Component v1.6.4) self (MCA v2.0, API v2.0, Component v1.6.4) C++ profiling: yes sm (MCA v2.0, API v2.0, Component v1.6.4) sm (MCA v2.0, API v2.0, Component v1.6.4) Fortran77 profiling: yes tcp (MCA v2.0, API v2.0, Component v1.6.4) tcp (MCA v2.0, API v2.0, Component v1.6.4) Fortran90 profiling: yes unity (MCA v2.0, API v2.0, Component v1.6.4) unity (MCA v2.0, API v2.0, Component v1.6.4) C++ exceptions: no pt2pt (MCA v2.0, API v2.0, Component v1.6.4) pt2pt (MCA v2.0, API v2.0, Component v1.6.4) Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no) rdma (MCA v2.0, API v2.0, Component v1.6.4) rdma (MCA v2.0, API v2.0, Component v1.6.4) Sparse Groups: no hnp (MCA v2.0, API v2.0, Component v1.6.4) hnp (MCA v2.0, API v2.0, Component v1.6.4) Internal debug support: no orted (MCA v2.0, API v2.0, Component v1.6.4) orted (MCA v2.0, API v2.0, Component v1.6.4) MPI interface warnings: no tool (MCA v2.0, API v2.0, Component v1.6.4) tool (MCA v2.0, API v2.0, Component v1.6.4) MPI parameter check: runtime tcp (MCA v2.0, API v2.0, Component v1.6.4) tcp (MCA v2.0, API v2.0, Component v1.6.4) Memory profiling support: no default (MCA v2.0, API v2.0, Component v1.6.4) default (MCA v2.0, API v2.0, Component v1.6.4) Memory debugging support: no cm (MCA v2.0, API v2.0, Component v1.6.4) cm (MCA v2.0, API v2.0, Component v1.6.4) libltdl support: yes loadleveler (MCA v2.0, API v2.0, Component v1.6.4) loadleveler (MCA v2.0, API v2.0, Component v1.6.4) Heterogeneous support: no slurm (MCA v2.0, API v2.0, Component v1.6.4) slurm (MCA v2.0, API v2.0, Component v1.6.4) mpirun default --prefix: no load_balance (MCA v2.0, API v2.0, Component v1.6.4) load_balance (MCA v2.0, API v2.0, Component v1.6.4) MPI I/O support: yes rank_file (MCA v2.0, API v2.0, Component v1.6.4) rank_file (MCA v2.0, API v2.0, Component v1.6.4) MPI_WTIME support: gettimeofday resilient (MCA v2.0, API v2.0, Component v1.6.4) resilient (MCA v2.0, API v2.0, Component v1.6.4) Symbol vis. support: yes round_robin (MCA v2.0, API v2.0, Component v1.6.4) round_robin (MCA v2.0, API v2.0, Component v1.6.4) Host topology support: yes seq (MCA v2.0, API v2.0, Component v1.6.4) seq (MCA v2.0, API v2.0, Component v1.6.4) MPI extensions: affinity example topo (MCA v2.0, API v2.0, Component v1.6.4) topo (MCA v2.0, API v2.0, Component v1.6.4) FT Checkpoint support: no (checkpoint thread: no) oob (MCA v2.0, API v2.0, Component v1.6.4) oob (MCA v2.0, API v2.0, Component v1.6.4) VampirTrace support: yes binomial (MCA v2.0, API v2.0, Component v1.6.4) binomial (MCA v2.0, API v2.0, Component v1.6.4) MPI_MAX_PROCESSOR_NAME: 256 cm (MCA v2.0, API v2.0, Component v1.6.4) cm (MCA v2.0, API v2.0, Component v1.6.4) MPI_MAX_ERROR_STRING: 256 direct (MCA v2.0, API v2.0, Component v1.6.4) direct (MCA v2.0, API v2.0, Component v1.6.4) MPI_MAX_OBJECT_NAME: 64 linear (MCA v2.0, API v2.0, Component v1.6.4) linear (MCA v2.0, API v2.0, Component v1.6.4) MPI_MAX_INFO_KEY: 36 radix (MCA v2.0, API v2.0, Component v1.6.4) radix (MCA v2.0, API v2.0, Component v1.6.4) MPI_MAX_INFO_VAL: 256 slave (MCA v2.0, API v2.0, Component v1.6.4) slave (MCA v2.0, API v2.0, Component v1.6.4) MPI_MAX_PORT_NAME: 1024 rsh (MCA v2.0, API v2.0, Component v1.6.4) rsh (MCA v2.0, API v2.0, Component v1.6.4) MPI_MAX_DATAREP_STRING: 128 slurm (MCA v2.0, API v2.0, Component v1.6.4) slurm (MCA v2.0, API v2.0, Component v1.6.4) Package: Open MPI root@bwslv02 Distribution rsh (MCA v2.0, API v2.0, Component v1.6.4) rsh (MCA v2.0, API v2.0, Component v1.6.4) Open MPI: 1.6.4 default (MCA v2.0, API v2.0, Component v1.6.4) default (MCA v2.0, API v2.0, Component v1.6.4) Open MPI SVN revision: r28081 env (MCA v2.0, API v2.0, Component v1.6.4) env (MCA v2.0, API v2.0, Component v1.6.4) Open MPI release date: Feb 19, 2013 hnp (MCA v2.0, API v2.0, Component v1.6.4) hnp (MCA v2.0, API v2.0, Component v1.6.4) Open RTE: 1.6.4 singleton (MCA v2.0, API v2.0, Component v1.6.4) singleton (MCA v2.0, API v2.0, Component v1.6.4) Open RTE SVN revision: r28081 slave (MCA v2.0, API v2.0, Component v1.6.4) slave (MCA v2.0, API v2.0, Component v1.6.4) Open RTE release date: Feb 19, 2013 slurm (MCA v2.0, API v2.0, Component v1.6.4) slurm (MCA v2.0, API v2.0, Component v1.6.4) OPAL: 1.6.4 slurmd (MCA v2.0, API v2.0, Component v1.6.4) slurmd (MCA v2.0, API v2.0, Component v1.6.4) OPAL SVN revision: r28081 tool (MCA v2.0, API v2.0, Component v1.6.4) tool (MCA v2.0, API v2.0, Component v1.6.4) OPAL release date: Feb 19, 2013 bad (MCA v2.0, API v2.0, Component v1.6.4) bad (MCA v2.0, API v2.0, Component v1.6.4) MPI API: 2.1 basic (MCA v2.0, API v2.0, Component v1.6.4) basic (MCA v2.0, API v2.0, Component v1.6.4) Ident string: 1.6.4 hier (MCA v2.0, API v2.0, Component v1.6.4) hier (MCA v2.0, API v2.0, Component v1.6.4) Prefix: /usr/local command (MCA v2.0, API v1.0, Component v1.6.4) command (MCA v2.0, API v1.0, Component v1.6.4) Configured architecture: x86_64-unknown-linux-gnu syslog (MCA v2.0, API v1.0, Component v1.6.4) syslog (MCA v2.0, API v1.0, Component v1.6.4) Configure host: bwslv02 Configured by: root Configured on: Wed May 8 20:56:29 JST 2013 Configure host: bwslv02 Built by: root Built on: Wed May 8 21:05:38 JST 2013 Built host: bwslv02 C bindings: yes C++ bindings: yes Fortran77 bindings: yes (all) Fortran90 bindings: yes Fortran90 bindings size: small C compiler: gcc C compiler absolute: /usr/bin/gcc C compiler family name: GNU C compiler version: 4.4.7 C++ compiler: g++ C++ compiler absolute: /usr/bin/g++ Fortran77 compiler: gfortran Fortran77 compiler abs: /usr/bin/gfortran Fortran90 compiler: gfortran Fortran90 compiler abs: /usr/bin/gfortran C profiling: yes C++ profiling: yes Fortran77 profiling: yes Fortran90 profiling: yes C++ exceptions: no Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no) Sparse Groups: no Internal debug support: no MPI interface warnings: no MPI parameter check: runtime Memory profiling support: no Memory debugging support: no libltdl support: yes Heterogeneous support: no mpirun default --prefix: no MPI I/O support: yes MPI_WTIME support: gettimeofday Symbol vis. support: yes Host topology support: yes MPI extensions: affinity example FT Checkpoint support: no (checkpoint thread: no) VampirTrace support: yes MPI_MAX_PROCESSOR_NAME: 256 MPI_MAX_ERROR_STRING: 256 MPI_MAX_OBJECT_NAME: 64 MPI_MAX_INFO_KEY: 36 MPI_MAX_INFO_VAL: 256 MPI_MAX_PORT_NAME: 1024 MPI_MAX_DATAREP_STRING: 128 MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6.4) MCA memory: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4) MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6.4) MCA carto: file (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6.4) MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6.4) MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4) MCA timer: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6.4) MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6.4) MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6.4) MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6.4) MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6.4) MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: inter (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: self (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: sync (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6.4) MCA io: romio (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: csum (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: v (MCA v2.0, API v2.0, Component v1.6.4) MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6.4) MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: self (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6.4) MCA topo: unity (MCA v2.0, API v2.0, Component v1.6.4) MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6.4) MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: orted (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: tool (MCA v2.0, API v2.0, Component v1.6.4) MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6.4) MCA odls: default (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: cm (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6.4) MCA rml: oob (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: cm (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: direct (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: linear (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: radix (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: slave (MCA v2.0, API v2.0, Component v1.6.4) MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6.4) MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6.4) MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: env (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slave (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: tool (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6.4) MCA notifier: command (MCA v2.0, API v1.0, Component v1.6.4) MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6.4) MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6.4) MCA memory: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4) MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6.4) MCA carto: file (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6.4) MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6.4) MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4) MCA timer: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6.4) MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6.4) MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6.4) MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6.4) MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6.4) MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: inter (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: self (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: sync (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6.4) MCA io: romio (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: csum (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: v (MCA v2.0, API v2.0, Component v1.6.4) MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6.4) MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: self (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6.4) MCA topo: unity (MCA v2.0, API v2.0, Component v1.6.4) MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6.4) MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: orted (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: tool (MCA v2.0, API v2.0, Component v1.6.4) MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6.4) MCA odls: default (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: cm (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6.4) MCA rml: oob (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: cm (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: direct (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: linear (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: radix (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: slave (MCA v2.0, API v2.0, Component v1.6.4) MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6.4) MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6.4) MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: env (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slave (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: tool (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6.4) MCA notifier: command (MCA v2.0, API v1.0, Component v1.6.4) MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6.4) MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6.4) MCA memory: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4) MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6.4) MCA carto: file (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6.4) MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6.4) MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6.4) MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4) MCA timer: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6.4) MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6.4) MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6.4) MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6.4) MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6.4) MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6.4) MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: inter (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: self (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: sync (MCA v2.0, API v2.0, Component v1.6.4) MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6.4) MCA io: romio (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6.4) MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: csum (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6.4) MCA pml: v (MCA v2.0, API v2.0, Component v1.6.4) MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6.4) MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: self (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: sm (MCA v2.0, API v2.0, Component v1.6.4) MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6.4) MCA topo: unity (MCA v2.0, API v2.0, Component v1.6.4) MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6.4) MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: orted (MCA v2.0, API v2.0, Component v1.6.4) MCA iof: tool (MCA v2.0, API v2.0, Component v1.6.4) MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6.4) MCA odls: default (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: cm (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6.4) MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6.4) MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6.4) MCA rml: oob (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: cm (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: direct (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: linear (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: radix (MCA v2.0, API v2.0, Component v1.6.4) MCA routed: slave (MCA v2.0, API v2.0, Component v1.6.4) MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6.4) MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6.4) MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: env (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slave (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6.4) MCA ess: tool (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6.4) MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6.4) MCA notifier: command (MCA v2.0, API v1.0, Component v1.6.4) MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6.4) (2013/05/16 9:12), Jeff Squyres (jsquyres) wrote:I am unable to replicate your error -- 1.6.4 has MPI_REDUCE defined on MPI_SUM properly. Are you absolutely sure you're using OMPI 1.6.4 on all nodes? Try this: mpirun ... ompi_info (insert whatever hostfile and -np value you're using for your fortran test) and see what is reported as the OMPI version from other nodes. On May 15, 2013, at 7:46 AM, Hayato KUNIIE <kuni...@oita.email.ne.jp> wrote:I using Ver, 1.6.4 in all nodes. (2013/05/15 7:10), Jeff Squyres (jsquyres) wrote:Are you sure that you have exactly the same version of Open MPI on all your nodes? On May 14, 2013, at 11:39 AM, Hayato KUNIIE <kuni...@oita.email.ne.jp> wrote:Hello I'm kuni255 I build bewulf type PC Cluster (Cent OS release 6.4). And I studing about MPI.(Open MPI Ver.1.6.4) I tried following sample which using MPI_REDUCE. Then, Error occured. This cluster system consist of one head node and 2 slave nodes. And sharing home directory in head node by NFS. so Open MPI is installed each nodes. When I test this program on only head node, program is run correctly. and output result. But When I test this program on only slave node, same error occured. Please tell me, good idea : ) Error message [bwslv01:30793] *** An error occurred in MPI_Reduce: the reduction operation MPI_SUM is not defined on the MPI_INTEGER datatype [bwslv01:30793] *** on communicator MPI_COMM_WORLD [bwslv01:30793] *** MPI_ERR_OP: invalid reduce operation [bwslv01:30793] *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort -------------------------------------------------------------------------- mpirun has exited due to process rank 1 with PID 30793 on node bwslv01 exiting improperly. There are two reasons this could occur: 1. this process did not call "init" before exiting, but others in the job did. This can cause a job to hang indefinitely while it waits for all processes to call "init". By rule, if one process calls "init", then ALL processes must call "init" prior to termination. 2. this process called "init", but exited without calling "finalize". By rule, all processes that call "init" MUST call "finalize" prior to exiting or it will be considered an "abnormal termination" This may have caused other processes in the application to be terminated by signals sent by mpirun (as reported here). -------------------------------------------------------------------------- [bwhead.clnet:02147] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal [bwhead.clnet:02147] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages Fortran90 source code include 'mpif.h' parameter(nmax=12) integer n(nmax) call mpi_init(ierr) call mpi_comm_size(MPI_COMM_WORLD, isize, ierr) call mpi_comm_rank(MPI_COMM_WORLD, irank, ierr) ista=irank*(nmax/isize) + 1 iend=ista+(nmax/isize-1) isum=0 do i=1,nmax n(i) = i isum = isum + n(i) end do call mpi_reduce(isum, itmp, 1, MPI_INTEGER, MPI_SUM, & 0, MPI_COMM_WORLD, ierr) if (irank == 0) then isum=itmp WRITE(*,*) isum endif call mpi_finalize(ierr) end _______________________________________________ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users_______________________________________________ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users_______________________________________________ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users
ompiReport.tar.xz
Description: Binary data