My operating system is opensuse 12.3 X86_64 I use openmpi  is 1.6-3.1.2
the message is
Open RTE was unable to open the hostfile:
    /usr/lib64/mpi/gcc/openmpi/etc/openmpi-default-hostfile
Check to make sure the path and filename are correct.
--------------------------------------------------------------------------
[linux-mwmw.site:01535] [[25215,0],0] ORTE_ERROR_LOG: Not found in file
base/ras_base_allocate.c at line 200
[linux-mwmw.site:01535] [[25215,0],0] ORTE_ERROR_LOG: Not found in file
base/plm_base_launch_support.c at line 99
[linux-mwmw.site:01535] [[25215,0],0] ORTE_ERROR_LOG: Not found in file
plm_rsh_module.c at line 1167



2013/5/28 Jeff Squyres (jsquyres) <jsquy...@cisco.com>

> Per the email that you forwarded below, I replied to you off list saying
> that we could figure it out without bothering people, and then post the
> final resolution back to the list (I do this sometimes when figuring out a
> problem is going to take a bunch of back-and-forth).  On May 25th, I
> replied to your mail asking for more information that you did not provide.
>
> Do you want to go reply to that mail from May 25?
>
>
>
> On May 27, 2013, at 7:05 AM, Hayato KUNIIE <kuni...@oita.email.ne.jp>
> wrote:
>
> > Hello
> >
> > I posted this topic in last week.
> > But Information about this problem was few.
> > And I post again with more information.
> >
> > I build bewulf type PC Cluster (Cent OS release 6.4). And I studing
> > about MPI.(Open MPI Ver.1.6.4) I tried following sample which using
> > MPI_REDUCE (FORTRAN).
> >
> > Then, following Error occured.
> > --------------------------------------
> > [bwslv01:30793] *** An error occurred in MPI_Reduce: the reduction
> > operation MPI_SUM is not defined on the MPI_INTEGER datatype
> > [bwslv01:30793] *** on communicator MPI_COMM_WORLD
> > [bwslv01:30793] *** MPI_ERR_OP: invalid reduce operation
> > [bwslv01:30793] *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort
> > -------------------------------------- All informaion is showd in
> attached file err.log. and source file is attached as main.f
> >
> > This cluster system consist of one head node and 2 slave nodes.
> > And sharing home directory in head node by NFS. so Open MPI is installed
> > each nodes.
> >
> > When I test this program on only head node, program is run correctly.
> > and output result.
> > But When I test this program on only slave node, same error occured.
> >
> > Please tell me, good idea or advise.
> >
> > Other information is included attached file.
> > Following is construction of directories .
> > ompiReport
> > ├── head
> > │ ├── config.log // Item 3 on help page
> > │ ├── ifocnfig.txt // Item 8 on help page
> > │ ├── lstopo.txt // Item 5 on help page
> > | |---- PATH.txt // Item 7 on help page
> > | |---- LD_LIBRARY_PATH.txt // Item 7 on help page
> > │ └── ompi_info_all.txt // Item 4 on help page
> > ├── ompi_info_full.txt // Item 6 on help page
> > |---- main.f // source file
> > |---- err.log // error message
> > ├── slv01
> > │ ├── config.log // Item 3 on help page
> > │ ├── ifconfig.txt // Item 8 on help page
> > │ ├── lstopo.txt // Item 5 on help page
> > | |---- PATH.txt // Item 7 on help page
> > | |---- LD_LIBRARY_PATH.txt // Item 7 on help page
> > │ └── ompi_info_all.txt // Item 4 on help page
> > └── slv02
> > ├── config.log // Item 3 on help page
> > ├── ifconfig.txt // Item 8 on help page
> > ├── lstopo.txt // Item 5 on help page
> > |---- PATH.txt // Item 7 on help page
> > |---- LD_LIBRARY_PATH.txt // Item 7 on help page
> > └── ompi_info_all.txt // Item 4 on help page
> >
> > 3 directories, 13 files
> >
> > Best regards
> >
> >
> >
> > (2013/05/16 23:24), Jeff Squyres (jsquyres) wrote:
> >> (OFF LIST)
> >>
> >> Let's figure this out off-list and post the final resolution back to
> the list.
> >>
> >> This is quite odd.
> >>
> >> You launched this mpirun from a single node, right?  I'm trying to make
> sure that you're doing non-interactive logins on the remote nodes to find
> the ompi_info's, because sometimes there's a difference between paths that
> are set for interactive and non-interactive logins.
> >>
> >> Can you send all the information listed here:
> >>
> >>     http://www.open-mpi.org/community/help/
> >>
> >>
> >>
> >> On May 16, 2013, at 9:53 AM, Hayato KUNIIE <kuni...@oita.email.ne.jp>
> wrote:
> >>
> >>> Following is result of mpirun ompi_info on three_nodes.
> >>>
> >>> three nodes version is same.
> >>>
> >>> Package: Open MPI root@bwhead.clnet Distribution  Open MPI
> root@bwslv01 Distribution  Open MPI root@bwslv02 Distribution
> >>> Open MPI: 1.6.4  1.6.4  1.6.4
> >>> Open MPI SVN revision: r28081  r28081  r28081
> >>> Open MPI release date: Feb 19, 2013  Feb 19, 2013  Feb 19, 2013
> >>> Open RTE: 1.6.4  1.6.4  1.6.4
> >>> Open RTE SVN revision: r28081  r28081  r28081
> >>> Open RTE release date: Feb 19, 2013  Feb 19, 2013  Feb 19, 2013
> >>> OPAL: 1.6.4  1.6.4  1.6.4
> >>> OPAL SVN revision: r28081  r28081  r28081
> >>> OPAL release date: Feb 19, 2013  Feb 19, 2013  Feb 19, 2013
> >>> MPI API: 2.1  2.1  2.1
> >>> Ident string: 1.6.4  1.6.4  1.6.4
> >>> Prefix: /usr/local  /usr/local  /usr/local
> >>> Configured architecture: x86_64-unknown-linux-gnu
> x86_64-unknown-linux-gnu  x86_64-unknown-linux-gnu
> >>> Configure host: bwhead.clnet  bwslv01  bwslv02
> >>> Configured by: root  root  root
> >>> Configured on: Wed May  8 20:38:14 JST 2013 45 JST 2013 29 JST 2013
> >>> Configure host: bwhead.clnet  bwslv01  bwslv02
> >>> Built by: root  root  root
> >>> Built on: Wed May  8 20:48:44 JST 2013 43 JST 2013 38 JST 2013
> >>> Built host: bwhead.clnet  bwslv01  bwslv02
> >>> C bindings: yes  yes  yes
> >>> C++ bindings: yes  yes  yes
> >>> Fortran77 bindings: yes (all)  yes (all)  yes (all)
> >>> Fortran90 bindings: yes  yes  yes
> >>> Fortran90 bindings size: small  small  small
> >>> C compiler: gcc  gcc  gcc
> >>> C compiler absolute: /usr/bin/gcc  /usr/bin/gcc  /usr/bin/gcc
> >>> C compiler family name: GNU  GNU  GNU
> >>> C compiler version: 4.4.7  4.4.7  4.4.7
> >>> C++ compiler: g++  g++  g++
> >>> C++ compiler absolute: /usr/bin/g++  /usr/bin/g++  /usr/bin/g++
> >>> Fortran77 compiler: gfortran  gfortran  gfortran
> >>> Fortran77 compiler abs: /usr/bin/gfortran  /usr/bin/gfortran
> /usr/bin/gfortran
> >>> Fortran90 compiler: gfortran  gfortran  gfortran
> >>> Fortran90 compiler abs: /usr/bin/gfortran  /usr/bin/gfortran
> /usr/bin/gfortran
> >>> C profiling: yes  yes  yes
> >>> C++ profiling: yes  yes  yes
> >>> Fortran77 profiling: yes  yes  yes
> >>> Fortran90 profiling: yes  yes  yes
> >>> C++ exceptions: no  no  no
> >>> Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no) no)  no)
> >>> Sparse Groups: no  no  no
> >>> Internal debug support: no  no  no
> >>> MPI interface warnings: no  no  no
> >>> MPI parameter check: runtime  runtime  runtime
> >>> Memory profiling support: no  no  no
> >>> Memory debugging support: no  no  no
> >>> libltdl support: yes  yes  yes
> >>> Heterogeneous support: no  no  no
> >>> mpirun default --prefix: no  no  no
> >>> MPI I/O support: yes  yes  yes
> >>> MPI_WTIME support: gettimeofday  gettimeofday  gettimeofday
> >>> Symbol vis. support: yes  yes  yes
> >>> Host topology support: yes  yes  yes
> >>> MPI extensions: affinity example  affinity example  affinity example
> >>> FT Checkpoint support: no (checkpoint thread: no)  no)  no)
> >>> VampirTrace support: yes  yes  yes
> >>> MPI_MAX_PROCESSOR_NAME: 256  256  256
> >>> MPI_MAX_ERROR_STRING: 256  256  256
> >>> MPI_MAX_OBJECT_NAME: 64  64  64
> >>> MPI_MAX_INFO_KEY: 36  36  36
> >>> MPI_MAX_INFO_VAL: 256  256  256
> >>> MPI_MAX_PORT_NAME: 1024  1024  1024
> >>> MPI_MAX_DATAREP_STRING: 128  128  128
> >>> Package: Open MPI root@bwslv01 Distribution  execinfo (MCA v2.0, API
> v2.0, Component v1.6.4)  execinfo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open MPI: 1.6.4  linux (MCA v2.0, API v2.0, Component v1.6.4) linux
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open MPI SVN revision: r28081  hwloc (MCA v2.0, API v2.0, Component
> v1.6.4)  hwloc (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open MPI release date: Feb 19, 2013  auto_detect (MCA v2.0, API v2.0,
> Component v1.6.4)  auto_detect (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open RTE: 1.6.4  file (MCA v2.0, API v2.0, Component v1.6.4)  file
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open RTE SVN revision: r28081  mmap (MCA v2.0, API v2.0, Component
> v1.6.4)  mmap (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open RTE release date: Feb 19, 2013  posix (MCA v2.0, API v2.0,
> Component v1.6.4)  posix (MCA v2.0, API v2.0, Component v1.6.4)
> >>> OPAL: 1.6.4  sysv (MCA v2.0, API v2.0, Component v1.6.4)  sysv (MCA
> v2.0, API v2.0, Component v1.6.4)
> >>> OPAL SVN revision: r28081  first_use (MCA v2.0, API v2.0, Component
> v1.6.4)  first_use (MCA v2.0, API v2.0, Component v1.6.4)
> >>> OPAL release date: Feb 19, 2013  hwloc (MCA v2.0, API v2.0, Component
> v1.6.4)  hwloc (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI API: 2.1  linux (MCA v2.0, API v2.0, Component v1.6.4)  linux (MCA
> v2.0, API v2.0, Component v1.6.4)
> >>> Ident string: 1.6.4  env (MCA v2.0, API v2.0, Component v1.6.4) env
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Prefix: /usr/local  config (MCA v2.0, API v2.0, Component v1.6.4)
> config (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Configured architecture: x86_64-unknown-linux-gnu  linux (MCA v2.0,
> API v2.0, Component v1.6.4)  linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Configure host: bwslv01  hwloc132 (MCA v2.0, API v2.0, Component
> v1.6.4)  hwloc132 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Configured by: root  orte (MCA v2.0, API v2.0, Component v1.6.4) orte
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Configured on: Wed May  8 20:56:45 JST 2013  orte (MCA v2.0, API v2.0,
> Component v1.6.4)  orte (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Configure host: bwslv01  basic (MCA v2.0, API v2.0, Component v1.6.4)
>  basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Built by: root  bucket (MCA v2.0, API v2.0, Component v1.6.4) bucket
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Built on: Wed May  8 21:05:43 JST 2013  basic (MCA v2.0, API v2.0,
> Component v1.6.4)  basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Built host: bwslv01  hierarch (MCA v2.0, API v2.0, Component v1.6.4)
>  hierarch (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C bindings: yes  inter (MCA v2.0, API v2.0, Component v1.6.4) inter
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C++ bindings: yes  self (MCA v2.0, API v2.0, Component v1.6.4) self
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Fortran77 bindings: yes (all)  sm (MCA v2.0, API v2.0, Component
> v1.6.4)  sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Fortran90 bindings: yes  sync (MCA v2.0, API v2.0, Component v1.6.4)
>  sync (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Fortran90 bindings size: small  tuned (MCA v2.0, API v2.0, Component
> v1.6.4)  tuned (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C compiler: gcc  romio (MCA v2.0, API v2.0, Component v1.6.4) romio
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C compiler absolute: /usr/bin/gcc  fake (MCA v2.0, API v2.0, Component
> v1.6.4)  fake (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C compiler family name: GNU  rdma (MCA v2.0, API v2.0, Component
> v1.6.4)  rdma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C compiler version: 4.4.7  sm (MCA v2.0, API v2.0, Component v1.6.4)
>  sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C++ compiler: g++  bfo (MCA v2.0, API v2.0, Component v1.6.4)  bfo
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C++ compiler absolute: /usr/bin/g++  csum (MCA v2.0, API v2.0,
> Component v1.6.4)  csum (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Fortran77 compiler: gfortran  ob1 (MCA v2.0, API v2.0, Component
> v1.6.4)  ob1 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Fortran77 compiler abs: /usr/bin/gfortran  v (MCA v2.0, API v2.0,
> Component v1.6.4)  v (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Fortran90 compiler: gfortran  r2 (MCA v2.0, API v2.0, Component
> v1.6.4)  r2 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Fortran90 compiler abs: /usr/bin/gfortran  vma (MCA v2.0, API v2.0,
> Component v1.6.4)  vma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C profiling: yes  self (MCA v2.0, API v2.0, Component v1.6.4) self
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C++ profiling: yes  sm (MCA v2.0, API v2.0, Component v1.6.4)  sm (MCA
> v2.0, API v2.0, Component v1.6.4)
> >>> Fortran77 profiling: yes  tcp (MCA v2.0, API v2.0, Component v1.6.4)
>  tcp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Fortran90 profiling: yes  unity (MCA v2.0, API v2.0, Component v1.6.4)
>  unity (MCA v2.0, API v2.0, Component v1.6.4)
> >>> C++ exceptions: no  pt2pt (MCA v2.0, API v2.0, Component v1.6.4) pt2pt
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no) rdma
> (MCA v2.0, API v2.0, Component v1.6.4)  rdma (MCA v2.0, API v2.0, Component
> v1.6.4)
> >>> Sparse Groups: no  hnp (MCA v2.0, API v2.0, Component v1.6.4)  hnp
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Internal debug support: no  orted (MCA v2.0, API v2.0, Component
> v1.6.4)  orted (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI interface warnings: no  tool (MCA v2.0, API v2.0, Component
> v1.6.4)  tool (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI parameter check: runtime  tcp (MCA v2.0, API v2.0, Component
> v1.6.4)  tcp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Memory profiling support: no  default (MCA v2.0, API v2.0, Component
> v1.6.4)  default (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Memory debugging support: no  cm (MCA v2.0, API v2.0, Component
> v1.6.4)  cm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> libltdl support: yes  loadleveler (MCA v2.0, API v2.0, Component
> v1.6.4)  loadleveler (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Heterogeneous support: no  slurm (MCA v2.0, API v2.0, Component
> v1.6.4)  slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> mpirun default --prefix: no  load_balance (MCA v2.0, API v2.0,
> Component v1.6.4)  load_balance (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI I/O support: yes  rank_file (MCA v2.0, API v2.0, Component v1.6.4)
>  rank_file (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI_WTIME support: gettimeofday  resilient (MCA v2.0, API v2.0,
> Component v1.6.4)  resilient (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Symbol vis. support: yes  round_robin (MCA v2.0, API v2.0, Component
> v1.6.4)  round_robin (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Host topology support: yes  seq (MCA v2.0, API v2.0, Component v1.6.4)
>  seq (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI extensions: affinity example  topo (MCA v2.0, API v2.0, Component
> v1.6.4)  topo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> FT Checkpoint support: no (checkpoint thread: no)  oob (MCA v2.0, API
> v2.0, Component v1.6.4)  oob (MCA v2.0, API v2.0, Component v1.6.4)
> >>> VampirTrace support: yes  binomial (MCA v2.0, API v2.0, Component
> v1.6.4)  binomial (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI_MAX_PROCESSOR_NAME: 256  cm (MCA v2.0, API v2.0, Component v1.6.4)
>  cm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI_MAX_ERROR_STRING: 256  direct (MCA v2.0, API v2.0, Component
> v1.6.4)  direct (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI_MAX_OBJECT_NAME: 64  linear (MCA v2.0, API v2.0, Component v1.6.4)
>  linear (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI_MAX_INFO_KEY: 36  radix (MCA v2.0, API v2.0, Component v1.6.4)
>  radix (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI_MAX_INFO_VAL: 256  slave (MCA v2.0, API v2.0, Component v1.6.4)
>  slave (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI_MAX_PORT_NAME: 1024  rsh (MCA v2.0, API v2.0, Component v1.6.4)
>  rsh (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI_MAX_DATAREP_STRING: 128  slurm (MCA v2.0, API v2.0, Component
> v1.6.4)  slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Package: Open MPI root@bwslv02 Distribution  rsh (MCA v2.0, API v2.0,
> Component v1.6.4)  rsh (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open MPI: 1.6.4  default (MCA v2.0, API v2.0, Component v1.6.4)
> default (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open MPI SVN revision: r28081  env (MCA v2.0, API v2.0, Component
> v1.6.4)  env (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open MPI release date: Feb 19, 2013  hnp (MCA v2.0, API v2.0,
> Component v1.6.4)  hnp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open RTE: 1.6.4  singleton (MCA v2.0, API v2.0, Component v1.6.4)
> singleton (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open RTE SVN revision: r28081  slave (MCA v2.0, API v2.0, Component
> v1.6.4)  slave (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Open RTE release date: Feb 19, 2013  slurm (MCA v2.0, API v2.0,
> Component v1.6.4)  slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> OPAL: 1.6.4  slurmd (MCA v2.0, API v2.0, Component v1.6.4)  slurmd
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> OPAL SVN revision: r28081  tool (MCA v2.0, API v2.0, Component v1.6.4)
>  tool (MCA v2.0, API v2.0, Component v1.6.4)
> >>> OPAL release date: Feb 19, 2013  bad (MCA v2.0, API v2.0, Component
> v1.6.4)  bad (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MPI API: 2.1  basic (MCA v2.0, API v2.0, Component v1.6.4)  basic (MCA
> v2.0, API v2.0, Component v1.6.4)
> >>> Ident string: 1.6.4  hier (MCA v2.0, API v2.0, Component v1.6.4) hier
> (MCA v2.0, API v2.0, Component v1.6.4)
> >>> Prefix: /usr/local  command (MCA v2.0, API v1.0, Component v1.6.4)
>  command (MCA v2.0, API v1.0, Component v1.6.4)
> >>> Configured architecture: x86_64-unknown-linux-gnu  syslog (MCA v2.0,
> API v1.0, Component v1.6.4)  syslog (MCA v2.0, API v1.0, Component v1.6.4)
> >>> Configure host: bwslv02
> >>> Configured by: root
> >>> Configured on: Wed May  8 20:56:29 JST 2013
> >>> Configure host: bwslv02
> >>> Built by: root
> >>> Built on: Wed May  8 21:05:38 JST 2013
> >>> Built host: bwslv02
> >>> C bindings: yes
> >>> C++ bindings: yes
> >>> Fortran77 bindings: yes (all)
> >>> Fortran90 bindings: yes
> >>> Fortran90 bindings size: small
> >>> C compiler: gcc
> >>> C compiler absolute: /usr/bin/gcc
> >>> C compiler family name: GNU
> >>> C compiler version: 4.4.7
> >>> C++ compiler: g++
> >>> C++ compiler absolute: /usr/bin/g++
> >>> Fortran77 compiler: gfortran
> >>> Fortran77 compiler abs: /usr/bin/gfortran
> >>> Fortran90 compiler: gfortran
> >>> Fortran90 compiler abs: /usr/bin/gfortran
> >>> C profiling: yes
> >>> C++ profiling: yes
> >>> Fortran77 profiling: yes
> >>> Fortran90 profiling: yes
> >>> C++ exceptions: no
> >>> Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no)
> >>> Sparse Groups: no
> >>> Internal debug support: no
> >>> MPI interface warnings: no
> >>> MPI parameter check: runtime
> >>> Memory profiling support: no
> >>> Memory debugging support: no
> >>> libltdl support: yes
> >>> Heterogeneous support: no
> >>> mpirun default --prefix: no
> >>> MPI I/O support: yes
> >>> MPI_WTIME support: gettimeofday
> >>> Symbol vis. support: yes
> >>> Host topology support: yes
> >>> MPI extensions: affinity example
> >>> FT Checkpoint support: no (checkpoint thread: no)
> >>> VampirTrace support: yes
> >>> MPI_MAX_PROCESSOR_NAME: 256
> >>> MPI_MAX_ERROR_STRING: 256
> >>> MPI_MAX_OBJECT_NAME: 64
> >>> MPI_MAX_INFO_KEY: 36
> >>> MPI_MAX_INFO_VAL: 256
> >>> MPI_MAX_PORT_NAME: 1024
> >>> MPI_MAX_DATAREP_STRING: 128
> >>> MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA memory: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA carto: file (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA timer: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: inter (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: self (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: sync (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA io: romio (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: csum (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: v (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: self (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA topo: unity (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: orted (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: tool (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA odls: default (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: cm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rml: oob (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: cm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: direct (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: linear (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: radix (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: slave (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: env (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slave (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: tool (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA notifier: command (MCA v2.0, API v1.0, Component v1.6.4)
> >>> MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6.4)
> >>> MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA memory: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA carto: file (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA timer: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: inter (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: self (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: sync (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA io: romio (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: csum (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: v (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: self (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA topo: unity (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: orted (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: tool (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA odls: default (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: cm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rml: oob (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: cm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: direct (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: linear (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: radix (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: slave (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: env (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slave (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: tool (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA notifier: command (MCA v2.0, API v1.0, Component v1.6.4)
> >>> MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6.4)
> >>> MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA memory: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA carto: file (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA timer: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: inter (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: self (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: sync (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA io: romio (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: csum (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA pml: v (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: self (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: sm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA topo: unity (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: orted (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA iof: tool (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA odls: default (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: cm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA rml: oob (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: cm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: direct (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: linear (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: radix (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA routed: slave (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: env (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slave (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA ess: tool (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6.4)
> >>> MCA notifier: command (MCA v2.0, API v1.0, Component v1.6.4)
> >>> MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6.4)
> >>>
> >>>
> >>>
> >>>
> >>> (2013/05/16 9:12), Jeff Squyres (jsquyres) wrote:
> >>>> I am unable to replicate your error -- 1.6.4 has MPI_REDUCE defined
> on MPI_SUM properly.
> >>>>
> >>>> Are you absolutely sure you're using OMPI 1.6.4 on all nodes?
> >>>>
> >>>> Try this:
> >>>>
> >>>>     mpirun ... ompi_info
> >>>>
> >>>> (insert whatever hostfile and -np value you're using for your fortran
> test) and see what is reported as the OMPI version from other nodes.
> >>>>
> >>>>
> >>>> On May 15, 2013, at 7:46 AM, Hayato KUNIIE <kuni...@oita.email.ne.jp>
> wrote:
> >>>>
> >>>>> I using Ver, 1.6.4 in all nodes.
> >>>>>
> >>>>> (2013/05/15 7:10), Jeff Squyres (jsquyres) wrote:
> >>>>>> Are you sure that you have exactly the same version of Open MPI on
> all your nodes?
> >>>>>>
> >>>>>>
> >>>>>> On May 14, 2013, at 11:39 AM, Hayato KUNIIE <
> kuni...@oita.email.ne.jp> wrote:
> >>>>>>
> >>>>>>> Hello I'm kuni255
> >>>>>>>
> >>>>>>> I build bewulf type PC Cluster (Cent OS release 6.4). And I studing
> >>>>>>> about MPI.(Open MPI Ver.1.6.4) I tried following sample which using
> >>>>>>> MPI_REDUCE.
> >>>>>>>
> >>>>>>> Then, Error occured.
> >>>>>>>
> >>>>>>> This cluster system consist of one head node and 2 slave nodes.
> >>>>>>> And sharing home directory in head node by NFS. so Open MPI is
> installed
> >>>>>>> each nodes.
> >>>>>>>
> >>>>>>> When I test this program on only head node, program is run
> correctly.
> >>>>>>> and output result.
> >>>>>>> But When I test this program on only slave node, same error
> occured.
> >>>>>>>
> >>>>>>> Please tell me, good idea : )
> >>>>>>>
> >>>>>>> Error message
> >>>>>>> [bwslv01:30793] *** An error occurred in MPI_Reduce: the reduction
> >>>>>>> operation MPI_SUM is not defined on the MPI_INTEGER datatype
> >>>>>>> [bwslv01:30793] *** on communicator MPI_COMM_WORLD
> >>>>>>> [bwslv01:30793] *** MPI_ERR_OP: invalid reduce operation
> >>>>>>> [bwslv01:30793] *** MPI_ERRORS_ARE_FATAL: your MPI job will now
> abort
> >>>>>>>
> --------------------------------------------------------------------------
> >>>>>>> mpirun has exited due to process rank 1 with PID 30793 on
> >>>>>>> node bwslv01 exiting improperly. There are two reasons this could
> occur:
> >>>>>>>
> >>>>>>> 1. this process did not call "init" before exiting, but others in
> >>>>>>> the job did. This can cause a job to hang indefinitely while it
> waits
> >>>>>>> for all processes to call "init". By rule, if one process calls
> "init",
> >>>>>>> then ALL processes must call "init" prior to termination.
> >>>>>>>
> >>>>>>> 2. this process called "init", but exited without calling
> "finalize".
> >>>>>>> By rule, all processes that call "init" MUST call "finalize" prior
> to
> >>>>>>> exiting or it will be considered an "abnormal termination"
> >>>>>>>
> >>>>>>> This may have caused other processes in the application to be
> >>>>>>> terminated by signals sent by mpirun (as reported here).
> >>>>>>>
> --------------------------------------------------------------------------
> >>>>>>> [bwhead.clnet:02147] 1 more process has sent help message
> >>>>>>> help-mpi-errors.txt / mpi_errors_are_fatal
> >>>>>>> [bwhead.clnet:02147] Set MCA parameter "orte_base_help_aggregate"
> to 0
> >>>>>>> to see all help / error messages
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>> Fortran90 source code
> >>>>>>> include 'mpif.h'
> >>>>>>> parameter(nmax=12)
> >>>>>>> integer n(nmax)
> >>>>>>>
> >>>>>>> call mpi_init(ierr)
> >>>>>>> call mpi_comm_size(MPI_COMM_WORLD, isize, ierr)
> >>>>>>> call mpi_comm_rank(MPI_COMM_WORLD, irank, ierr)
> >>>>>>> ista=irank*(nmax/isize) + 1
> >>>>>>> iend=ista+(nmax/isize-1)
> >>>>>>> isum=0
> >>>>>>> do i=1,nmax
> >>>>>>> n(i) = i
> >>>>>>> isum = isum + n(i)
> >>>>>>> end do
> >>>>>>> call mpi_reduce(isum, itmp, 1, MPI_INTEGER, MPI_SUM,
> >>>>>>> & 0, MPI_COMM_WORLD, ierr)
> >>>>>>>
> >>>>>>> if (irank == 0) then
> >>>>>>> isum=itmp
> >>>>>>> WRITE(*,*) isum
> >>>>>>> endif
> >>>>>>> call mpi_finalize(ierr)
> >>>>>>> end
> >>>>>>> _______________________________________________
> >>>>>>> users mailing list
> >>>>>>> us...@open-mpi.org
> >>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>> _______________________________________________
> >>>>> users mailing list
> >>>>> us...@open-mpi.org
> >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>> _______________________________________________
> >>> users mailing list
> >>> us...@open-mpi.org
> >>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>
> >
> >
> > <ompiReport.tar.xz>_______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> --
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to