I am unable to replicate your error -- 1.6.4 has MPI_REDUCE defined on MPI_SUM 
properly.

Are you absolutely sure you're using OMPI 1.6.4 on all nodes?

Try this:

    mpirun ... ompi_info

(insert whatever hostfile and -np value you're using for your fortran test) and 
see what is reported as the OMPI version from other nodes.


On May 15, 2013, at 7:46 AM, Hayato KUNIIE <kuni...@oita.email.ne.jp> wrote:

> I using Ver, 1.6.4 in all nodes.
> 
> (2013/05/15 7:10), Jeff Squyres (jsquyres) wrote:
>> Are you sure that you have exactly the same version of Open MPI on all your 
>> nodes?
>> 
>> 
>> On May 14, 2013, at 11:39 AM, Hayato KUNIIE <kuni...@oita.email.ne.jp> wrote:
>> 
>>> Hello I'm kuni255
>>> 
>>> I build bewulf type PC Cluster (Cent OS release 6.4). And I studing
>>> about MPI.(Open MPI Ver.1.6.4) I tried following sample which using
>>> MPI_REDUCE.
>>> 
>>> Then, Error occured.
>>> 
>>> This cluster system consist of one head node and 2 slave nodes.
>>> And sharing home directory in head node by NFS. so Open MPI is installed
>>> each nodes.
>>> 
>>> When I test this program on only head node, program is run correctly.
>>> and output result.
>>> But When I test this program on only slave node, same error occured.
>>> 
>>> Please tell me, good idea : )
>>> 
>>> Error message
>>> [bwslv01:30793] *** An error occurred in MPI_Reduce: the reduction
>>> operation MPI_SUM is not defined on the MPI_INTEGER datatype
>>> [bwslv01:30793] *** on communicator MPI_COMM_WORLD
>>> [bwslv01:30793] *** MPI_ERR_OP: invalid reduce operation
>>> [bwslv01:30793] *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort
>>> --------------------------------------------------------------------------
>>> mpirun has exited due to process rank 1 with PID 30793 on
>>> node bwslv01 exiting improperly. There are two reasons this could occur:
>>> 
>>> 1. this process did not call "init" before exiting, but others in
>>> the job did. This can cause a job to hang indefinitely while it waits
>>> for all processes to call "init". By rule, if one process calls "init",
>>> then ALL processes must call "init" prior to termination.
>>> 
>>> 2. this process called "init", but exited without calling "finalize".
>>> By rule, all processes that call "init" MUST call "finalize" prior to
>>> exiting or it will be considered an "abnormal termination"
>>> 
>>> This may have caused other processes in the application to be
>>> terminated by signals sent by mpirun (as reported here).
>>> --------------------------------------------------------------------------
>>> [bwhead.clnet:02147] 1 more process has sent help message
>>> help-mpi-errors.txt / mpi_errors_are_fatal
>>> [bwhead.clnet:02147] Set MCA parameter "orte_base_help_aggregate" to 0
>>> to see all help / error messages
>>> 
>>> 
>>> 
>>> 
>>> Fortran90 source code
>>> include 'mpif.h'
>>> parameter(nmax=12)
>>> integer n(nmax)
>>> 
>>> call mpi_init(ierr)
>>> call mpi_comm_size(MPI_COMM_WORLD, isize, ierr)
>>> call mpi_comm_rank(MPI_COMM_WORLD, irank, ierr)
>>> ista=irank*(nmax/isize) + 1
>>> iend=ista+(nmax/isize-1)
>>> isum=0
>>> do i=1,nmax
>>> n(i) = i
>>> isum = isum + n(i)
>>> end do
>>> call mpi_reduce(isum, itmp, 1, MPI_INTEGER, MPI_SUM,
>>> & 0, MPI_COMM_WORLD, ierr)
>>> 
>>> if (irank == 0) then
>>> isum=itmp
>>> WRITE(*,*) isum
>>> endif
>>> call mpi_finalize(ierr)
>>> end
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/


Reply via email to