Further, if one specifies a single process then it works, but as soon
as more than one process starts the program hangs like I mentioned
before.
Daniel
On 10/3/08, Daniel Gruner <[EMAIL PROTECTED]> wrote:
> Well, I just did the same, but with NO success... The processes are
> apparently started, run at the beginning, but then they hang and do
> not finalize. For example, running the "hellow" example from the
> mvapich2 distribution:
>
> [EMAIL PROTECTED] examples]# cat hellow.c
> /* -*- Mode: C; c-basic-offset:4 ; -*- */
> /*
> * (C) 2001 by Argonne National Laboratory.
> * See COPYRIGHT in top-level directory.
> */
>
> #include <stdio.h>
> #include "mpi.h"
>
> int main( int argc, char *argv[] )
> {
> int rank;
> int size;
>
> MPI_Init( 0, 0 );
> MPI_Comm_rank(MPI_COMM_WORLD, &rank);
> MPI_Comm_size(MPI_COMM_WORLD, &size);
> printf( "Hello world from process %d of %d\n", rank, size );
> MPI_Finalize();
> return 0;
> }
>
> [EMAIL PROTECTED] examples]# make hellow
> ../bin/mpicc -I../src/include -I../src/include -c hellow.c
> ../bin/mpicc -o hellow hellow.o
> [EMAIL PROTECTED] examples]# ./hellow
>
> Hello world from process 0 of 1
>
>
> (this was fine, just running on the master). Running on the two nodes
> requires that the xmvapich process be killed (ctrl-C):
>
> [EMAIL PROTECTED] examples]# xmvapich -ap ./hellow
> n0000: Hello world from process 0 of 2
> n0001: Hello world from process 1 of 2
> [EMAIL PROTECTED] examples]#
>
> I have tried other codes, both in C and Fortran, with the same
> behaviour. I don't know if the issue is with xmvapich or with
> mvapich2. Communication is just GigE.
>
>
> Daniel
>
>
>
> On 9/30/08, Abhishek Kulkarni <[EMAIL PROTECTED]> wrote:
> >
> > Just gave this a quick try, and xmvapich seems to run MPI apps compiled
> > with mpich2 without any issues.
> >
> > $ xmvapich -a ./mpihello
> > blender: Hello World from process 0 of 1
> > eregion: Hello World from process 0 of 1
> >
> > Hope that helps,
> >
> >
> > -- Abhishek
> >
> >
> > On Tue, 2008-09-30 at 17:02 +0200, Stefan Boresch wrote:
> > > Thanks for the quick reply!
> > >
> > > On Tue, Sep 30, 2008 at 07:34:37AM -0700, ron minnich wrote:
> > > >
> > > > On Tue, Sep 30, 2008 at 1:57 AM, stefan <[EMAIL PROTECTED]> wrote:
> > > > >
> > > > > the state of xcpu support with MPI libraries -- either of the common
> > > > > free ones
> > > > > is fine (e.g., openmpi, mpich2)
> > > > >
> > > >
> > > > there is now support for mpich2. openmpi is not supported as openmpi
> > > > is (once again) in flux. it has been supported numerous times and has
> > > > changed out from under us numerous times. I no longer use openmpi if I
> > > > have a working mvapich or mpich available.
> > > >
> > >
> > > I am slightly confused. I guess I had inferred the openmpi issues from
> > > the various mailing lists. But I just looked at the latest mpich2
> > > prerelease and found no mentioning of (s)xcpu(2). I thought that some
> > > patches/support on the side of the mpi library are necessary (as, e.g.,
> > > openmpi provides for bproc ...) Or am I completely misunderstanding
> > > something here, and this is somehow handled by xcpu itself ...
> > > I guess there is some difference between
> > >
> > > xrx 192.168.19.2 /bin/date
> > >
> > > and
> > >
> > > xrx 192.168.19.2 <pathto>/mpiexec ...
> > >
> > > and the latter seems too magic to me to run out of the box (it sure
> > > would be nice though ...)
> > >
> > > Sorry for making myself a nuisance -- thanks,
> > >
> > > Stefan Boresch
> > >
> > >
> > >
> >
> >
>