OK, I think I got things sorted out.  Thanx for your inputs.

Ted

--- On Fri, 2/6/09, Jeff Squyres <jsquy...@cisco.com> wrote:
From: Jeff Squyres <jsquy...@cisco.com>
Subject: Re: [OMPI users] Global Communicator
To: "Open MPI Users" <us...@open-mpi.org>
Cc: ted...@wag.caltech.edu
List-Post: users@lists.open-mpi.org
Date: Friday, February 6, 2009, 2:46 PM

On Feb 6, 2009, at 11:59 AM, McCalla, Mac wrote:

> Does the "default mpirun command" implementation match the build
environment for quest_ompi.x ?
> ie., what mpi implementation (mpich, LAM/MPI, OPENMPI, or other) was
quest_ompi.x  compiled/linked with? and does that match the result of
"which mpirun"?  You might try running a job using your PBS system
that simply executes the which mpirun command and see what you get.

Mac is dead-on in his analysis.

You must compile your MPI application with exactly the same MPI implementation
that you run it -- there is currently no binary compatibility between MPI
implementations.  So if you compile your MPI application with Open MPI, then you
must run it with Open MPI's mpirun/mpiexec (which, as the man page notes,
are exact synonyms of each other in Open MPI -- this is not necessarily true in
other MPI implementations).

If you compile your application with MPICH, then you must use MPICH's
mpirun/mpiexec -- and you'll need to ask them for help if things don't
work (i.e., it's an entirely different project and we're not qualified
to give you help with their software).

Finally, note that Open MPI "understands" PBS/Torque.  If your system
administrator build Open MPI with PBS/Torque support, you don't need to list
the hostfile on Open MPI's mpirun/mpiexec command line -- Open MPI will
"just figure it out" and know a) what hosts to use and b) how many
processes to launch.  Specifically, you can use this:

     mpiexec ${CODE} >/ul/tedhyu/fuelcell/HOH/test/HH.out

(or, as Mac suggests, you might want to provide a full path name to mpiexec to
ensure that you're getting the "Right" mpiexec)

You can check your Open MPI installation to see if it has PBS/Torque support
with the following:

    ompi_info | grep tm

If you see anything in the output like this (although the exact version numbers
may be different):

                 MCA ras: tm (MCA v2.0, API v2.0, Component v1.4)

then your Open MPI has this native PBS/Torque support built-in and you can
avoid specifying a hostfile and -np arguments.

--Jeff Squyres
Cisco Systems




      

Reply via email to