Could you try one of the later nightly 1.2 tarballs? We just fixed a shared memory race condition, for example:

        http://www.open-mpi.org/nightly/v1.2/


On Feb 16, 2007, at 12:12 AM, Eric Thibodeau wrote:

Hello devs,

Thought I would let you know there seems to be a problem with 1.2b3r13112 when running the "bisection" test on a Tyan VX50 machine (the 8 DualCore model with 32Gigs of RAM).

OpenMPI was compiled with (as seen from config.log):
configure:116866: running /bin/sh './configure' CFLAGS="-O3 - DNDEBUG -finline-functions -fno-strict-aliasing -pthread" CPPFLAGS=" " FFLAGS="" LDFLAGS=" " --enable-shared --disable- static --prefix=/export/livia/home/parallel/eric/openmpi_x86_64 -- with-mpi=open_mpi --cache-file=/dev/null --srcdir=.

MPPTEST (1.3c) was compiled with:
./configure --with-mpi=$HOME/openmpi_`uname -m`

...which, for some reason, works fine on that system that doesn't have any other MPI implementation (ie: doesn't have LAM-MPI as per this thread).

Then I ran a few tests but this one ran for over it's allowed time (1800 seconds and was going over 50minutes...) and was up to 16Gigs of RAM:

orterun -np 9 ./perftest-1.3c/mpptest -max_run_time 1800 -bisect - size 0 4096 1 -gnuplot -fname HyperTransport/ Global_bisect_0_4096_1.gpl

I had to CTRL-\ the process as CTRL-C wasn't sufficient. 2 mpptest processes and 1 orterun process were using 100% CPU ou of of the 16 cores.

If any of this can be indicative of an OpenMPI bug and if I can help in tracking it down, don't hesitate to ask for details.

And, finally, Anthony, thanks for the MPICC and --with-mpich pointers, I will try those to simplify the build process!

Eric

Le jeudi 15 février 2007 19:51, Anthony Chan a écrit :

As long as mpicc is working, try configuring mpptest as

mpptest/configure MPICC=<OpenMPI-install-dir>/bin/mpicc

or

mpptest/configure  --with-mpich=<OpenMPI-install-dir>

A.Chan

On Thu, 15 Feb 2007, Eric Thibodeau wrote:

Hi Jeff,

        Thanks for your response, I eventually figured it out, here is the
only way I got mpptest to compile:

export LD_LIBRARY_PATH="$HOME/openmpi_`uname -m`/lib"
CC="$HOME/openmpi_`uname -m`/bin/mpicc" ./configure --with- mpi="$HOME/openmpi_`uname -m`"

And, yes I know I should use the mpicc wrapper and all (I do RTFM :P ) but mpptest is less than cooperative and hasn't been updated lately AFAIK.

I'll keep you posted on some results as I get some results out (testing TCP/IP as well as the HyperTransport on a Tyan Beast). Up to now, LAM-MPI seems less efficient at async communications and shows no improovments with persistant communications under TCP/IP. OpenMPI, on the other hand,
seems more efficient using persistant communications when in a
HyperTransport (shmem) environment... I know I am crossing many test
boudaries but I will post some PNGs of my results (as well as how I got to
them ;)

Eric

On Thu, 15 Feb 2007, Jeff Squyres wrote:

I think you want to add $HOME/openmpi_`uname -m`/lib to your
LD_LIBRARY_PATH. This should allow executables created by mpicc (or any derivation thereof, such as extracting flags via showme) to find
the Right shared libraries.

Let us know if that works for you.

FWIW, we do recommend using the wrapper compilers over extracting the flags via --showme whenever possible (it's just simpler and should do
what you need).


On Feb 15, 2007, at 3:38 PM, Eric Thibodeau wrote:

Hello all,


I have been attempting to compile mpptest on my nodes in vain. Here
is my current setup:


Openmpi is in "$HOME/openmpi_`uname -m`" which translates to "/
export/home/eric/openmpi_i686/". I tried the following approaches
(you can see some of these were out of desperation):


CFLAGS=`mpicc --showme:compile` LDFLAGS=`mpicc --showme:link` ./
configure


Configure fails on:

checking whether the C compiler works... configure: error: cannot
run C compiled programs.


The log shows that:

./a.out: error while loading shared libraries: liborte.so.0: cannot
open shared object file: No such file or directory



CC="/export/home/eric/openmpi_i686/bin/mpicc" ./configure --with-
mpi=$HOME/openmpi_`uname -m`

Same problems as above...


LDFLAGS="$HOME/openmpi_`uname -m`/lib" ./configure --with-mpi= $HOME/
openmpi_`uname -m`


Configure fails on:

checking for C compiler default output file name... configure:
error: C compiler cannot create executables


And...finally (not that all of this was done in the presented order):

./configure --with-mpi=$HOME/openmpi_`uname -m`


Which ends with:


checking for library containing MPI_Init... no

configure: error: Could not find MPI library


Anyone can help me with this one...?


Note that LAM-MPI is also installed on these systems...


Eric Thibodeau


_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users




_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


--
Eric Thibodeau
Neural Bucket Solutions Inc.
T. (514) 736-1436
C. (514) 710-0517

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


--
Jeff Squyres
Server Virtualization Business Unit
Cisco Systems


Reply via email to