The problem with doing it that way is that is disallows our in-hose
code teams from using their compilers of choice. Prior to open-mpi we
have been using LA-MPI. LA-MPI has always been compiled in such a
way that it wouldn't matter what other compilers were used to build
mpi applications provided the necessary include and lib files were
linked in the compilations.
-david
On Apr 10, 2006, at 8:00 AM, Ralph Castain wrote:
I'm not an expert on the configure system, but one thing jumps out
at me immediately - you used "gcc" to compile your program. You
really need to use "mpicc" to do so.
I think that might be the source of your errors.
Ralph
David Gunter wrote:
After much fiddling around, I managed to create a version of open-mpi
that would actually build. Unfortunately, I can't run the simplest
of applications with it. Here's the setup I used:
export CC=gcc
export CXX=g++
export FC=gfortran
export F77=gfortran
export CFLAGS="-m32"
export CXXFLAGS="-m32"
export FFLAGS="-m32"
export FCFLAGS="-m32"
export LDFLAGS="-L/usr/lib"
./configure --prefix=/net/scratch1/dog/flash64/openmpi/
openmpi-1.0.2-32b --build=i686-pc-linux-gnu --with-bproc --with-g
m --enable-io-romio --with-romio --with-io-romio-flags='--build=i686-
pc-linux-gnu'
Configure completes, as does 'make' and then 'make install'. Next I
tried to compile a simple MPI_Send test application, which fails
to run:
(flashc 104%) gcc -m32 -I/net/scratch1/dog/flash64/openmpi/
openmpi-1.0.2-32b/include -o send4 send4.c -L/net/scratch1/dog/
flash64/openmpi/openmpi-1.0.2-32b/lib -lmpi
/net/scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/lib/libopal.so.0:
warning: epoll_wait is not implemented and will always fail
/net/scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/lib/libopal.so.0:
warning: epoll_ctl is not implemented and will always fail
/net/scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/lib/libopal.so.0:
warning: epoll_create is not implemented and will always fail
(flashc 105%) which mpiexec
/net/scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/bin/mpiexec
(flashc 106%) mpiexec -n 4 ./send4
[flashc.lanl.gov:32373] mca: base: component_find: unable to open: /
lib/libc.so.6: version `GLIBC_2.3.4' not found (required by /net/
scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/lib/openmpi/
mca_paffinity_linux.so) (ignored)
[flashc.lanl.gov:32373] mca: base: component_find: unable to open:
libbproc.so.4: cannot open shared object file: No such file or
directory (ignored)
[flashc.lanl.gov:32373] mca: base: component_find: unable to open:
libbproc.so.4: cannot open shared object file: No such file or
directory (ignored)
[flashc.lanl.gov:32373] mca: base: component_find: unable to open:
libbproc.so.4: cannot open shared object file: No such file or
directory (ignored)
[flashc.lanl.gov:32373] mca: base: component_find: unable to open:
libbproc.so.4: cannot open shared object file: No such file or
directory (ignored)
[flashc.lanl.gov:32373] mca: base: component_find: unable to open:
libbproc.so.4: cannot open shared object file: No such file or
directory (ignored)
mpiexec: relocation error: /net/scratch1/dog/flash64/openmpi/
openmpi-1.0.2-32b/lib/openmpi/mca_soh_bproc.so: undefined symbol:
bproc_nodelist
I'm still open to suggestions.
-david
On Apr 10, 2006, at 7:11 AM, David R. (Chip) Kent IV wrote:
When running the tests, is the LD_LIBRARY_PATH getting set to lib64
instead of lib or something like that?
Chip
On Sat, Apr 08, 2006 at 02:45:01AM -0600, David Gunter wrote:
I am trying to build a 32-bit compatible OpenMPI for our 64-bit
Bproc
Opteron systems. I saw the thread from last August-September 2005
regarding this but didn't see where it ever succeeded or if any of
the problems had been fixed. Most importantly, romio is
required to
work as well.
Is this possible and how is it done? Here's what I have tried so
far:
setenv CFLAGS -m32
setenv CXXFLAGS -m32
setenv FFLAGS -m32
setenv F90FLAGS -m32
I have used the '--build=i686-pc-linux-gnu' option to the configure
setup as well as --with-io-romio-flags="--build=i686-pc-linux-gnu"
configure halts with errors when trying to run the Fortran 77
tests.
If I remove those env settings and just use the --build option,
configure will proceed to the end but the make will eventually halt
with errors due to a mix of lib64 libs being accessed at some
point.
Any ideas?
-david
--
David Gunter
CCN-8: HPC Environments: Parallel Tools Team
Los Alamos National Laboratory
------------------------------------------------------------
listmanager [ptools_team] Options:
To: listmana...@listserv.lanl.gov
Body: ptools_team
------------------------------------------------------------
--
-----------------------------------------------------
David R. "Chip" Kent IV
Parallel Tools Team
High Performance Computing Environments Group (CCN-8)
Los Alamos National Laboratory
(505)665-5021
drk...@lanl.gov
-----------------------------------------------------
This message is "Technical data or Software Publicly
Available" or "Correspondence".
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users