On Thu, Dec/25/2008 08:12:49AM, Jeff Squyres wrote: > It's quite possible that we don't handle this situation properly. Won't > you need to libdir's (one for the 32 bit OMPI executables, and one for the > 64 bit MPI apps)?
I don't need an OPAL environment variable for the executables, just a single OPAL_LIBDIR var for the libraries. (One set of 32-bit executables runs with both 32-bit and 64-bit libraries.) I'm guessing OPAL_LIBDIR will not work for you if you configure with a non-standard --libdir option. -Ethan > > On Dec 23, 2008, at 3:58 PM, Ethan Mallove wrote: > >> I think the problem is that I am doing a multi-lib build. I have >> 32-bit libraries in lib/, and 64-bit libraries in lib/64. I assume I >> do not see the issue for 32-bit tests, because all the dependencies >> are where Open MPI expects them to be. For the 64-bit case, I tried >> setting OPAL_LIBDIR to /opt/openmpi-relocated/lib/lib64, but no luck. >> Given the below configure arguments, what do my OPAL_* env vars need >> to be? (Also, could using --enable-orterun-prefix-by-default interfere >> with OPAL_PREFIX?) >> >> $ ./configure CC=cc CXX=CC F77=f77 FC=f90 --with-openib >> --without-udapl --disable-openib-ibcm --enable-heterogeneous >> --enable-cxx-exceptions --enable-shared --enable-orterun-prefix-by-default >> --with-sge --enable-mpi-f90 --with-mpi-f90-size=small >> --disable-mpi-threads --disable-progress-threads --disable-debug >> CFLAGS="-m32 -xO5" CXXFLAGS="-m32 -xO5" FFLAGS="-m32 -xO5" FCFLAGS="-m32 >> -xO5" >> --prefix=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install >> >> --mandir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/man >> >> --libdir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/lib >> >> --includedir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/include >> >> --without-mx --with-tm=/ws/ompi-tools/orte/torque/current/shared-install32 >> --with-contrib-vt-flags="--prefix=/workspace/em162155/hpc/mtt-scratch/burl-ct-v! >> 20z-12/ompi-tarball-testing/installs/DGQx/install >> --mandir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/man >> >> --libdir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/lib >> >> --includedir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/include >> >> LDFLAGS=-R/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/lib" >> >> $ ./confgiure CC=cc CXX=CC F77=f77 FC=f90 --with-openib >> --without-udapl --disable-openib-ibcm --enable-heterogeneous >> --enable-cxx-exceptions --enable-shared --enable-orterun-prefix-by-default >> --with-sge --enable-mpi-f90 --with-mpi-f90-size=small >> --disable-mpi-threads --disable-progress-threads --disable-debug >> CFLAGS="-m64 -xO5" CXXFLAGS="-m64 -xO5" FFLAGS="-m64 -xO5" FCFLAGS="-m64 >> -xO5" >> --prefix=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install >> >> --mandir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/man >> >> --libdir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/lib/lib64 >> >> --includedir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/include/64 >> >> --without-mx --with-tm=/ws/ompi-tools/orte/torque/current/shared-install64 >> --with-contrib-vt-flags="--prefix=/workspace/em162155/hpc/mtt-scratch/! >> burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install >> --mandir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/man >> >> --libdir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/lib/lib64 >> >> --includedir=/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/include/64 >> >> LDFLAGS=-R/workspace/em162155/hpc/mtt-scratch/burl-ct-v20z-12/ompi-tarball-testing/installs/DGQx/install/lib" >> >> --disable-binaries >> >> -Ethan >> >>> >>> >>> On Dec 22, 2008, at 12:42 PM, Ethan Mallove wrote: >>> >>>> Can anyone get OPAL_PREFIX to work on Linux? A simple test is to see >>>> if the following works for any mpicc/mpirun: >>>> >>>> $ mv <openmpi-installation> /tmp/foo >>>> $ set OPAL_PREFIX /tmp/foo >>>> $ mpicc ... >>>> $ mpirun ... >>>> >>>> If you are able to get the above to run successfully, I'm interested >>>> in your config.log file. >>>> >>>> Thanks, >>>> Ethan >>>> >>>> >>>> On Thu, Dec/18/2008 11:03:25AM, Ethan Mallove wrote: >>>>> Hello, >>>>> >>>>> The below FAQ lists instructions on how to use a relocated Open MPI >>>>> installation: >>>>> >>>>> http://www.open-mpi.org/faq/?category=building#installdirs >>>>> >>>>> On Solaris, OPAL_PREFIX and friends (documented in the FAQ) work for >>>>> me with both MPI (hello_c) and non-MPI (hostname) programs. On Linux, >>>>> I can only get the non-MPI case to work. Here are the environment >>>>> variables I am setting: >>>>> >>>>> $ cat setenv_opal_prefix.csh >>>>> set opal_prefix = "/opt/openmpi-relocated" >>>>> >>>>> setenv OPAL_PREFIX $opal_prefix >>>>> setenv OPAL_BINDIR $opal_prefix/bin >>>>> setenv OPAL_SBINDIR $opal_prefix/sbin >>>>> setenv OPAL_DATAROOTDIR $opal_prefix/share >>>>> setenv OPAL_SYSCONFDIR $opal_prefix/etc >>>>> setenv OPAL_SHAREDSTATEDIR $opal_prefix/com >>>>> setenv OPAL_LOCALSTATEDIR $opal_prefix/var >>>>> setenv OPAL_LIBDIR $opal_prefix/lib >>>>> setenv OPAL_INCLUDEDIR $opal_prefix/include >>>>> setenv OPAL_INFODIR $opal_prefix/info >>>>> setenv OPAL_MANDIR $opal_prefix/man >>>>> >>>>> setenv PATH $opal_prefix/bin:$PATH >>>>> setenv LD_LIBRARY_PATH $opal_prefix/lib:$opal_prefix/lib/64 >>>>> >>>>> Here is the error I get: >>>>> >>>>> $ mpirun -np 2 hello_c >>>>> >>>>> -------------------------------------------------------------------------- >>>>> It looks like opal_init failed for some reason; your parallel process >>>>> is >>>>> likely to abort. There are many reasons that a parallel process can >>>>> fail during opal_init; some of which are due to configuration or >>>>> environment problems. This failure appears to be an internal failure; >>>>> here's some additional information (which may only be relevant to an >>>>> Open MPI developer): >>>>> >>>>> opal_carto_base_select failed >>>>> --> Returned value -13 instead of OPAL_SUCCESS >>>>> >>>>> -------------------------------------------------------------------------- >>>>> [burl-ct-v20z-0:27737] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found in >>>>> file runtime/orte_init.c at line 77 >>>>> >>>>> Any ideas on what's going on? >>>>> >>>>> Thanks, >>>>> Ethan >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> >>> -- >>> Jeff Squyres >>> Cisco Systems >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > -- > Jeff Squyres > Cisco Systems > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users