Am 22.04.2014 13:08, schrieb Jed Brown:
Niklas Fischer <[email protected]> writes:
Hello,
I have attached a small test case for a problem I am experiencing. What
this dummy program does is it reads a vector and a matrix from a text
file and then solves Ax=b. The same data is available in two forms:
- everything is in one file (matops.s.0 and vops.s.0)
- the matrix and vector are split between processes (matops.0,
matops.1, vops.0, vops.1)
The serial version of the program works perfectly fine but unfortunately
errors occure, when running the parallel version:
make && mpirun -n 2 a.out matops vops
mpic++ -DPETSC_CLANGUAGE_CXX -isystem
/home/data/fischer/libs/petsc-3.4.3/arch-linux2-c-debug/include -isystem
/home/data/fischer/libs/petsc-3.4.3/include petsctest.cpp -Werror -Wall
-Wpedantic -std=c++11 -L
/home/data/fischer/libs/petsc-3.4.3/arch-linux2-c-debug/lib -lpetsc
/usr/bin/ld: warning: libmpi_cxx.so.0, needed by
/home/data/fischer/libs/petsc-3.4.3/arch-linux2-c-debug/lib/libpetsc.so,
may conflict with libmpi_cxx.so.1
/usr/bin/ld: warning: libmpi.so.0, needed by
/home/data/fischer/libs/petsc-3.4.3/arch-linux2-c-debug/lib/libpetsc.so,
may conflict with libmpi.so.1
librdmacm: couldn't read ABI version.
librdmacm: assuming: 4
CMA: unable to get RDMA device list
--------------------------------------------------------------------------
[[43019,1],0]: A high-performance Open MPI point-to-point messaging module
was unable to find any relevant network interfaces:
Module: OpenFabrics (openib)
Host: dornroeschen.igpm.rwth-aachen.de
CMA: unable to get RDMA device list
It looks like your MPI is either broken or some of the code linked into
your application was compiled with a different MPI or different version.
Make sure you can compile and run simple MPI programs in parallel.
Hello Jed,
thank you for your inputs. Unfortunately MPI does not seem to be the
issue here. The attachment contains a simple MPI hello world program
which runs flawlessly (I will append the output to this mail) and I have
not encountered any problems with other MPI programs. My question still
stands.
Greetings,
Niklas Fischer
mpirun -np 2 ./mpitest
librdmacm: couldn't read ABI version.
librdmacm: assuming: 4
CMA: unable to get RDMA device list
--------------------------------------------------------------------------
[[44086,1],0]: A high-performance Open MPI point-to-point messaging module
was unable to find any relevant network interfaces:
Module: OpenFabrics (openib)
Host: dornroeschen.igpm.rwth-aachen.de
Another transport will be used instead, although this may result in
lower performance.
--------------------------------------------------------------------------
librdmacm: couldn't read ABI version.
librdmacm: assuming: 4
CMA: unable to get RDMA device list
Hello world from processor dornroeschen.igpm.rwth-aachen.de, rank 0 out
of 2 processors
Hello world from processor dornroeschen.igpm.rwth-aachen.de, rank 1 out
of 2 processors
[dornroeschen.igpm.rwth-aachen.de:128141] 1 more process has sent help
message help-mpi-btl-base.txt / btl:no-nics
[dornroeschen.igpm.rwth-aachen.de:128141] Set MCA parameter
"orte_base_help_aggregate" to 0 to see all help / error messages