Hi,

i try to solve a problem in parallel with MUMPS as the direct solver. As
long as I run the program on only 1 node with 6 processors everything works
fine! But using 2 nodes with 3 processors each gets mumps stuck in the
factorization.

For the purpose of testing I run the ex2.c on a resolution of 100x100
(which is of course way to small for a direct solver in parallel).

The code is run with :
mpirun ./ex2 -on_error_abort -pc_type lu -pc_factor_mat_solver_package
mumps -ksp_type preonly -log_summary -options_left -m 100 -n 100
-mat_mumps_icntl_4 3

The petsc-configuration I used is:
--prefix=/opt/Petsc/3.4.4.extended --with-mpi=yes
--with-mpi-dir=/opt/Openmpi/1.9a/ --with-debugging=no --download-mumps
 --download-scalapack --download-parmetis --download-metis

Is this common behavior? Or is there an error in the petsc configuration I
am using here?

Best,
Gunnar

Reply via email to