Hello all,
I am recently encountering an issue with MUMPS solver.
The error is coming from the dealii::PETScWrappers::SparseDirectMUMPS class
in these lines.
dealii::PETScWrappers::SparseDirectMUMPS
solver(solver_control, mpi_communicator);
solver.solve (system_matrix,
locally_owned_solution,
system_rhs); // error while solving
The error that is printed is
ERR: ERROR : NBROWS > NBROWF
ERR: INODE = 15525
ERR: NBROW= 1 NBROWF= 0
ERR: ROW_LIST= 2
application called MPI_Abort(MPI_COMM_WORLD, -99) - process 0
[cli_0]: aborting job:
application called MPI_Abort(MPI_COMM_WORLD, -99) - process 0
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 0
[cli_0]: aborting job:
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 0
My code uses MPI based parallelism so it uses
parallel::distributed::triangulation.
I am using deal.ii8.3.0 , petsc 3.6.0 and p4est 1.1 .
Any help will be appreciated.
Thanks.
--
The deal.II project is located at http://www.dealii.org/
For mailing list/forum options, see
https://groups.google.com/d/forum/dealii?hl=en
---
You received this message because you are subscribed to the Google Groups
"deal.II User Group" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
For more options, visit https://groups.google.com/d/optout.