Yes, I did. I got this error message - the configure log shows that I installed mumps.
0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: No support for this operation for this object type! [0]PETSC ERROR: Matrix format seqdense does not have a solver package mumps for LU. Perhaps you must ./configure with --download-mumps! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Development HG revision: cc5f6de4d644fb53ec2bbf114fa776073e3e8534 HG Date: Fri Dec 21 11:22:24 2012 -0600 [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: /home/gokhale/WAIGEN/GDEB-WAIGEN2012/bin/waiproj on a linux-gcc named asd1.wai.com by gokhale Wed Jan 2 14:29:19 2013 [0]PETSC ERROR: Libraries linked from /opt/petsc/petsc-dev/linux-gcc-gpp-mpich-mumps-complex-elemental/lib [0]PETSC ERROR: Configure run at Fri Dec 21 14:30:56 2012 [0]PETSC ERROR: Configure options --with-x=0 --with-mpi=1 --download-mpich=yes --with-x11=0 --with-debugging=0 --with-clanguage=C++ --with-shared-libraries=1 --download-mumps=yes --download-f-blas-lapack=1 --download-parmetis=1 --download-metis --download-scalapack=1 --download-blacs=1 --with-cmake=/usr/bin/cmake28 --with-scalar-type=complex --download-elemental [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: MatGetFactor() line 3944 in /opt/petsc/petsc-dev/src/mat/interface/matrix.c [0]PETSC ERROR: PCSetUp_LU() line 133 in /opt/petsc/petsc-dev/src/ksp/pc/impls/factor/lu/lu.c [0]PETSC ERROR: PCSetUp() line 832 in /opt/petsc/petsc-dev/src/ksp/pc/interface/precon.c [0]PETSC ERROR: KSPSetUp() line 267 in /opt/petsc/petsc-dev/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: KSPSolve() line 376 in /opt/petsc/petsc-dev/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: main() line 169 in src/examples/waiproj.c application called MPI_Abort(MPI_COMM_WORLD, 56) - process 0 [cli_0]: aborting job: application called MPI_Abort(MPI_COMM_WORLD, 56) - process 0 -Nachiket On Wed, Jan 2, 2013 at 2:33 PM, Jed Brown <jedbrown at mcs.anl.gov> wrote: > Did you try it? Yes, it works. > > > On Wed, Jan 2, 2013 at 1:31 PM, Nachiket Gokhale <gokhalen at gmail.com> > wrote: >> >> Does MUMPS work with PETSC in serial i.e. one MPI process? I need to >> run in serial because I have to perform certain dense matrix >> multiplications which do not work in parallel. If mumps does not >> work, I think I will try superlu. >> >> -Nachiket > >
