Victoria, please keep the list in copy. > I am not understanding how can I switch to ParMetis if it does not appear in > the options of -mat_mumps_icntl_7.In the options I only have Metis and not > ParMetis.
You need to use -mat_mumps_icntl_28 2 -mat_mumps_icntl_29 2 Barry, I don’t think we can programmatically shut off this warning, it’s guarded by a bunch of KEEP() values, see src/dana_driver.F:4707, which are only settable/gettable by people with access to consortium releases. I’ll ask the MUMPS people for confirmation. Note that this warning is only printed to screen with the option -mat_mumps_icntl_4 2 (or higher), so this won’t show up for standard runs. Thanks, Pierre > On 1 Nov 2023, at 5:52 PM, Barry Smith <[email protected]> wrote: > > > Pierre, > > Could the PETSc MUMPS interface "turn-off" ICNTL(6) in this situation so > as to not trigger the confusing warning message from MUMPS? > > Barry > >> On Nov 1, 2023, at 12:17 PM, Pierre Jolivet <[email protected]> wrote: >> >> >> >>> On 1 Nov 2023, at 3:33 PM, Zhang, Hong via petsc-users >>> <[email protected]> wrote: >>> >>> Victoria, >>> "** Maximum transversal (ICNTL(6)) not allowed because matrix is distributed >>> Ordering based on METIS" >> >> This warning is benign and appears for every run using a sequential >> partitioner in MUMPS with a MATMPIAIJ. >> (I’m not saying switching to ParMETIS will not make the issue go away) >> >> Thanks, >> Pierre >> >> $ ../../../../arch-darwin-c-debug-real/bin/mpirun -n 2 ./ex2 -pc_type lu >> -mat_mumps_icntl_4 2 >> Entering DMUMPS 5.6.2 from C interface with JOB, N = 1 56 >> executing #MPI = 2, without OMP >> >> ================================================= >> MUMPS compiled with option -Dmetis >> MUMPS compiled with option -Dparmetis >> MUMPS compiled with option -Dpord >> MUMPS compiled with option -Dptscotch >> MUMPS compiled with option -Dscotch >> ================================================= >> L U Solver for unsymmetric matrices >> Type of parallelism: Working host >> >> ****** ANALYSIS STEP ******** >> >> ** Maximum transversal (ICNTL(6)) not allowed because matrix is distributed >> Processing a graph of size: 56 with 194 edges >> Ordering based on AMF >> WARNING: Largest root node of size 26 not selected for parallel >> execution >> >> Leaving analysis phase with ... >> INFOG(1) = 0 >> INFOG(2) = 0 >> […] >> >>> Try parmetis. >>> Hong >>> From: petsc-users <[email protected]> on behalf of Victoria >>> Rolandi <[email protected]> >>> Sent: Tuesday, October 31, 2023 10:30 PM >>> To: [email protected] <[email protected]> >>> Subject: [petsc-users] Error using Metis with PETSc installed with MUMPS >>> >>> Hi, >>> >>> I'm solving a large sparse linear system in parallel and I am using PETSc >>> with MUMPS. I am trying to test different options, like the ordering of the >>> matrix. Everything works if I use the -mat_mumps_icntl_7 2 or >>> -mat_mumps_icntl_7 0 options (with the first one, AMF, performing better >>> than AMD), however when I test METIS -mat_mumps_icntl_7 5 I get an error >>> (reported at the end of the email). >>> >>> I have configured PETSc with the following options: >>> >>> --with-cc=mpiicc --with-cxx=mpiicpc --with-fc=mpiifort >>> --with-scalar-type=complex --with-debugging=0 --with-precision=single >>> --download-mumps --download-scalapack --download-parmetis --download-metis >>> >>> and the installation didn't give any problems. >>> >>> Could you help me understand why metis is not working? >>> >>> Thank you in advance, >>> Victoria >>> >>> Error: >>> >>> ****** ANALYSIS STEP ******** >>> ** Maximum transversal (ICNTL(6)) not allowed because matrix is distributed >>> Processing a graph of size: 699150 with 69238690 edges >>> Ordering based on METIS >>> 510522 37081376 [100] [10486 699150] >>> Error! Unknown CType: -1 >> >
