I already done VecAssemblyBegin/End(). However, only mpirun case these outputs are represented. There are more error outputs as below. -------------------------------------------------------------------------- MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_SELF with errorcode 73.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. You may or may not see output from other processes, depending on exactly when Open MPI kills them. -------------------------------------------------------------------------- [ubuntu:02473] PMIX ERROR: UNREACHABLE in file ../../../src/server/pmix_server.c at line 2193 [ubuntu:02473] PMIX ERROR: UNREACHABLE in file ../../../src/server/pmix_server.c at line 2193 [ubuntu:02473] PMIX ERROR: UNREACHABLE in file ../../../src/server/pmix_server.c at line 2193 [ubuntu:02473] 3 more processes have sent help message help-mpi-api.txt / mpi-abort [ubuntu:02473] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages Could this be the cause of the former petsc error?? Thanks, Hyung Kim 2022년 12월 6일 (화) 오후 10:58, Matthew Knepley <[email protected]>님이 작성: > On Tue, Dec 6, 2022 at 6:45 AM 김성익 <[email protected]> wrote: > >> Hello, >> >> >> There is a code which can run in not mpirun and also it can run in >> mpi_linear_solver_server. >> However, it has an error in just mpirun case such as mpirun -np ./program. >> The error output is as below. >> [0]PETSC ERROR: --------------------- Error Message >> -------------------------------------------------------------- >> [0]PETSC ERROR: Object is in wrong state >> [0]PETSC ERROR: Not for unassembled vector >> [0]PETSC ERROR: See https://petsc.org/release/faq/ for trouble shooting. >> [0]PETSC ERROR: Petsc Release Version 3.18.1, unknown >> [0]PETSC ERROR: ./app on a arch-linux-c-debug named ubuntu by ksi2443 Tue >> Dec 6 03:39:13 2022 >> [0]PETSC ERROR: Configure options -download-mumps -download-scalapack >> -download-parmetis -download-metis >> [0]PETSC ERROR: #1 VecCopy() at >> /home/ksi2443/petsc/src/vec/vec/interface/vector.c:1625 >> [0]PETSC ERROR: #2 KSPInitialResidual() at >> /home/ksi2443/petsc/src/ksp/ksp/interface/itres.c:60 >> [0]PETSC ERROR: #3 KSPSolve_GMRES() at >> /home/ksi2443/petsc/src/ksp/ksp/impls/gmres/gmres.c:227 >> [0]PETSC ERROR: #4 KSPSolve_Private() at >> /home/ksi2443/petsc/src/ksp/ksp/interface/itfunc.c:899 >> [0]PETSC ERROR: #5 KSPSolve() at >> /home/ksi2443/petsc/src/ksp/ksp/interface/itfunc.c:1071 >> [0]PETSC ERROR: #6 main() at /home/ksi2443/Downloads/coding/a1.c:450 >> [0]PETSC ERROR: No PETSc Option Table entries >> [0]PETSC ERROR: ----------------End of Error Message -------send entire >> error message to [email protected] >> >> How can I fix this?? >> > > It looks like we do not check the assembled state in parallel, since it > cannot cause a problem, but every time you > update values with VecSetValues(), you should call VecAssemblyBegin/End(). > > Thanks > > Matt > > >> Thanks, >> Hyung Kim >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > > https://www.cse.buffalo.edu/~knepley/ > <http://www.cse.buffalo.edu/~knepley/> >
