Nope. There is a back trace:
** thread #1: tid = 0x3b477b4, 0x00007fffb306cd42 libsystem_kernel.dylib`__pthread_kill + 10, queue = 'com.apple.main-thread', stop reason = signal SIGABRT * frame #0: 0x00007fffb306cd42 libsystem_kernel.dylib`__pthread_kill + 10 frame #1: 0x00007fffb315a457 libsystem_pthread.dylib`pthread_kill + 90 frame #2: 0x00007fffb2fd2420 libsystem_c.dylib`abort + 129 frame #3: 0x00000001057ff30a libpetsc.3.07.dylib`Petsc_MPI_AbortOnError(comm=<unavailable>, flag=<unavailable>) + 26 at init.c:185 [opt] frame #4: 0x0000000106bd3245 libpmpi.12.dylib`MPIR_Err_return_comm + 533 frame #5: 0x00000001068defd4 libmpi.12.dylib`MPI_Comm_create + 3492 frame #6: 0x00000001061345d9 libpetsc.3.07.dylib`hypre_GenerateSubComm(comm=-1006627852, participate=<unavailable>, new_comm_ptr=<unavailable>) + 409 at gen_redcs_mat.c:531 [opt] frame #7: 0x000000010618f8ba libpetsc.3.07.dylib`hypre_GaussElimSetup(amg_data=0x00007fe7ff857a00, level=<unavailable>, relax_type=9) + 74 at par_relax.c:4209 [opt] frame #8: 0x0000000106140e93 libpetsc.3.07.dylib`hypre_BoomerAMGSetup(amg_vdata=<unavailable>, A=0x00007fe80842aff0, f=0x00007fe80842a980, u=0x00007fe80842a510) + 17699 at par_amg_setup.c:2108 [opt] frame #9: 0x0000000105ec773c libpetsc.3.07.dylib`PCSetUp_HYPRE(pc=<unavailable>) + 2540 at hypre.c:226 [opt] frame #10: 0x0000000105eea68d libpetsc.3.07.dylib`PCSetUp(pc=0x00007fe805553f50) + 797 at precon.c:968 [opt] frame #11: 0x0000000105ee9fe5 libpetsc.3.07.dylib`PCApply(pc=0x00007fe805553f50, x=0x00007fe80052d420, y=0x00007fe800522c20) + 181 at precon.c:478 [opt] frame #12: 0x00000001015cf218 libmesh_opt.0.dylib`libMesh::PetscPreconditioner<double>::apply(libMesh::NumericVector<double> const&, libMesh::NumericVector<double>&) + 24 frame #13: 0x00000001009c7998 libmoose-opt.0.dylib`PhysicsBasedPreconditioner::apply(libMesh::NumericVector<double> const&, libMesh::NumericVector<double>&) + 520 frame #14: 0x00000001016ad701 libmesh_opt.0.dylib`libmesh_petsc_preconditioner_apply + 129 frame #15: 0x0000000105e7e715 libpetsc.3.07.dylib`PCApply_Shell(pc=0x00007fe8052623f0, x=0x00007fe806805a20, y=0x00007fe806805420) + 117 at shellpc.c:123 [opt] frame #16: 0x0000000105eea079 libpetsc.3.07.dylib`PCApply(pc=0x00007fe8052623f0, x=0x00007fe806805a20, y=0x00007fe806805420) + 329 at precon.c:482 [opt] frame #17: 0x0000000105eeb611 libpetsc.3.07.dylib`PCApplyBAorAB(pc=0x00007fe8052623f0, side=PC_RIGHT, x=0x00007fe806805a20, y=0x00007fe806806020, work=0x00007fe806805420) + 945 at precon.c:714 [opt] frame #18: 0x0000000105f31658 libpetsc.3.07.dylib`KSPGMRESCycle [inlined] KSP_PCApplyBAorAB(ksp=0x00007fe806022220, x=<unavailable>, y=0x00007fe806806020, w=<unavailable>) + 191 at kspimpl.h:295 [opt] frame #19: 0x0000000105f31599 libpetsc.3.07.dylib`KSPGMRESCycle(itcount=<unavailable>, ksp=<unavailable>) + 553 at gmres.c:156 [opt] frame #20: 0x0000000105f326bd libpetsc.3.07.dylib`KSPSolve_GMRES(ksp=<unavailable>) + 221 at gmres.c:240 [opt] frame #21: 0x0000000105f5f671 libpetsc.3.07.dylib`KSPSolve(ksp=0x00007fe806022220, b=0x00007fe7fd946220, x=<unavailable>) + 1345 at itfunc.c:677 [opt] frame #22: 0x0000000105fd0251 libpetsc.3.07.dylib`SNESSolve_NEWTONLS(snes=<unavailable>) + 1425 at ls.c:230 [opt] frame #23: 0x0000000105fa10ca libpetsc.3.07.dylib`SNESSolve(snes=<unavailable>, b=<unavailable>, x=0x00007fe7fd865e20) + 858 at snes.c:4128 [opt] frame #24: 0x00000001016b63c3 libmesh_opt.0.dylib`libMesh::PetscNonlinearSolver<double>::solve(libMesh::SparseMatrix<double>&, libMesh::NumericVector<double>&, libMesh::NumericVector<double>&, double, unsigned int) + 835 frame #25: 0x00000001016fc244 libmesh_opt.0.dylib`libMesh::NonlinearImplicitSystem::solve() + 324 frame #26: 0x0000000100a71dc8 libmoose-opt.0.dylib`NonlinearSystem::solve() + 472 frame #27: 0x00000001009fe815 libmoose-opt.0.dylib`FEProblemBase::solve() + 117 frame #28: 0x0000000100761fba libmoose-opt.0.dylib`Steady::execute() + 266 frame #29: 0x0000000100b78ac3 libmoose-opt.0.dylib`MooseApp::run() + 259 frame #30: 0x00000001003843aa moose_test-opt`main + 122 frame #31: 0x00007fffb2f3e235 libdyld.dylib`start + 1* Fande, On Mon, Apr 2, 2018 at 4:02 PM, Stefano Zampini <stefano.zamp...@gmail.com> wrote: > maybe this will fix ? > > > *diff --git a/src/ksp/pc/impls/hypre/hypre.c > b/src/ksp/pc/impls/hypre/hypre.c* > > *index 28addcf533..6a756d4c57 100644* > > *--- a/src/ksp/pc/impls/hypre/hypre.c* > > *+++ b/src/ksp/pc/impls/hypre/hypre.c* > > @@ -142,8 +142,7 @@ static PetscErrorCode PCSetUp_HYPRE(PC pc) > > > > ierr = PetscObjectTypeCompare((PetscObject)pc->pmat,MATHYPRE, > &ishypre);CHKERRQ(ierr); > > if (!ishypre) { > > - ierr = MatDestroy(&jac->hpmat);CHKERRQ(ierr); > > - ierr = MatConvert(pc->pmat,MATHYPRE,MAT_INITIAL_MATRIX,&jac-> > hpmat);CHKERRQ(ierr); > > + ierr = MatConvert(pc->pmat,MATHYPRE,jac->hpmat ? MAT_REUSE_MATRIX : > MAT_INITIAL_MATRIX,&jac->hpmat);CHKERRQ(ierr); > > } else { > > ierr = PetscObjectReference((PetscObject)pc->pmat);CHKERRQ(ierr); > > ierr = MatDestroy(&jac->hpmat);CHKERRQ(ierr); > > > > 2018-04-02 23:46 GMT+02:00 Kong, Fande <fande.k...@inl.gov>: > >> Hi All, >> >> I am trying to upgrade PETSc from 3.7.6 to 3.8.3 for MOOSE and its >> applications. I have a error message for a standard test: >> >> >> >> >> >> >> >> >> >> *preconditioners/pbp.lots_of_variables: MPI had an >> errorpreconditioners/pbp.lots_of_variables: >> ------------------------------------------------preconditioners/pbp.lots_of_variables: >> Other MPI error, error stack:preconditioners/pbp.lots_of_variables: >> PMPI_Comm_dup(177)..................: MPI_Comm_dup(comm=0x84000001, >> new_comm=0x97d1068) failedpreconditioners/pbp.lots_of_variables: >> PMPI_Comm_dup(162)..................: >> preconditioners/pbp.lots_of_variables: >> MPIR_Comm_dup_impl(57)..............: >> preconditioners/pbp.lots_of_variables: >> MPIR_Comm_copy(739).................: >> preconditioners/pbp.lots_of_variables: >> MPIR_Get_contextid_sparse_group(614): Too many communicators (0/2048 free >> on this process; ignore_id=0)* >> >> >> I did "git bisect', and the following commit introduces this issue: >> >> >> >> >> >> >> >> >> *commit 49a781f5cee36db85e8d5b951eec29f10ac13593Author: Stefano Zampini >> <stefano.zamp...@gmail.com <stefano.zamp...@gmail.com>>Date: Sat Nov 5 >> 20:15:19 2016 +0300 PCHYPRE: use internal Mat of type MatHYPRE >> hpmat already stores two HYPRE vectors* >> >> Before I debug line-by-line, anyone has a clue on this? >> >> >> Fande, >> > > > > -- > Stefano >