Perhaps you have done enough memory preallocation for the multiprocessors and it is going very--very slowly with memory allocations?
Barry On Oct 12, 2009, at 8:03 AM, Andreas Grassl wrote: > Hello, > > I'm trying to work with MUMPS-Cholesky as direct solver and CG as > iterative > solver and have problems with SBAIJ-Matrices. > > I have a MatSetValues-line not regarding the symmetry but force it > by the > command line option -mat_ignore_lower_triangular. > > The problems show up at "bigger" problems (~450000 DOF) with "more" > processes (1 > and 2 work fine, 4 gives a hangup) and/or trying to save the matrix. > The machine > has 4 cores and 12 GB memory. NNZ per row is ~80. > > Attached you find the output of the program run with -info. > > Any hints where to search? > > Cheers, > > ando > > -- > /"\ Grassl Andreas > \ / ASCII Ribbon Campaign Uni Innsbruck Institut f. Mathematik > X against HTML email Technikerstr. 13 Zi 709 > / \ +43 (0)512 507 6091 > same call as 2procinfo without -ksp_info_binary > > mpiexec -np 2 externalsolver > [0] PetscInitialize(): PETSc successfully started: number of > processors = 2 > [1] PetscInitialize(): PETSc successfully started: number of > processors = 2 > [1] PetscInitialize(): Running on machine: mat1.uibk.ac.at > [0] PetscInitialize(): Running on machine: mat1.uibk.ac.at > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [0] PetscCommDuplicate(): returning tag 2147483647 > [1] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [1] PetscCommDuplicate(): returning tag 2147483647 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483646 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483646 > [0] MatSetUpPreallocation(): Warning not preallocating matrix storage > [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 > -2080374783 max tags = 2147483647 > [0] PetscCommDuplicate(): returning tag 2147483647 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): Duplicating a communicator 1140850689 > -2080374783 max tags = 2147483647 > [1] PetscCommDuplicate(): returning tag 2147483647 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483646 > [0] PetscCommDuplicate(): returning tag 2147483646 > [0] MatSetOption_MPISBAIJ(): Option SYMMETRIC ignored > [0] MatStashScatterBegin_Private(): No of messages: 1 > [0] MatStashScatterBegin_Private(): Mesg_to: 1: size: 30544064 > [0] MatStashScatterBegin_Private(): No of messages: 0 > [0] MatAssemblyBegin_MPISBAIJ(): Stash has 3818007 entries,uses 8 > mallocs. > [0] MatAssemblyBegin_MPISBAIJ(): Block-Stash has 3818007 entries, > uses 8 mallocs. > [0] MatAssemblyEnd_SeqSBAIJ(): Matrix size: 221545 X 221545, block > size 1; storage space: 16676676 unneeded, 5477824 used > [0] MatAssemblyEnd_SeqSBAIJ(): Number of mallocs during MatSetValues > is 0 > [0] MatAssemblyEnd_SeqSBAIJ(): Most nonzeros blocks in any row is 62 > [1] MatAssemblyEnd_SeqSBAIJ(): Matrix size: 225367 X 225367, block > size 1; storage space: 16793278 unneeded, 5743422 used > [1] MatAssemblyEnd_SeqSBAIJ(): Number of mallocs during MatSetValues > is 0 > [1] MatAssemblyEnd_SeqSBAIJ(): Most nonzeros blocks in any row is 65 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483645 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483644 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483643 > [1] PetscCommDuplicate(): returning tag 2147483639 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483645 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483644 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483643 > [0] PetscCommDuplicate(): returning tag 2147483639 > [0] PetscCommDuplicate(): returning tag 2147483634 > [1] PetscCommDuplicate(): returning tag 2147483634 > [1] PetscCommDuplicate(): returning tag 2147483630 > [1] PetscCommDuplicate(): returning tag 2147483625 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483642 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483641 > [1] PetscCommDuplicate(): returning tag 2147483620 > [0] VecScatterCreateCommon_PtoS(): Using blocksize 1 scatter > [0] VecScatterCreate(): General case: MPI to Seq > [0] PetscCommDuplicate(): returning tag 2147483630 > [0] PetscCommDuplicate(): returning tag 2147483625 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483642 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483641 > [0] PetscCommDuplicate(): returning tag 2147483620 > [0] VecScatterCreateCommon_PtoS(): Using blocksize 1 scatter > [0] VecScatterCreate(): General case: MPI to MPI > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483640 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483639 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDupli[1] PetscCommDuplicate(): Using internal PETSc > communicator 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483640 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483639 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483638 > [1] MatAssemblyEnd_SeqBAIJ(): Matrix size: 225367 X 0, block size 1; > storage space: 22536700 unneeded, 0 used > [1] MatAssemblyEnd_SeqBAIJ(): Number of mallocs during MatSetValues > is 0 > [1] MatAssemblyEnd_SeqBAIJ(): Most nonzeros blocks in any row is 0 > [1] Mat_CheckCompressedRow(): Found the ratio (num_zerorows 225367)/ > (num_localrows 225367) > 0.6. Use CompressedRow routines. > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483614 > [1] PetscCommDuplicate(): returning tag 2147483613 > [1]before KSPsetup > [1] PetscCommDuplicate(): Using cate(): returning tag 2147483638 > [0] MatAssemblyEnd_SeqBAIJ(): Matrix size: 221545 X 6275, block size > 1; storage space: 22060506 unneeded, 93994 used > [0] MatAssemblyEnd_SeqBAIJ(): Number of mallocs during MatSetValues > is 0 > [0] MatAssemblyEnd_SeqBAIJ(): Most nonzeros blocks in any row is 32 > [0] Mat_CheckCompressedRow(): Found the ratio (num_zerorows 216534)/ > (num_localrows 221545) > 0.6. Use CompressedRow routines. > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483614 > [0] PetscCommDuplicate(): returning tag 2147483613 > [0]before KSPsetup > [0] PCSetUp(): Setting up new PC > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483637 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483636 > [0] PetscCommDuplicate(): returning tag 2147483612 > [0] PetscCommDuplicate(): Using internainternal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483637 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483636 > [1] PetscCommDuplicate(): returning tag 2147483612 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483635 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483634 > l PETSc communicator 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483635 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483634 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483633 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483632 > [1] PetscCommDuplicate(): returning tag 2147483607 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483633 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483632 > [0] PetscCommDuplicate(): returning tag 2147483607 > [0] PetscCommDuplicate(): returning tag 2147483602 > [0] VecScatterCreate(): Special case: processor zero gets entire > parallel vector, rest get none > [1] PetscCommDuplicate(): returning tag 2147483602 > > DMUMPS 4.8.4 > L D L^T Solver for symmetric positive definite matrices > Type of parallelism: Working host > > ****** ANALYSIS STEP ******** > > Density: NBdense, Average, Median = 0 49 50 > Ordering based on METIS > ** Peak of sequential stack size (number of real entries) : > 16763905. > A root of estimated size 2965 has been selected for > Scalapack. > > Leaving analysis phase with ... > INFOG(1) = 0 > INFOG(2) = 0 > -- (20) Number of entries in factors (estim.) = 120657071 > -- (3) Storage of factors (REAL, estimated) = 137364268 > -- (4) Storage of factors (INT , estimated) = 6138804 > -- (5) Maximum frontal size (estimated) = 3705 > -- (6) Number of nodes in the tree = 21863 > -- (7) Ordering option effectively used = 5 > ICNTL(6) Maximum transversal option = 0 > ICNTL(7) Pivot order option = 7 > Percentage of memory relaxation (effective) = 200 > Number of level 2 nodes = 1 > Number of split nodes = 0 > RINFO(1) Operations during elimination (estim) = 1.114D+11 > Distributed matrix entry format (ICNTL(18)) = 3 > ** Rank of proc needing largest memory in IC facto : 0 > ** Estimated corresponding MBYTES for IC facto : 2085 > ** Estimated avg. MBYTES per work. proc at facto (IC) : 2044 > ** TOTAL space in MBYTES for IC factorization : 4088 > ** Rank of proc needing largest memory for OOC facto : 1 > ** Estimated corresponding MBYTES for OOC facto : 644 > ** Estimated avg. MBYTES per work. proc at facto (OOC) : 630 > ** TOTAL space in MBYTES for OOC factorization : 1261 > > ****** FACTORIZATION STEP ******** > > > GLOBAL STATISTICS PRIOR NUMERICAL FACTORIZATION ... > NUMBER OF WORKING PROCESSES = 2 > REAL SPACE FOR FACTORS = 137364268 > INTEGER SPACE FOR FACTORS = 6138804 > MAXIMUM FRONTAL SIZE (ESTIMATED) = 3705 > NUMBER OF NODES IN THE TREE = 21863 > Maximum effective relaxed size of S = 239060000 > Average effective relaxed size of S = 234725641 > > REDISTRIB: TOTAL DATA LOCAL/SENT = 9086109 2238491 > GLOBAL TIME FOR MATRIX DISTRIBUTION = 0.7322 > ** Memory relaxation parameter ( ICNTL(14) ) : 200 > ** Rank of processor needing largest memory in facto : 0 > ** Space in MBYTES used by this processor for facto : 2085 > ** Avg. Space in MBYTES per working proc during facto : 2044 > > ELAPSED TIME FOR FACTORIZATION = 28.1724 > Maximum effective space used in S (KEEP(67) = 77516871 > Average effective space used in S (KEEP(67) = 75395654 > ** EFF Min: Rank of processor needing largest memory : 0 > ** EFF Min: Space in MBYTES used by this processor : 741 > ** EFF Min: Avg. Space in MBYTES per working proc : 720 > > GLOBAL STATISTICS > RINFOG(2) OPERATIONS DURING NODE ASSEMBLY = 2.565D+08 > ------(3) OPERATIONS DURING NODE ELIMINATION = 1.114D+11 > INFOG (9) REAL SPACE FOR FACTORS = 137364268 > INFOG(10) INTEGER SPACE FOR FACTORS = 6138809 > INFOG(11) MAXIMUM FRONT SIZE = 3705 > INFOG(29) NUMBER OF ENTRIES IN FACTORS = 116259976 > INFOG(12) NB OF NEGATIVE PIVOTS = 0 > INFOG(14) NUMBER OF MEMORY COMPRESS = 0 > [0]after KSPsetup > [0]RHS setup > [1]after KSPsetup > [1]RHS setup > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483601 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483601 > [0] PetscCommDuplicate(): returning tag 2147483596 > [1] PetscCommDuplicate(): returning tag 2147483596 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483591 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483591 > [0] VecAssemblyBegin_MPI(): Stash has 225367 entries, uses 12 mallocs. > [0] VecAssemblyBegin_MPI(): Block-Stash has 0 entries, uses 0 mallocs. > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483631 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483631 > > > ****** SOLVE & CHECK STEP ******** > > > STATISTICS PRIOR SOLVE PHASE ........... > NUMBER OF RIGHT-HAND-SIDES = 1 > BLOCKING FACTOR FOR MULTIPLE RHS = 1 > ICNTL (9) = 1 > --- (10) = 0 > --- (11) = 0 > --- (20) = 0 > --- (21) = 1 > ** Rank of processor needing largest memory in solve : 0 > ** Space in MBYTES used by this processor for solve : 2013 > ** Avg. Space in MBYTES per working proc during solve : 1975 > > > LEAVING SOLVER WITH: INFOG(1) ............ = 0 > INFOG(2) ............ = 0 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483630 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483629 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483630 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483629 > [0] PetscCommDuplicate(): returning tag 2147483590 > [1] PetscCommDuplicate(): returning tag 2147483590 > [1] VecScatterCreate(): General case: Seq to MPI > [0] VecScatterCreateCommon_PtoS(): Using blocksize 1 scatter > [0] VecScatterCreate(): General case: Seq to MPI > [0] PetscCommDuplicate(): returning tag 2147483586 > KSP Object: > type: preonly > maximum iterations=10000, initial guess is zero > tolerances: relative=1e-06, absolute=1e-50, divergence=10000 > left preconditioning > PC Object: > type: cholesky > Cholesky: out-of-place factorization > matrix ordering: natural > Cholesky: factor fill ratio needed 0 > Factored matrix follows > Matrix Object: > type=mpisbaij, rows=446912, cols=446912 > package used to perform factorization: mumps > [1] PetscCommDuplicate(): returning tag 2147483586 > [1] PetscCommDuplicate(): returning tag 2147483585 > [1] PetscCommDuplicate(): returning tag 2147483584 > [1] PetscCommDuplicate(): returning tag 2147483583 > [1] PetscCommDuplicate(): returning tag 2147483582 > [1] PetscCommDuplicate(): returning tag 2147483581 > [1] PetscCommDuplicate(): returning tag 2147483580 > total: nonzeros=0, allocated nonzeros=893824 > MUMPS run parameters: > SYM (matrix type): 1 > PAR (host participation): 1 > ICNTL(1) (output for error): 6 > ICNTL(2) (output of diagnostic msg):0 > ICNTL(3) (output for global info): 6 > ICNTL(4) (level of printing): -1 > ICNTL(5) (input mat struct): 0 > ICNTL(6) (matrix prescaling): 0 > ICNTL(7) (matrix ordering): 7 > ICNTL(8) (scalling strategy): 77 > ICNTL(9) (A/A^T x=b is solved): 1 > ICNTL(10) (max num of refinements): 0 > ICNTL(11) (error analysis): 0 > ICNTL(12) (efficiency control): 1 > ICNTL(13) (efficiency control): 0 > ICNTL(14) (percentage of estimated workspace increase): > 200 > ICNTL(18)[1] PetscCommDuplicate(): returning tag > 2147483579 > (input mat struct): 3 > ICNTL(19) (Shur complement info): 0 > ICNTL(20) (rhs sparse pattern): 0 > ICNTL(21) (solution struct): 1 > ICNTL(22) (in-core/out-of-core facility): 0 > ICNTL(23) (max size of memory can be allocated locally):0 > ICNTL(24) (detection of null pivot rows): 0 > ICNTL(25) (computation of a null space basis): 0 > ICNTL(26) (Schur options for rhs or solution): 0 > ICNTL(27) (experimental parameter): > -8 > CNTL(1) (relative pivoting threshold): 0 > CNTL(2) (stopping criterion of refinement): 1.49012e-08 > CNTL(3) (absolute pivoting threshold): 0 > CNTL(4) (value of static pivoting): -1 > CNTL(5) (fixation for null pivots): 0 > RINFO(1) (local estimated flops for the elimination after > analysis): > [0] 5.16132e+10 > [0] PetscCommDuplicate(): returning tag 2147483585 > [1] 5.97957e+10 > RINFO(2) (local estimated flops for the assembly after > factorization): > [0] 1.33366e+08 > [0] PetscCommDuplicate(): returning tag 2147483584 > [1] 1.23105e+08 > RINFO(3) (local estimated flops for the elimination after > factorization): > [0] 5.16124e+10 > [0] PetscCommDuplicate(): returning tag 2147483583 > [1] 5.9795e+10 > INFO(15) (estimated size of (in MB) MUMPS internal data for > running numerical factorization): > [0] 2085 > [0] PetscCommDuplicate(): returning tag 2147483582 > [1] 2003 > INFO(16) (size of (in MB) MUMPS internal data used during > numerical factorization): > [0] 2085 > [0] PetscCommDuplicate(): returning tag 2147483581 > [1] 2003 > INFO(23) (num of pivots eliminated on this processor after > factorization): > [0] 236610 > [0] PetscCommDuplicate(): returning tag 2147483580 > [1] 210302 > RINFOG(1) (global estimated flops for the elimination > after analysis): 1.11409e+11 > RINFOG(2) (global estimated flops for the assembly > after factorization): 2.56471e+08 > RINFOG(3) (global estimated flops for the elimination > after factorization): 1.11407e+11 > INFOG(3) (estimated real workspace for factors on all > processors after analysis): 137364268 > INFOG(4) (estimated integer workspace for factors on > all processors after analysis): 6138804 > INFOG(5) (estimated maximum front size in the complete > tree): 3705 > INFOG(6) (number of nodes in the complete tree): 21863 > INFOG(7) (ordering option effectively uese after > analysis): 5 > INFOG(8) (structural symmetry in percent of the > permuted matrix after analysis): 100 > INFOG(9) (total real/complex workspace to store the > matrix factors after factorization): 137364268 > INFOG(10) (total integer space store the matrix factors > after factorization): 6138809 > INFOG(11) (order of largest frontal matrix after > factorization): 3705 > INFOG(12) (number of off-diagonal pivots): 0 > INFOG(13) (number of delayed pivots after > factorization): 0 > INFOG(14) (number of memory compress after > factorization): 0 > INFOG(15) (number of steps of iterative refinement > after solution): 0 > INFOG(16) (estimated size (in MB) of all MUMPS internal > data for factorization after analysis: value on the most memory > consuming processor): 2085 > INFOG(17) (estimated size of all MUMPS internal data > for factorization after analysis: sum over all processors): 4088 > INFOG(18) (size of all MUMPS internal data allocated > during factorization: value on the most memory consuming processor): > 2085 > INFOG(19) (size of all MUMPS internal data allocated > during factorization: sum over all processors): 4088 > INFOG(20) (estimated number of entries in the factors): > 120657071 > INFOG(21) (size in MB of memory effectively used during > factorization - value on the most memory consuming processor): 741 > INFOG(22) (size in MB of memory effectively used during > factorization - sum over all processors): 1441 > INFOG(23) (after analysis: value of ICNTL(6) > effectively used): 0 > INFOG(24) (after analysis: value of ICNTL(12) > effectively used): 1 > INFOG(25) (after factorization: number of pivots > modified by static pivoting): 0 > linear system matrix = precond matrix: > Matrix Object: > type=mpisbaij, rows=446912, cols=446912 > total: nonzeros=11315240, allocated nonzeros=89382400 > block size is 1 > [0] PetscCommDuplicate(): returning tag 2147483579 > [1] PetscCommDuplicate(): returning tag 2147483578 > [1] PetscCommDuplicate(): returning tag 2147483577 > [1] Petsc_DelViewer(): Deleting viewer data in an MPI_Comm -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483578 > [0] PetscCommDuplicate(): returning tag 2147483577 > [0] Petsc_DelViewer(): Deleting viewer data in an MPI_Comm -2080374784 > [0] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm 1140850689 > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374783 > [0] Petsc_DelTag(): Deleting tag data in an MPI_Comm -2080374783 > [0] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483576 > [1] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm 1140850689 > [1] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374783 > [1] Petsc_DelTag(): Deleting tag data in an MPI_Comm -2080374783 > [0] Petsc_DelViewer(): Deleting viewer data in an MPI_Comm -2080374784 > [0] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm 1140850688 > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > [0] Petsc_DelTag(): Deleting tag data in an MPI_Comm -2080374784 > [0] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm -2080374784 > [0] Petsc_DelViewer(): Deleting viewer data in an MPI_Comm -2080374784 > ************************************************************************************************************************ > *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript - > r -fCourier9' to print this document *** > ************************************************************************************************************************ > > ---------------------------------------------- PETSc Performance > Summary: ---------------------------------------------- > > [1] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483576 > [1] Petsc_DelViewer(): Deleting viewer data in an MPI_Comm -2080374784 > [1] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm 1140850688 > [1] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > [1] Petsc_DelTag(): Deleting tag data in an MPI_Comm -2080374784 > [1] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm -2080374784 > [1] Petsc_DelViewer(): Deleting viewer data in an MPI_Comm -2080374784 > externalsolver on a linux32-i named mat1.uibk.ac.at with 2 > processors, by csae1801 Mon Oct 12 12:08:38 2009 > Using Petsc Release Version 3.0.0, Patch 8, Fri Aug 21 14:02:12 CDT > 2009 > > Max Max/Min Avg Total > Time (sec): 3.989e+01 1.00000 3.989e+01 > Objects: 3.600e+01 1.00000 3.600e+01 > Flops: 0.000e+00 0.00000 0.000e+00 0.000e+00 > Flops/sec: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Messages: 2.550e+01 1.08511 2.450e+01 4.900e+01 > MPI Message Lengths: 6.596e+07 1.00053 2.692e+06 1.319e+08 > MPI Reductions: 5.800e+01 1.00000 > > Flop counting convention: 1 flop = 1 real number operation of type > (multiply/divide/add/subtract) > e.g., VecAXPY() for real vectors of > length N --> 2N flops > and VecAXPY() for complex vectors of > length N --> 8N flops > > Summary of Stages: ----- Time ------ ----- Flops ----- --- > Messages --- -- Message Lengths -- -- Reductions -- > Avg %Total Avg %Total counts > %Total Avg %Total counts %Total > 0: Main Stage: 2.8402e-01 0.7% 0.0000e+00 0.0% 1.000e > +00 2.0% 8.163e-02 0.0% 0.000e+00 0.0% > 1: decomposition: 4.7451e-01 1.2% 0.0000e+00 0.0% 0.000e > +00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% > 2: Matrix Create: 7.2766e-01 1.8% 0.0000e+00 0.0% 0.000e > +00 0.0% 0.000e+00 0.0% 2.000e+00 3.4% > 3: Matrix fill: 2.5720e+00 6.4% 0.0000e+00 0.0% 1.600e+01 > 32.7% 2.486e+06 92.4% 1.500e+01 25.9% > 4: Solver setup: 3.2841e+01 82.3% 0.0000e+00 0.0% 0.000e > +00 0.0% 0.000e+00 0.0% 8.000e+00 13.8% > 5: RHS setup: 3.2125e-02 0.1% 0.0000e+00 0.0% 4.000e > +00 8.2% 1.094e+05 4.1% 4.000e+00 6.9% > 6: Solve: 4.9751e-01 1.2% 0.0000e+00 0.0% 2.500e+01 > 51.0% 5.944e+04 2.2% 6.000e+00 10.3% > 7: Postproc: 2.4653e+00 6.2% 0.0000e+00 0.0% 3.000e > +00 6.1% 3.679e+04 1.4% 0.000e+00 0.0% > > ------------------------------------------------------------------------------------------------------------------------ > See the 'Profiling' chapter of the users' manual for details on > interpreting output. > Phase summary info: > Count: number of times phase was executed > Time and Flops: Max - maximum over all processors > Ratio - ratio of maximum to minimum over all > processors > Mess: number of messages sent > Avg. len: average message length > Reduct: number of global reductions > Global: entire computation > Stage: stages of a computation. Set stages with > PetscLogStagePush() and PetscLogStagePop(). > %T - percent time in this phase %F - percent flops in > this phase > %M - percent messages in this phase %L - percent message > lengths in this phase > %R - percent reductions in this phase > Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max > time over all processors) > ------------------------------------------------------------------------------------------------------------------------ > Event Count Time (sec) > Flops --- Global --- --- Stage --- > Total > Max Ratio Max Ratio Max Ratio Mess Avg > len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s > ------------------------------------------------------------------------------------------------------------------------ > > --- Event Stage 0: Main Stage > > PetscBarrier 1 1.0 3.1269e-0328.4 0.00e+00 0.0 0.0e+00 0.0e > +00 0.0e+00 0 0 0 0 0 1 0 0 0 0 0 > > --- Event Stage 1: decomposition > > > --- Event Stage 2: Matrix Create > > > --- Event Stage 3: Matrix fill > > MatAssemblyBegin 1 1.0 3.3235e-01 1.9 0.00e+00 0.0 6.0e+00 2.0e > +07 3.0e+00 1 0 12 92 5 10 0 38100 20 0 > MatAssemblyEnd 1 1.0 1.0536e+00 1.0 0.00e+00 0.0 1.0e+01 1.8e > +04 1.2e+01 3 0 20 0 21 41 0 62 0 80 0 > > --- Event Stage 4: Solver setup > > MatCholFctrSym 1 1.0 4.0531e-06 2.1 0.00e+00 0.0 0.0e+00 0.0e > +00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatCholFctrNum 1 1.0 3.2830e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e > +00 5.0e+00 82 0 0 0 9 100 0 0 0 62 0 > KSPSetup 1 1.0 6.9141e-06 1.4 0.00e+00 0.0 0.0e+00 0.0e > +00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > PCSetUp 1 1.0 3.2841e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e > +00 8.0e+00 82 0 0 0 14 100 0 0 0100 0 > > --- Event Stage 5: RHS setup > > VecAssemblyBegin 1 1.0 6.2160e-03 1.6 0.00e+00 0.0 4.0e+00 1.3e > +06 3.0e+00 0 0 8 4 5 16 0100100 75 0 > VecAssemblyEnd 1 1.0 8.4589e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e > +00 0.0e+00 0 0 0 0 0 26 0 0 0 0 0 > > --- Event Stage 6: Solve > > MatSolve 1 1.0 4.9525e-01 1.0 0.00e+00 0.0 6.0e+00 4.8e > +05 4.0e+00 1 0 12 2 7 100 0 24 98 67 0 > MatView 2 1.0 1.4269e-03 1.0 0.00e+00 0.0 1.8e+01 2.7e > +03 2.0e+00 0 0 37 0 3 0 0 72 2 33 0 > VecSet 1 1.0 8.1515e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e > +00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecScatterBegin 2 1.0 6.7840e-03 1.3 0.00e+00 0.0 2.0e+00 1.3e > +06 0.0e+00 0 0 4 2 0 1 0 8 86 0 0 > VecScatterEnd 1 1.0 2.1610e-03 1.7 0.00e+00 0.0 0.0e+00 0.0e > +00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > KSPSolve 1 1.0 4.9603e-01 1.0 0.00e+00 0.0 6.0e+00 4.8e > +05 4.0e+00 1 0 12 2 7 100 0 24 98 67 0 > PCApply 1 1.0 4.9526e-01 1.0 0.00e+00 0.0 6.0e+00 4.8e > +05 4.0e+00 1 0 12 2 7 100 0 24 98 67 0 > > --- Event Stage 7: Postproc > > VecView 1 1.0 2.1809e+00 2.0 0.00e+00 0.0 2.0e+00 9.0e > +05 0.0e+00 4 0 4 1 0 66 0 67100 0 0 > ------------------------------------------------------------------------------------------------------------------------ > > Memory usage is given in bytes: > > Object Type Creations Destructions Memory Descendants' > Mem. > > --- Event Stage 0: Main Stage > > IS L to G Mapping 0 1 889756 0 > Matrix 0 3 547690752 0 > Vec 0 6 3702888 0 > Vec Scatter 0 2 1736 0 > Viewer 0 1 544 0 > > --- Event Stage 1: decomposition > > > --- Event Stage 2: Matrix Create > > IS L to G Mapping 1 0 0 0 > Matrix 3 0 0 0 > > --- Event Stage 3: Matrix fill > > Index Set 4 4 152680 0 > Vec 7 1 1304 0 > Vec Scatter 2 0 0 0 > > --- Event Stage 4: Solver setup > > Index Set 3 2 1008 0 > Matrix 3 0 0 0 > Vec 2 1 1773664 0 > Vec Scatter 1 0 0 0 > Krylov Solver 1 0 0 0 > Preconditioner 1 0 0 0 > > --- Event Stage 5: RHS setup > > Vec 2 0 0 0 > Viewer 1 0 0 0 > > --- Event Stage 6: Solve > > Index Set 2 2 947456 0 > Vec 1 0 0 0 > Vec Scatter 1 0 0 0 > Viewer 1 0 0 0 > > --- Event Stage 7: Postproc > > Index Set 0 1 504 0 > Matrix 0 3 8870136 0 > Vec 0 4 7125104 0 > Vec Scatter 0 2 1320 0 > Krylov Solver 0 1 832 0 > Preconditioner 0 1 728 0 > Viewer 0 1 544 0 > = > = > = > = > = > = > = > = > = > = > = > = > = > = > = > =================================================================[1] > PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [1] PetscCommDuplicate(): returning tag 2147483647 > [1] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm 1140850688 > [1] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > [1] Petsc_DelTag(): Deleting tag data in an MPI_Comm -2080374784 > [1] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm -2080374784 > ======================================== > Average time to get PetscTime(): 1.3113e-06 > Average time for MPI_Barrier(): 0.000144005 > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [0] PetscCommDuplicate(): returning tag 2147483647 > Average time for zero size MPI_Send(): 6.54459e-05 > [0] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm 1140850688 > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > [0] Petsc_DelTag(): Deleting tag data in an MPI_Comm -2080374784 > [0] Petsc_DelComm(): Deleting PETSc communicator imbedded in a user > MPI_Comm -2080374784 > #PETSc Option Table entries: > -dump_data false > -info > -ksp_compute_singularvalues > -ksp_rtol 1e-6 > -ksp_type preonly > -ksp_view > -log_summary > -mat_ignore_lower_triangular > -mat_mumps_icntl_14 200 > -mat_mumps_sym 1 > -mat_type sbaij > -monitorname errormonitor.dat > -nz 100 > -options_left > -pc_factor_mat_solver_package mumps > -pc_type cholesky > -preload false > -verb false > #End o PETSc Option Table entries > Compiled without FORTRAN kernels > Compiled with full precision matrices (default) > sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 > sizeof(PetscScalar) 8 > Configure run at: Wed Oct 7 18:16:40 2009 > Configure options: --with-scalar-type=real --with-debugging=0 --with- > precision=double --with-cc=icc --with-fc=ifort --with-cxx=icpc -- > with-shared=0 --with-mpi=1 --with-external-packages-dir=/home/lux/ > csae1801/petsc/externalpackages --download-mpich=ifneeded --with- > scalapack=1 --download-scalapack=ifneeded --download-f-blas- > lapack=ifneeded --with-blacs=1 --download-blacs=ifneeded --with- > parmetis=1 --download-parmetis=ifneeded --with-mumps=1 --download- > mumps=ifneeded --with-hypre=1 --with-hypre-dir=/home/lux/csae1801/ > petsc/externalpackages/hypre-2.4.0b/src/hypre --with-spooles=1 -- > download-spooles=ifneeded --with-superlu_dist=1 --download- > superlu_dist=ifneeded PETSC_ARCH=linux32-intel-c-prod > ----------------------------------------- > Libraries compiled on Mi 7. Okt 18:17:41 CEST 2009 on mat1.uibk.ac.at > Machine characteristics: Linux mat1.uibk.ac.at 2.6.18-53.el5xen #1 > SMP Wed Oct 10 16:48:44 EDT 2007 x86_64 x86_64 x86_64 GNU/Linux > Using PETSc directory: /home/lux/csae1801/petsc/petsc-3.0.0-p8 > Using PETSc arch: linux32-intel-c-prod > ----------------------------------------- > Using C compiler: /home/lux/csae1801/petsc/petsc-3.0.0-p8/linux32- > intel-c-prod/bin/mpicc -O > Using Fortran compiler: /home/lux/csae1801/petsc/petsc-3.0.0-p8/ > linux32-intel-c-prod/bin/mpif90 -O > ----------------------------------------- > Using include paths: -I/home/lux/csae1801/petsc/petsc-3.0.0-p8/ > linux32-intel-c-prod/include -I/home/lux/csae1801/petsc/petsc-3.0.0- > p8/include -I/home/lux/csae1801/petsc/petsc-3.0.0-p8/linux32-intel-c- > prod/include -I/home/lux/csae1801/petsc/externalpackages/ > hypre-2.4.0b/src/hypre/include > ------------------------------------------ > Using C linker: /home/lux/csae1801/petsc/petsc-3.0.0-p8/linux32- > intel-c-prod/bin/mpicc -O > Using Fortran linker: /home/lux/csae1801/petsc/petsc-3.0.0-p8/ > linux32-intel-c-prod/bin/mpif90 -O > Using libraries: -Wl,-rpath,/home/lux/csae1801/petsc/petsc-3.0.0-p8/ > linux32-intel-c-prod/lib -L/home/lux/csae1801/petsc/petsc-3.0.0-p8/ > linux32-intel-c-prod/lib -lpetscts -lpetscsnes -lpetscksp -lpetscdm - > lpetscmat -lpetscvec -lpetsc -lX11 -Wl,-rpath,/home/lux/ > csae1801/petsc/petsc-3.0.0-p8/linux32-intel-c-prod/lib -L/home/lux/ > csae1801/petsc/petsc-3.0.0-p8/linux32-intel-c-prod/lib - > lsuperlu_dist_2.3 -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common > -lpord -lparmetis -lmetis -lscalapack -lblacs -Wl,-rpath,/home/lux/ > csae1801/petsc/externalpackages/hypre-2.4.0b/src/hypre/lib -L/home/ > lux/csae1801/petsc/externalpackages/hypre-2.4.0b/src/hypre/lib - > lHYPRE -lmpichcxx -lstdc++ -lspooles -lflapack -lfblas -lnsl -laio - > lrt -lPEPCF90 -ldl -L/home/lux/csae1801/petsc/petsc-3.0.0-p8/linux32- > intel-c-prod/lib -lmpich -lpthread -lrt -L/opt/intel/cmkl/10.1.1.019/ > lib/em64t -L/opt/intel/Compiler/11.0/074/ipp/em64t/lib -L/opt/intel/ > Compiler/11.0/074/mkl/lib/em64t -L/opt/intel/Compiler/11.0/074/tbb/ > em64t/cc4.1.0_libc2.4_kernel2.6.16.21/lib -L/opt/intel/Compiler/ > 11.0/074/lib/intel64 -L/usr/lib/gcc/x86_64-redhat-linux/4.1.2 -limf - > lsvml -lipgo -ldecimal -lirc -lgcc_s -lirc_s -lmpichf90 -lifport - > lifcore -lm -lm -lmpichcxx -lstdc++ -lmpichcxx -lstdc++ -ldl -lmpich > -lpthread -lrt -limf -lsvml -lipgo -ldecimal -lirc -lgcc_s -lirc_s - > ldl > ------------------------------------------ > #PETSc Option Table entries: > -dump_data false > -info > -ksp_compute_singularvalues > -ksp_rtol 1e-6 > -ksp_type preonly > -ksp_view > -log_summary > -mat_ignore_lower_triangular > -mat_mumps_icntl_14 200 > -mat_mumps_sym 1 > -mat_type sbaij > -monitorname errormonitor.dat > -nz 100 > -options_left > -pc_factor_mat_solver_package mumps > -pc_type cholesky > -preload false > -verb false > #End o PETSc Option Table entries > There are no unused options. > > 1olarr > > before sol put > sol put succes > before fullsol > /DIANA/AP/LS41 12:08:39 35.66-CPU 11.16-IO SOLVE > /DIANA/AP/LS41 12:08:39 35.66-CPU 11.18-IO FILLUP > /DIANA/AP/LS41 12:08:39 35.66-CPU 11.18-IO FILLUP > /DIANA/AP/LS41 12:08:39 35.66-CPU 11.18-IO POST > /DIANA/AP/LS41 12:08:41 36.79-CPU 12.27-IO POST > /DIANA/AP/LS41 12:08:41 36.79-CPU 12.27-IO LINSTA > /DIANA/DC/END 12:08:41 36.79-CPU 12.27-IO STOP > DIANA JOB 3446 finished > > 1 proc externalsolver stuck at 100% with 12%memory out of 12GB > 1 proc externalsolver at 0% > > mpiexec -np 2 externalsolver > [0] PetscInitialize(): PETSc successfully started: number of > processors = 2 > [0] PetscInitialize(): Running on machine: mat1.uibk.ac.at > [1] PetscInitialize(): PETSc successfully started: number of > processors = 2 > [1] PetscInitialize(): Running on machine: mat1.uibk.ac.at > [1] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [1] PetscCommDuplicate(): returning tag 2147483647 > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [0] PetscCommDuplicate(): returning tag 2147483647 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483646 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483646 > [0] MatSetUpPreallocation(): Warning not preallocating matrix storage > [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 > -2080374783 max tags = 2147483647 > [1] PetscCommDuplicate(): Duplicating a communicator 1140850689 > -2080374783 max tags = 2147483647 > [1] PetscCommDuplicate(): returning tag 2147483647 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483646 > [0] PetscCommDuplicate(): returning tag 2147483647 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483646 > [0] MatSetOption_MPISBAIJ(): Option SYMMETRIC ignored > [0] MatStashScatterBegin_Private(): No of messages: 1 > [0] MatStashScatterBegin_Private(): Mesg_to: 1: size: 30544064 > [0] MatStashScatterBegin_Private(): No of messages: 0 > [0] MatAssemblyBegin_MPISBAIJ(): Stash has 3818007 entries,uses 8 > mallocs. > [0] MatAssemblyBegin_MPISBAIJ(): Block-Stash has 3818007 entries, > uses 8 mallocs. > [0] MatAssemblyEnd_SeqSBAIJ(): Matrix size: 221545 X 221545, block > size 1; storage space: 16676676 unneeded, 5477824 used > [0] MatAssemblyEnd_SeqSBAIJ(): Number of mallocs during MatSetValues > is 0 > [0] MatAssemblyEnd_SeqSBAIJ(): Most nonzeros blocks in any row is 62 > [1] MatAssemblyEnd_SeqSBAIJ(): Matrix size: 225367 X 225367, block > size 1; storage space: 16793278 unneeded, 5743422 used > [1] MatAssemblyEnd_SeqSBAIJ(): Number of mallocs during MatSetValues > is 0 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483645 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483644 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483643 > [0] PetscCommDuplicate(): returning tag 2147483639 > [0] PetscCommDuplicate(): returning tag 2147483634 > [0] VecScatterCreateCommon_PtoS(): Using blocksize 1 scatter > [0] VecScatterCreate(): General case: MPI to Seq > [0] PetscCommDuplicate(): returning tag 2147483630 > [0] PetscCommDuplicate(): returning tag 2147483625 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483642 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483641 > [0] PetscCommDuplicate[1] MatAssemblyEnd_SeqSBAIJ(): Most nonzeros > blocks in any row is 65 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483645 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483644 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483643 > [1] PetscCommDuplicate(): returning tag 2147483639 > [1] PetscCommDuplicate(): returning tag 2147483634 > [1] PetscCommDuplicate(): returning tag 2147483630 > [1] PetscCommDuplicate(): returning tag 2147483625 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483642 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483641 > [1] PetscCommDuplicate(): returning tag 2147483620 > [1] PetscC(): returning tag 2147483620 > [0] VecScatterCreateCommon_PtoS(): Using blocksize 1 scatter > [0] VecScatterCreate(): General case: MPI to MPI > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483640 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483639 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483638 > [0] MatAssemblyEnd_SeqBAIJ(): Matrix size: 221545 X 6275, block size > 1; storage space: 22060506 unneeded, 93994 used > [0] MatAssemblyEnd_SeqBAIJ(): Number of mallocs during MatSetValues > is 0 > [0] MatAssemblyEnd_SeqBAIJ(): Most nonzeros blocks in any row is 32 > [0] Mat_CheckCompressedRow(): Found the ratio (num_zerorows 216534)/ > (num_localrows 221545) > 0.6. Use CompressedRow routines. > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetsommDuplicate(): Using internal PETSc communicator 1140850689 > -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483640 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483639 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483638 > [1] MatAssemblyEnd_SeqBAIJ(): Matrix size: 225367 X 0, block size 1; > storage space: 22536700 unneeded, 0 used > [1] MatAssemblyEnd_SeqBAIJ(): Number of mallocs during MatSetValues > is 0 > [1] MatAssemblyEnd_SeqBAIJ(): Most nonzeros blocks in any row is 0 > [1] Mat_CheckCompressedRow(): Found the ratio (num_zerorows 225367)/ > (num_localrows 225367) > 0.6. Use CompressedRow routines. > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483614 > [1] PetscCommDuplicate(): returning tag 2147483613 > [1]before KSPsetup > [1] PetscCommDuplicate(): Using internal PcCommDuplicate(): > returning tag 2147483614 > [0] PetscCommDuplicate(): returning tag 2147483613 > [0]before KSPsetup > [0] PCSetUp(): Setting up new PC > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483637 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483636 > [0] PetscCommDuplicate(): returning tag 2147483612 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483635 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483634 > ETSc communicator 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483637 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483636 > [1] PetscCommDuplicate(): returning tag 2147483612 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483635 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483634 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483633 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483632 > [0] PetscCommDuplicate(): returning tag 2147483607 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483633 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483632 > [1] PetscCommDuplicate(): returning tag 2147483607 > [0] PetscCommDuplicate(): returning tag 2147483602 > [0] VecScatterCreate(): Special case: processor zero gets entire > parallel vector, rest get none > [1] PetscCommDuplicate(): returning tag 2147483602 > > DMUMPS 4.8.4 > L D L^T Solver for symmetric positive definite matrices > Type of parallelism: Working host > > ****** ANALYSIS STEP ******** > > Density: NBdense, Average, Median = 0 49 50 > Ordering based on METIS > ** Peak of sequential stack size (number of real entries) : > 16763905. > A root of estimated size 2965 has been selected for > Scalapack. > > Leaving analysis phase with ... > INFOG(1) = 0 > INFOG(2) = 0 > -- (20) Number of entries in factors (estim.) = 120657071 > -- (3) Storage of factors (REAL, estimated) = 137364268 > -- (4) Storage of factors (INT , estimated) = 6138804 > -- (5) Maximum frontal size (estimated) = 3705 > -- (6) Number of nodes in the tree = 21863 > -- (7) Ordering option effectively used = 5 > ICNTL(6) Maximum transversal option = 0 > ICNTL(7) Pivot order option = 7 > Percentage of memory relaxation (effective) = 200 > Number of level 2 nodes = 1 > Number of split nodes = 0 > RINFO(1) Operations during elimination (estim) = 1.114D+11 > Distributed matrix entry format (ICNTL(18)) = 3 > ** Rank of proc needing largest memory in IC facto : 0 > ** Estimated corresponding MBYTES for IC facto : 2085 > ** Estimated avg. MBYTES per work. proc at facto (IC) : 2044 > ** TOTAL space in MBYTES for IC factorization : 4088 > ** Rank of proc needing largest memory for OOC facto : 1 > ** Estimated corresponding MBYTES for OOC facto : 644 > ** Estimated avg. MBYTES per work. proc at facto (OOC) : 630 > ** TOTAL space in MBYTES for OOC factorization : 1261 > > ****** FACTORIZATION STEP ******** > > > GLOBAL STATISTICS PRIOR NUMERICAL FACTORIZATION ... > NUMBER OF WORKING PROCESSES = 2 > REAL SPACE FOR FACTORS = 137364268 > INTEGER SPACE FOR FACTORS = 6138804 > MAXIMUM FRONTAL SIZE (ESTIMATED) = 3705 > NUMBER OF NODES IN THE TREE = 21863 > Maximum effective relaxed size of S = 239060000 > Average effective relaxed size of S = 234725641 > > REDISTRIB: TOTAL DATA LOCAL/SENT = 9086109 2238491 > GLOBAL TIME FOR MATRIX DISTRIBUTION = 0.8155 > ** Memory relaxation parameter ( ICNTL(14) ) : 200 > ** Rank of processor needing largest memory in facto : 0 > ** Space in MBYTES used by this processor for facto : 2085 > ** Avg. Space in MBYTES per working proc during facto : 2044 > > ELAPSED TIME FOR FACTORIZATION = 28.2595 > Maximum effective space used in S (KEEP(67) = 77516871 > Average effective space used in S (KEEP(67) = 75395654 > ** EFF Min: Rank of processor needing largest memory : 0 > ** EFF Min: Space in MBYTES used by this processor : 741 > ** EFF Min: Avg. Space in MBYTES per working proc : 720 > > GLOBAL STATISTICS > RINFOG(2) OPERATIONS DURING NODE ASSEMBLY = 2.565D+08 > ------(3) OPERATIONS DURING NODE ELIMINATION = 1.114D+11 > INFOG (9) REAL SPACE FOR FACTORS = 137364268 > INFOG(10) INTEGER SPACE FOR FACTORS = 6138809 > INFOG(11) MAXIMUM FRONT SIZE = 3705 > INFOG(29) NUMBER OF ENTRIES IN FACTORS = 116259976 > INFOG(12) NB OF NEGATIVE PIVOTS = 0 > INFOG(14) NUMBER OF MEMORY COMPRESS = 0 > [0]after KSPsetup > [0]RHS setup > [1]after KSPsetup > [1]RHS setup > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483601 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483601 > [1] PetscCommDuplicate(): returning tag 2147483596 > [0] PetscCommDuplicate(): returning tag 2147483596 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483591 > [0] PetscCommDuplicate(): returning tag 2147483591 > [0] VecAssemblyBegin_MPI(): Stash has 225367 entries, uses 12 mallocs. > [0] VecAssemblyBegin_MPI(): Block-Stash has 0 entries, uses 0 mallocs. > [0] PetscCommDuplicate(): returning tag 2147483590 > [1] PetscCommDuplicate(): returning tag 2147483590 > [1] PetscCommDuplicate(): returning tag 2147483589 > [0] PetscCommDuplicate(): returning tag 2147483589 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483631 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483630 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483631 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483630 > > 4 processes of externalsolver stuck with 0% cpu > > > mpiexec -np 4 externalsolver > [1] PetscInitialize(): PETSc successfully started: number of > processors = 4 > [1] PetscInitialize(): Running on machine: mat1.uibk.ac.at > [3] PetscInitialize(): PETSc successfully started: number of > processors = 4 > [3] PetscInitialize(): Running on machine: mat1.uibk.ac.at > [0] PetscInitialize(): PETSc successfully started: number of > processors = 4 > [0] PetscInitialize(): Running on machine: mat1.uibk.ac.at > [2] PetscInitialize(): PETSc successfully started: number of > processors = 4 > [2] PetscInitialize(): Running on machine: mat1.uibk.ac.at > [3] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [3] PetscCommDuplicate(): returning tag 2147483647 > [1] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [1] PetscCommDuplicate(): returning tag 2147483647 > [2] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [2] PetscCommDuplicate(): returning tag 2147483647 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [2] PetscCommDuplicate(): returning tag 2147483646 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [3] PetscCommDuplicate(): returning tag 2147483646 > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > [0] PetscCommDuplicate(): returning tag 2147483647 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483646 > [0] MatSetUpPreallocation(): Warning not preallocating matrix storage > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483646 > [1] PetscCommDuplicate(): Duplicating a communicator 1140850689 > -2080374783 max tags = 2147483647 > [3] PetscCommDuplicate(): Duplicating a communicator 1140850689 > -2080374783 max tags = 2147483647 > [3] PetscCommDuplicate(): returning tag 2147483647 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483646 > [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 > -2080374783 max tags = 2147483647 > [0] PetscCommDuplicate(): returning tag 2147483647 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483646 > [1] PetscCommDuplicate(): returning tag 2147483647 > [2] PetscCommDuplicate(): Duplicating a communicator 1140850689 > -2080374783 max tags = 2147483647 > [2] PetscCommDuplicate(): returning tag 2147483647 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483646 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483646 > [0] MatSetOption_MPISBAIJ(): Option SYMMETRIC ignored > [0] MatStashScatterBegin_Private(): No of messages: 3 > [0] MatStashScatterBegin_Private(): Mesg_to: 1: size: 11261544 > [0] MatStashScatterBegin_Private(): Mesg_to: 2: size: 10384840 > [0] MatStashScatterBegin_Private(): Mesg_to: 3: size: 393344 > [0] MatStashScatterBegin_Private(): No of messages: 0 > [0] MatAssemblyBegin_MPISBAIJ(): Stash has 2754963 entries,uses 8 > mallocs. > [0] MatAssemblyBegin_MPISBAIJ(): Block-Stash has 2754963 entries, > uses 8 mallocs. > [3] MatAssemblyEnd_SeqSBAIJ(): Matrix size: 112413 X 112413, block > size 1; storage space: 8404901 unneeded, 2836399 used > [3] MatAssemblyEnd_SeqSBAIJ(): Number of mallocs during MatSetValues > is 0 > [3] MatAssemblyEnd_SeqSBAIJ(): Most nonzeros blocks in any row is 65 > [1] MatAssemblyEnd_SeqSBAIJ(): Matrix size: 113299 X 113299, block > size 1; storage space: 8647559 unneeded, 2682341 used > [1] MatAssemblyEnd_SeqSBAIJ(): Number of mallocs during MatSetValues > is 0 > [1] MatAssemblyEnd_SeqSBAIJ(): Most nonzeros blocks in any row is 62 > [0] MatAssemblyEnd_SeqSBAIJ(): Matrix size: 112871 X 112871, block > size 1; storage space: 8670798 unneeded, 2616302 used > [0] MatAssemblyEnd_SeqSBAIJ(): Number of mallocs during MatSetValues > is 0 > [0] MatAssemblyEnd_SeqSBAIJ(): Most nonzeros blocks in any row is 52 > [2] MatAssemblyEnd_SeqSBAIJ(): Matrix size: 108329 X 108329, block > size 1; storage space: 8247737 unneeded, 2585163 used > [2] MatAssemblyEnd_SeqSBAIJ(): Number of mallocs during MatSetValues > is 0 > [2] MatAssemblyEnd_SeqSBAIJ(): Most nonzeros blocks in any row is 52 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483645 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483644 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483643 > [1] PetscCommDuplicate(): returning tag 2147483639 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483645 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483644 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483643 > [3] PetscCommDuplicate(): returning tag 2147483639 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483645 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483644 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483643 > [2] PetscCommDuplicate(): returning tag 2147483639 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483645 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483644 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483643 > [0] PetscCommDuplicate(): returning tag 2147483639 > [0] PetscCommDuplicate(): returning tag 2147483634 > [2] PetscCommDuplicate(): returning tag 2147483634 > [1] PetscCommDuplicate(): returning tag 2147483634 > [3] PetscCommDuplicate(): returning tag 2147483634 > [0] VecScatterCreateCommon_PtoS(): Using blocksize 1 scatter > [0] VecScatterCreate(): General case: MPI to Seq > [0] PetscCommDuplicate(): returning tag 2147483630 > [0] PetscCommDuplicate(): returning tag 2147483625 > [1] PetscCommDuplicate(): returning tag 2147483630 > [1] PetscCommDuplicate(): returning tag 2147483625 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483642 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483641 > [1] PetscCommDuplicate(): returning tag 2147483620 > [2] PetscCommDuplicate(): returning tag 2147483630 > [2] PetscCommDuplicate(): returning tag 2147483625 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483642 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483641 > [2] PetscCommDuplicate(): returning tag 2147483620 > [3] PetscCommDuplicate(): returning tag 2147483630 > [3] PetscCommDuplicate(): returning tag 2147483625 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483642 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483641 > [3] PetscCommDuplicate(): returning tag 2147483620 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483640 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483639 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483638 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483642 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483641 > [0] PetscCommDuplicate(): returning tag 2147483620 > [0] VecScatterCreateCommon_PtoS(): Using blocksize 1 scatter > [0] VecScatterCreate(): General case: MPI to MPI > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483640 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483639 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483638 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483640 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483639 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483638 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483640 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483639 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483638 > [3] MatAssemblyEnd_SeqBAIJ(): Matrix size: 112413 X 0, block size 1; > storage space: 11241300 unneeded, 0 used > [3] MatAssemblyEnd_SeqBAIJ(): Number of mallocs during MatSetValues > is 0 > [3] MatAssemblyEnd_SeqBAIJ(): Most nonzeros blocks in any row is 0 > [3] Mat_CheckCompressedRow(): Found the ratio (num_zerorows 112413)/ > (num_localrows 112413) > 0.6. Use CompressedRow routines. > [1] MatAssemblyEnd_SeqBAIJ(): Matrix size: 113299 X 2952, block size > 1; storage space: 11286236 unneeded, 43664 used > [1] MatAssemblyEnd_SeqBAIJ(): Number of mallocs during MatSetValues > is 0 > [1] MatAssemblyEnd_SeqBAIJ(): Most nonzeros blocks in any row is 29 > [1] Mat_CheckCompressedRow(): Found the ratio (num_zerorows 110902)/ > (num_localrows 113299) > 0.6. Use CompressedRow routines. > [2] MatAssemblyEnd_SeqBAIJ(): Matrix size: 108329 X 9258, block size > 1; storage space: 10624060 unneeded, 208840 used > [2] MatAssemblyEnd_SeqBAIJ(): Number of mallocs during MatSetValues > is 0 > [2] MatAssemblyEnd_SeqBAIJ(): Most nonzeros blocks in any row is 43 > [2] Mat_CheckCompressedRow(): Found the ratio (num_zerorows 92865)/ > (num_localrows 108329) > 0.6. Use CompressedRow routines. > [0] MatAssemblyEnd_SeqBAIJ(): Matrix size: 112871 X 17849, block > size 1; storage space: 10944569 unneeded, 342531 used > [0] MatAssemblyEnd_SeqBAIJ(): Number of mallocs during MatSetValues > is 0 > [0] MatAssemblyEnd_SeqBAIJ(): Most nonzeros blocks in any row is 43 > [0] Mat_CheckCompressedRow(): Found the ratio (num_zerorows 92260)/ > (num_localrows 112871) > 0.6. Use CompressedRow routines. > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [0] PetscCommDuplicate(): returning tag 2147483614 > [0] PetscCommDuplicate(): returning tag 2147483613 > [0]before KSPsetup > [0] PCSetUp(): Setting up new PC > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483637 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483636 > [0] PetscCommDuplicate(): returning tag 2147483612 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [1] PetscCommDuplicate(): returning tag 2147483614 > [1] PetscCommDuplicate(): returning tag 2147483613 > [1]before KSPsetup > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483637 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483636 > [1] PetscCommDuplicate(): returning tag 2147483612 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [2] PetscCommDuplicate(): returning tag 2147483614 > [2] PetscCommDuplicate(): returning tag 2147483613 > [2]before KSPsetup > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483637 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483636 > [2] PetscCommDuplicate(): returning tag 2147483612 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483635 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483634 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483635 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483634 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 -2080374784 > [3] PetscCommDuplicate(): returning tag 2147483614 > [3] PetscCommDuplicate(): returning tag 2147483613 > [3]before KSPsetup > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483637 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483636 > [3] PetscCommDuplicate(): returning tag 2147483612 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483635 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483634 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483635 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483634 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483633 > [1] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [1] PetscCommDuplicate(): returning tag 2147483632 > [1] PetscCommDuplicate(): returning tag 2147483607 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483633 > [2] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [2] PetscCommDuplicate(): returning tag 2147483632 > [2] PetscCommDuplicate(): returning tag 2147483607 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483633 > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [0] PetscCommDuplicate(): returning tag 2147483632 > [0] PetscCommDuplicate(): returning tag 2147483607 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483633 > [3] PetscCommDuplicate(): Using internal PETSc communicator > 1140850689 -2080374783 > [3] PetscCommDuplicate(): returning tag 2147483632 > [3] PetscCommDuplicate(): returning tag 2147483607 > [0] PetscCommDuplicate(): returning tag 2147483602 > [1] PetscCommDuplicate(): returning tag 2147483602 > [2] PetscCommDuplicate(): returning tag 2147483602 > [3] PetscCommDuplicate(): returning tag 2147483602 > [0] VecScatterCreate(): Special case: processor zero gets entire > parallel vector, rest get none > > DMUMPS 4.8.4 > L D L^T Solver for symmetric positive definite matrices > Type of parallelism: Working host > > ****** ANALYSIS STEP ******** > > Density: NBdense, Average, Median = 0 49 50 > Ordering based on METIS > ** Peak of sequential stack size (number of real entries) : > 16763905. > A root of estimated size 2965 has been selected for > Scalapack. > > Leaving analysis phase with ... > INFOG(1) = 0 > INFOG(2) = 0 > -- (20) Number of entries in factors (estim.) = 120657071 > -- (3) Storage of factors (REAL, estimated) = 137362626 > -- (4) Storage of factors (INT , estimated) = 6167135 > -- (5) Maximum frontal size (estimated) = 3705 > -- (6) Number of nodes in the tree = 21863 > -- (7) Ordering option effectively used = 5 > ICNTL(6) Maximum transversal option = 0 > ICNTL(7) Pivot order option = 7 > Percentage of memory relaxation (effective) = 200 > Number of level 2 nodes = 5 > Number of split nodes = 0 > RINFO(1) Operations during elimination (estim) = 1.114D+11 > Distributed matrix entry format (ICNTL(18)) = 3 > ** Rank of proc needing largest memory in IC facto : 0 > ** Estimated corresponding MBYTES for IC facto : 1112 > ** Estimated avg. MBYTES per work. proc at facto (IC) : 1083 > ** TOTAL space in MBYTES for IC factorization : 4333 > ** Rank of proc needing largest memory for OOC facto : 1 > ** Estimated corresponding MBYTES for OOC facto : 465 > ** Estimated avg. MBYTES per work. proc at facto (OOC) : 421 > ** TOTAL space in MBYTES for OOC factorization : 1684 > > ****** FACTORIZATION STEP ******** > > > GLOBAL STATISTICS PRIOR NUMERICAL FACTORIZATION ... > NUMBER OF WORKING PROCESSES = 4 > REAL SPACE FOR FACTORS = 137362626 > INTEGER SPACE FOR FACTORS = 6167135 > MAXIMUM FRONTAL SIZE (ESTIMATED) = 3705 > NUMBER OF NODES IN THE TREE = 21863 > Maximum effective relaxed size of S = 125678965 > Average effective relaxed size of S = 122625031 > > REDISTRIB: TOTAL DATA LOCAL/SENT = 2497290 8939630 > GLOBAL TIME FOR MATRIX DISTRIBUTION = 0.6324 > ** Memory relaxation parameter ( ICNTL(14) ) : 200 > ** Rank of processor needing largest memory in facto : 0 > ** Space in MBYTES used by this processor for facto : 1112 > ** Avg. Space in MBYTES per working proc during facto : 1083 >