Run with -start_in_debugger noxterm (since it seems to crash on one process) 
and type where when it crashes in the debugger. Also could run in valgrind 
http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind


On Oct 6, 2013, at 10:22 PM, Jose David Bermeol <[email protected]> wrote:

> No it doesn't
> 
> Nonzeros in L       10
>       Nonzeros in U       10
>       nonzeros in L+U     10
>       nonzeros in LSUB    10
>        Mat conversion(PETSc->SuperLU_DIST) time (max/min/avg): 
>                              5.91278e-05 / 5.91278e-05 / 5.91278e-05
>       EQUIL time             0.00
>       ROWPERM time           0.00
>       COLPERM time           0.00
>       SYMBFACT time          0.00
>       DISTRIBUTE time        0.00
>       FACTOR time            0.00
>       Factor flops    1.000000e+02    Mflops      0.29
>       SOLVE time             0.00
> [0]PETSC ERROR: 
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, 
> probably memory access out of range
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
> [0]PETSC ERROR: or see 
> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC ERROR: 
> or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory 
> corruption errors
> [0]PETSC ERROR: likely location of problem given in stack below
> [0]PETSC ERROR: ---------------------  Stack Frames 
> ------------------------------------
> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available,
> [0]PETSC ERROR:       INSTEAD the line number of the start of the function
> [0]PETSC ERROR:       is given.
> [0]PETSC ERROR: [0] SuperLU_DIST:pzgssvx_ABglobal line 228 
> /home/jbermeol/petsc/compile_pardiso/petsc-3.4.2/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
> [0]PETSC ERROR: [0] MatMatSolve_SuperLU_DIST line 198 
> /home/jbermeol/petsc/compile_pardiso/petsc-3.4.2/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
> [0]PETSC ERROR: [0] MatMatSolve line 3207 
> /home/jbermeol/petsc/compile_pardiso/petsc-3.4.2/src/mat/interface/matrix.c
> [0]PETSC ERROR: --------------------- Error Message 
> ------------------------------------
> [0]PETSC ERROR: Signal received!
> [0]PETSC ERROR: 
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Petsc Release Version 3.4.2, Jul, 02, 2013 
> [0]PETSC ERROR: See docs/changes/index.html for recent updates.
> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
> [0]PETSC ERROR: See docs/index.html for manual pages.
> [0]PETSC ERROR: 
> ------------------------------------------------------------------------
> [0]PETSC ERROR: ./test_superlu_dist on a arch-linux2-cxx-debug named 
> carter-fe00.rcac.purdue.edu by jbermeol Sun Oct  6 23:19:49 2013
> [0]PETSC ERROR: Libraries linked from 
> /home/jbermeol/petsc/compile_pardiso/petsc-3.4.2/arch-linux2-cxx-debug/lib
> [0]PETSC ERROR: Configure run at Sun Oct  6 22:05:55 2013
> [0]PETSC ERROR: Configure options --with-cc=mpiicc --with-cxx=mpiicpc 
> --with-fc=mpiifort --with-scalar-type=complex --with-shared-libraries=1 
> --with-debugging=1 --with-pic=1 --with-clanguage=C++ --with-fortran=1 
> --with-fortran-kernels=0 
> --with-blas-lapack-dir=/apps/rhel6/intel/composer_xe_2013.3.163/mkl 
> COPTFLAGS=-O CXXOPTFLAGS=-O FOPTFLAGS=-O --download-superlu_dist=yes 
> --download-superlu=yes --download-parmetis=yes --download-metis
> [0]PETSC ERROR: 
> ------------------------------------------------------------------------
> [0]PETSC ERROR: User provided function() line 0 in unknown directory unknown 
> file
> application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0
> 
> ----- Original Message -----
> From: "Barry Smith" <[email protected]>
> To: "Jose David Bermeol" <[email protected]>
> Cc: "Matthew Knepley" <[email protected]>, [email protected], "For 
> users of the development version of PETSc" <[email protected]>
> Sent: Sunday, October 6, 2013 4:44:32 PM
> Subject: Re: [petsc-dev] [petsc-users] Superlu_dist error
> 
> 
>  If you install with -O instead of -O3 (use a different PETSC_ARCH so it 
> doesn't overwrite) does the problem go away?
> 
>   Barry
> 
> On Oct 6, 2013, at 12:52 PM, Jose David Bermeol <[email protected]> wrote:
> 
>> Hi again, now I'm getting the error again. The problem happens when I'm 
>> using the flag -O3 for the compiler. So what should I do next to solve this??
>> 
>> Attached is my code, and for tis example I'm creating the matrix in the code.
>> 
>> Thanks  
>> 
>> Nonzeros in L       10
>>      Nonzeros in U       10
>>      nonzeros in L+U     10
>>      nonzeros in LSUB    10
>>      NUMfact space (MB) sum(procs):  L\U     0.00    all     0.03
>>      Total highmark (MB):  All       0.03    Avg     0.02    Max     0.02
>>       Mat conversion(PETSc->SuperLU_DIST) time (max/min/avg): 
>>                             0.000124216 / 4.81606e-05 / 8.61883e-05
>>      EQUIL time             0.00
>>      ROWPERM time           0.00
>>      COLPERM time           0.00
>>      SYMBFACT time          0.00
>>      DISTRIBUTE time        0.00
>>      FACTOR time            0.00
>>      Factor flops    1.000000e+02    Mflops      0.31
>>      SOLVE time             0.00
>> [1]PETSC ERROR: 
>> ------------------------------------------------------------------------
>> [1]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, 
>> probably memory access out of range
>> [1]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
>> [1]PETSC ERROR: or see 
>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[1]PETSC ERROR: 
>> or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory 
>> corruption errors
>> [1]PETSC ERROR: likely location of problem given in stack below
>> [1]PETSC ERROR: ---------------------  Stack Frames 
>> ------------------------------------
>> [1]PETSC ERROR: Note: The EXACT line numbers in the stack are not available,
>> [1]PETSC ERROR:       INSTEAD the line number of the start of the function
>> [1]PETSC ERROR:       is given.
>> [1]PETSC ERROR: [1] SuperLU_DIST:pzgssvx line 234 
>> /home/jbermeol/petsc/petsc_superlu_dist/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
>> [1]PETSC ERROR: [1] MatMatSolve_SuperLU_DIST line 198 
>> /home/jbermeol/petsc/petsc_superlu_dist/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
>> [1]PETSC ERROR: [1] MatMatSolve line 3207 
>> /home/jbermeol/petsc/petsc_superlu_dist/src/mat/interface/matrix.c
>> [1]PETSC ERROR: --------------------- Error Message 
>> ------------------------------------
>> [1]PETSC ERROR: Signal received!
>> [1]PETSC ERROR: 
>> ------------------------------------------------------------------------
>> [1]PETSC ERROR: Petsc Release Version 3.4.2, Jul, 02, 2013 
>> [1]PETSC ERROR: See docs/changes/index.html for recent updates.
>> [1]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
>> [1]PETSC ERROR: See docs/index.html for manual pages.
>> [1]PETSC ERROR: 
>> ------------------------------------------------------------------------
>> [1]PETSC ERROR: ./test_solver on a arch-linux2-c-debug named 
>> carter-fe00.rcac.purdue.edu by jbermeol Sun Oct  6 13:43:00 2013
>> [1]PETSC ERROR: Libraries linked from 
>> /home/jbermeol/petsc/petsc_superlu_dist/arch-linux2-c-debug/lib
>> [1]PETSC ERROR: Configure run at Sun Oct  6 13:38:20 2013
>> [1]PETSC ERROR: Configure options --with-cc=mpiicc --with-cxx=mpiicpc 
>> --with-fc=mpiifort --with-scalar-type=complex --with-shared-libraries=1 
>> --with-debugging=1 --download-f-blas-lapack --download-superlu_dist=yes 
>> --download-superlu=yes --download-parmetis=yes --download-metis 
>> COPTFLAGS=-O3 CXXOPTFLAGS=-O3 FOPTFLAGS=-O3
>> [1]PETSC ERROR: 
>> ------------------------------------------------------------------------
>> [1]PETSC ERROR: User provided function() line 0 in unknown directory unknown 
>> file
>> [0]PETSC ERROR: 
>> ------------------------------------------------------------------------
>> [0]PETSC ERROR: application called MPI_Abort(MPI_COMM_WORLD, 59) - process 1
>> Caught signal number 11 SEGV: Segmentation Violation, probably memory access 
>> out of range
>> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
>> [0]PETSC ERROR: or see 
>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC ERROR: 
>> or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory 
>> corruption errors
>> [0]PETSC ERROR: likely location of problem given in stack below
>> [0]PETSC ERROR: ---------------------  Stack Frames 
>> ------------------------------------
>> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available,
>> [0]PETSC ERROR:       INSTEAD the line number of the start of the function
>> [0]PETSC ERROR:       is given.
>> [0]PETSC ERROR: [0] SuperLU_DIST:pzgssvx line 234 
>> /home/jbermeol/petsc/petsc_superlu_dist/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
>> [0]PETSC ERROR: [0] MatMatSolve_SuperLU_DIST line 198 
>> /home/jbermeol/petsc/petsc_superlu_dist/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
>> [0]PETSC ERROR: [0] MatMatSolve line 3207 
>> /home/jbermeol/petsc/petsc_superlu_dist/src/mat/interface/matrix.c
>> [0]PETSC ERROR: --------------------- Error Message 
>> ------------------------------------
>> [0]PETSC ERROR: Signal received!
>> [0]PETSC ERROR: 
>> ------------------------------------------------------------------------
>> [0]PETSC ERROR: Petsc Release Version 3.4.2, Jul, 02, 2013 
>> [0]PETSC ERROR: See docs/changes/index.html for recent updates.
>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
>> [0]PETSC ERROR: See docs/index.html for manual pages.
>> [0]PETSC ERROR: 
>> ------------------------------------------------------------------------
>> [0]PETSC ERROR: ./test_solver on a arch-linux2-c-debug named 
>> carter-fe00.rcac.purdue.edu by jbermeol Sun Oct  6 13:43:00 2013
>> [0]PETSC ERROR: Libraries linked from 
>> /home/jbermeol/petsc/petsc_superlu_dist/arch-linux2-c-debug/lib
>> [0]PETSC ERROR: Configure run at Sun Oct  6 13:38:20 2013
>> [0]PETSC ERROR: Configure options --with-cc=mpiicc --with-cxx=mpiicpc 
>> --with-fc=mpiifort --with-scalar-type=complex --with-shared-libraries=1 
>> --with-debugging=1 --download-f-blas-lapack --download-superlu_dist=yes 
>> --download-superlu=yes --download-parmetis=yes --download-metis 
>> COPTFLAGS=-03 CXXOPTFLAGS=-03 FOPTFLAGS=-03
>> [0]PETSC ERROR: 
>> ------------------------------------------------------------------------
>> [0]PETSC ERROR: User provided function() line 0 in unknown directory unknown 
>> file
>> application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0
>> 
>> ----- Original Message -----
>> From: "Matthew Knepley" <[email protected]>
>> To: "Jose David Bermeol" <[email protected]>
>> Cc: "For users of the development version of PETSc" <[email protected]>, 
>> [email protected]
>> Sent: Sunday, October 6, 2013 8:19:30 AM
>> Subject: Re: [petsc-users] Superlu_dist error
>> 
>> 
>> On Sun, Oct 6, 2013 at 12:10 AM, Jose David Bermeol < [email protected] > 
>> wrote: 
>> 
>> 
>> 
>> 
>> Hi again, I compile with the following configuration: 
>> --with-cc=mpiicc --with-cxx=mpiicpc --with-fc=mpiifort 
>> --with-scalar-type=complex --with-shared-libraries=1 --with-debugging=1 
>> --with-pic=1 --with-clanguage=C++ --with-fortran=1 --with-fortran-kernels=0 
>> --download-f-blas-lapack --download-superlu_dist=yes --download-superlu=yes 
>> --download-parmetis=yes --download-metis 
>> 
>> 
>> 
>> Get rid of: 
>> 
>> 
>> --with-pic=1 --with-fortran=1 --with-fortran-kernels=0 
>> 
>> 
>> since they do not really do anything, and just put back MKL. I suspect you 
>> will get a crash, and then 
>> it sounds like an MKL bug, or a bizarre incompatibility between SuperLU and 
>> MKL. If not, we explore further. 
>> 
>> 
>> Thanks, 
>> 
>> 
>> Matt 
>> 
>> 
>> And my code run perfec, so that means is a MKL problem or a mismatch between 
>> versions, so how to test what is the problem?? 
>> 
>> Thanks 
>> 
>> ----- Original Message ----- 
>> From: "Matthew Knepley" < [email protected] > 
>> To: "Jose David Bermeol" < [email protected] > 
>> Cc: "For users of the development version of PETSc" < [email protected] 
>> >, [email protected] 
>> Sent: Saturday, October 5, 2013 11:55:24 PM 
>> Subject: Re: [petsc-users] Superlu_dist error 
>> 
>> 
>> On Sat, Oct 5, 2013 at 10:49 PM, Jose David Bermeol < [email protected] > 
>> wrote: 
>> 
>> 
>> 
>> 
>> Hi I'm runnig petsc trying to solve a linear system with superlu_dist. 
>> However i have a memory violation, atached is the code, and here is the 
>> output. Email me if you need something else to figured out what is 
>> happening. 
>> 
>> 
>> 
>> So it looks like SuperLU_Dist is bombing during an LAPACK operation. It 
>> could be an MKL problem, or a SuperLU_Dist problem, or our problem, 
>> or a mismatch between versions. I would try to simplify the configuration in 
>> order to cut down on the possibilities. Eliminate everything that is not 
>> necessary for SuperLU_dist first. Then change to --download-f-blas-lapack. 
>> If you still have a crash, send us the matrix since that should be 
>> reproducible and we can report a SuperLU_dist bug or fix our code. 
>> 
>> 
>> Thanks, 
>> 
>> 
>> Matt 
>> 
>> 
>> Thanks 
>> 
>> mpiexec -n 2 ./test_solver -mat_superlu_dist_statprint 
>> -mat_superlu_dist_matinput distributed 
>> Nonzeros in L 10 
>> Nonzeros in U 10 
>> nonzeros in L+U 10 
>> nonzeros in LSUB 10 
>> NUMfact space (MB) sum(procs): L\U 0.00 all 0.03 
>> Total highmark (MB): All 0.03 Avg 0.02 Max 0.02 
>> Mat conversion(PETSc->SuperLU_DIST) time (max/min/avg): 
>> 0.000146866 / 0.000145912 / 0.000146389 
>> EQUIL time 0.00 
>> ROWPERM time 0.00 
>> COLPERM time 0.00 
>> SYMBFACT time 0.00 
>> DISTRIBUTE time 0.00 
>> FACTOR time 0.00 
>> Factor flops 1.000000e+02 Mflops 0.31 
>> SOLVE time 0.00 
>> [0]PETSC ERROR: 
>> ------------------------------------------------------------------------ 
>> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, 
>> probably memory access out of range 
>> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger 
>> [0]PETSC ERROR: or see 
>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC ERROR: 
>> [1]PETSC ERROR: 
>> ------------------------------------------------------------------------ 
>> [1]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, 
>> probably memory access out of range 
>> [1]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X 
>> to find memory corruption errors 
>> Try option -start_in_debugger or -on_error_attach_debugger 
>> [1]PETSC ERROR: or see 
>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[1]PETSC ERROR: 
>> or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory 
>> corruption errors 
>> [0]PETSC ERROR: likely location of problem given in stack below 
>> [1]PETSC ERROR: likely location of problem given in stack below 
>> [1]PETSC ERROR: --------------------- Stack Frames 
>> ------------------------------------ 
>> [1]PETSC ERROR: Note: The EXACT line numbers in the stack are not available, 
>> [0]PETSC ERROR: --------------------- Stack Frames 
>> ------------------------------------ 
>> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available, 
>> [0]PETSC ERROR: [1]PETSC ERROR: INSTEAD the line number of the start of the 
>> function 
>> [1]PETSC ERROR: is given. 
>> [1]PETSC ERROR: [1] SuperLU_DIST:pzgssvx line 234 
>> /home/jbermeol/Nemo5/libs/petsc/build-cplx/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
>>  
>> [1]PETSC ERROR: [1] MatMatSolve_SuperLU_DIST line 198 
>> /home/jbermeol/Nemo5/libs/petsc/build-cplx/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
>>  
>> [1]PETSC ERROR: INSTEAD the line number of the start of the function 
>> [0]PETSC ERROR: is given. 
>> [0]PETSC ERROR: [0] SuperLU_DIST:pzgssvx line 234 
>> /home/jbermeol/Nemo5/libs/petsc/build-cplx/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
>>  
>> [0]PETSC ERROR: [1] MatMatSolve line 3207 
>> /home/jbermeol/Nemo5/libs/petsc/build-cplx/src/mat/interface/matrix.c 
>> [1]PETSC ERROR: --------------------- Error Message 
>> ------------------------------------ 
>> [1]PETSC ERROR: [0] MatMatSolve_SuperLU_DIST line 198 
>> /home/jbermeol/Nemo5/libs/petsc/build-cplx/src/mat/impls/aij/mpi/superlu_dist/superlu_dist.c
>>  
>> [0]PETSC ERROR: [0] MatMatSolve line 3207 
>> /home/jbermeol/Nemo5/libs/petsc/build-cplx/src/mat/interface/matrix.c 
>> Signal received! 
>> [1]PETSC ERROR: 
>> ------------------------------------------------------------------------ 
>> [1]PETSC ERROR: Petsc Release Version 3.4.2, Jul, 02, 2013 
>> [1]PETSC ERROR: See docs/changes/index.html for recent updates. 
>> [1]PETSC ERROR: See docs/faq.html for hints about trouble shooting. 
>> [1]PETSC ERROR: See docs/index.html for manual pages. 
>> [1]PETSC ERROR: 
>> ------------------------------------------------------------------------ 
>> [1]PETSC ERROR: ./test_solver on a linux-complex named 
>> carter-fe02.rcac.purdue.edu by jbermeol Sat Oct 5 23:45:21 2013 
>> [1]PETSC ERROR: [0]PETSC ERROR: --------------------- Error Message 
>> ------------------------------------ 
>> [0]PETSC ERROR: Libraries linked from 
>> /home/jbermeol/Nemo5/libs/petsc/build-cplx/linux-complex/lib 
>> [1]PETSC ERROR: Configure run at Sat Oct 5 11:19:36 2013 
>> [1]PETSC ERROR: Configure options --with-cc=mpiicc --with-cxx=mpiicpc 
>> --with-fc=mpiifort --with-scalar-type=complex --with-shared-libraries=1 
>> --with-debugging=1 --with-pic=1 --with-clanguage=C++ --with-fortran=1 
>> --with-fortran-kernels=0 
>> --with-blas-lapack-dir=/apps/rhel6/intel/composer_xe_2013.3.163/mkl 
>> --with-blacs-lib=/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64/libmkl_blacs_intelmpi_lp64.so
>>  --with-blacs-include=/apps/rhel6/intel/composer_xe_2013.3.163/mkl/include 
>> --with-scalapack-lib="-L/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64
>>  -lmkl_scalapack_lp64 -lmkl_blacs_intelmpi_lp64" 
>> --with-scalapack-include=/apps/rhel6/intel/composer_xe_2013.3.163/mkl/include
>>  --with-valgrind-dir=/apps/rhel6/valgrind/3.8.1 --COPTFLAGS=-O3 
>> --CXXOPTFLAGS=-O3 --FOPTFLAGS=-O3 
>> --with-mkl-include=/apps/rhel6/intel/composer_xe_2013.3.163/mkl/include 
>> --with-mkl-lib="[/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64/libmkl_intel_lp64.so,/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64/libmkl_intel_thread.so,/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64/libmkl_core.so,/apps/rhel6/intel/composer_xe_2013.3.163/mkl/../compiler/lib/intel64/libiomp5.so]"
>>  --with-cpardiso-dir=/home/jbermeol/testPetscSolvers/intel_mkl_cpardiso 
>> --with-hdf5 --download-hdf5=yes --download-metis=yes --download-parmetis=yes 
>> --download-superlu_dist=yes --download-superlu=yes --download-mumps=yes 
>> --download-spooles=yes --download-pastix=yes --download-ptscotch=yes 
>> --download-umfpack=yes --download-sowing 
>> Signal received! 
>> [0]PETSC ERROR: 
>> ------------------------------------------------------------------------ 
>> [0]PETSC ERROR: Petsc Release Version 3.4.2, Jul, 02, 2013 
>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. 
>> [0]PETSC ERROR: [1]PETSC ERROR: 
>> ------------------------------------------------------------------------ 
>> [1]PETSC ERROR: User provided function() line 0 in unknown directory unknown 
>> file 
>> See docs/faq.html for hints about trouble shooting. 
>> [0]PETSC ERROR: See docs/index.html for manual pages. 
>> [0]PETSC ERROR: 
>> ------------------------------------------------------------------------ 
>> [0]PETSC ERROR: ./test_solver on a linux-complex named 
>> carter-fe02.rcac.purdue.edu by jbermeol Sat Oct 5 23:45:21 2013 
>> [0]PETSC ERROR: Libraries linked from 
>> /home/jbermeol/Nemo5/libs/petsc/build-cplx/linux-complex/lib 
>> [0]PETSC ERROR: Configure run at Sat Oct 5 11:19:36 2013 
>> [0]PETSC ERROR: application called MPI_Abort(MPI_COMM_WORLD, 59) - process 1 
>> Configure options --with-cc=mpiicc --with-cxx=mpiicpc --with-fc=mpiifort 
>> --with-scalar-type=complex --with-shared-libraries=1 --with-debugging=1 
>> --with-pic=1 --with-clanguage=C++ --with-fortran=1 --with-fortran-kernels=0 
>> --with-blas-lapack-dir=/apps/rhel6/intel/composer_xe_2013.3.163/mkl 
>> --with-blacs-lib=/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64/libmkl_blacs_intelmpi_lp64.so
>>  --with-blacs-include=/apps/rhel6/intel/composer_xe_2013.3.163/mkl/include 
>> --with-scalapack-lib="-L/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64
>>  -lmkl_scalapack_lp64 -lmkl_blacs_intelmpi_lp64" 
>> --with-scalapack-include=/apps/rhel6/intel/composer_xe_2013.3.163/mkl/include
>>  --with-valgrind-dir=/apps/rhel6/valgrind/3.8.1 --COPTFLAGS=-O3 
>> --CXXOPTFLAGS=-O3 --FOPTFLAGS=-O3 
>> --with-mkl-include=/apps/rhel6/intel/composer_xe_2013.3.163/mkl/include 
>> --with-mkl-lib="[/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64/libmkl_intel_lp64.so,/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64/libmkl_intel_thread.so,/apps/rhel6/intel/composer_xe_2013.3.163/mkl/lib/intel64/libmkl_core.so,/apps/rhel6/intel/composer_xe_2013.3.163/mkl/../compiler/lib/intel64/libiomp5.so]"
>>  --with-cpardiso-dir=/home/jbermeol/testPetscSolvers/intel_mkl_cpardiso 
>> --with-hdf5 --download-hdf5=yes --download-metis=yes --download-parmetis=yes 
>> --download-superlu_dist=yes --download-superlu=yes --download-mumps=yes 
>> --download-spooles=yes --download-pastix=yes --download-ptscotch=yes 
>> --download-umfpack=yes --download-sowing 
>> [0]PETSC ERROR: 
>> ------------------------------------------------------------------------ 
>> [0]PETSC ERROR: User provided function() line 0 in unknown directory unknown 
>> file 
>> application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0 
>> 
>> 
>> 
>> -- 
>> What most experimenters take for granted before they begin their experiments 
>> is infinitely more interesting than any results to which their experiments 
>> lead. 
>> -- Norbert Wiener 
>> 
>> 
>> 
>> 
>> -- 
>> What most experimenters take for granted before they begin their experiments 
>> is infinitely more interesting than any results to which their experiments 
>> lead. 
>> -- Norbert Wiener<test_superlu_dist.c>

Reply via email to