On Sat, 7 Jul 2018, Hector E Barrios Molano wrote: > Thanks Barry and Satish for your answers. > > I installed the correct version of hypre. Also, I changed the paths to short > dos paths as Satish suggested. Now PETSc compiles without problems and the > tests are ok. > > Regarding Satish notes: > > I had to include -LIBS because otherwise the configure script stops while > testing blas-lapack with the following error: > > =============================================================================== > TESTING: checkLib from > config.packages.BlasLapack(config/BuildSystem/config/packages/BlasLapack.py:114)******************************************************************************* > UNABLE to CONFIGURE with GIVEN OPTIONS (see configure.log for > details): > ------------------------------------------------------------------------------- > You set a value for --with-blaslapack-lib=<lib>, but > ['-L/cygdrive/c/PROGRA~2/INTELS~1/COMPIL~1/windows/mkl/lib/intel64', > 'mkl_intel_lp64.lib', 'mkl_core.lib', 'mkl_intel_thread.lib'] cannot be used > *******************************************************************************
Well send us configure.log for this failure [and also configure.log for the subsequent successful one with LIB options] > > Do I need to turn on the MKL sparse functionality? I'm not sure what you are using PETSc for. If you elaborate - perhaps Richard might be able to suggest if MKL_SPARSE is useful for you. > what is the difference > between MKL_SPARSE, MKL_SPARSE_OPTIMIZE and MKL_SPARSE_SP2M in the configure > options? Older MKL versions have limited sparse functionality. So these flags [detected by configure] control the level of MKL sparse functionality used. > Why is better to use sequential MKL? Because PETSc primarily uses MPI model - and uses sequential blas in this model. > Is it possible to use a hybrid MPI - > OpenMP approach? for example MPI for internode communication and OpenMP for > intranode computation? In that case would it be good to use threaded MKL + > MPI? You might want to check the trhead "GAMG error with MKL" on petsc-dev list. Satish