You mean, ./configure --with-cc=gcc --with-cxx=g++ --with-fc=ifort --download-mpich=1 --download-hypre=1 --download-openmpi=1 thanks.
David On Mon, Mar 22, 2010 at 3:58 PM, Matthew Knepley <knepley at gmail.com> wrote: > On Mon, Mar 22, 2010 at 3:54 PM, David sheehan <david.sheehanjr at > gmail.com>wrote: > >> I don't have icc and icpc. I only have ifort and gnu compilers such >> g77,gcc >> and g++. Since my application code works very well with ifort, can I build >> >> up PETSC plus Hypre with ifort and gnu compilers? thanks. >> > > It might be possible (though not guaranteed if ifort conflicts with GNU > somehow). You > jsut provide these compilers to the configure, and use --download-openmpi. > > Matt > > >> >> David >> >> On Mon, Mar 22, 2010 at 3:29 PM, Satish Balay <balay at mcs.anl.gov> wrote: >> >>> If you have installation issues - send the relavent logs [in this case >>> configure.log] to petsc-maint. >>> >>> Here you appear to try too many things.. But its not clear if you are >>> installing openmpi/lam yourself - or using default from Red Hat 3.4. >>> [note - default openmpi will be built with gnu compilers - so >>> unuseable from ifort]. >>> >>> Thing to do: >>> >>> - decide on compilers you want to use. >>> - build PETSc and mpi with these compilers >>> >>> [for eg: configure CC=icc CXX=icpc FC=ifort --download-mpich=1 >>> --download-hypre=1] >>> >>> Note: alternative f90 compilers seable on linux are gfortran, g95 >>> >>> Satish >>> >>> On Mon, 22 Mar 2010, David sheehan wrote: >>> >>> > Hi, >>> > I am trying to compile the PETSC with hypre. I have Intel FORTRAN, >>> openmpi >>> > and lam >>> > MPI (gcc version 3.4.6 (Red Hat 3.4.6-3)) available. >>> > >>> > Since my application code is in FORTRAN with dynamic memory allocation >>> in >>> > the code, >>> > so I have to use Intel FORTRAN(ifort) as the compiler to link PETSC >>> with my >>> > application >>> > code. I can build up PETSC(petsc-3.0.0-p11) plus hypre with Lam MPI but >>> > without ifort >>> > successfully. However I can not use ifort to link the PETSC with my >>> > application success- >>> > fully. The link always shows erros about "undefined reference" to MPI >>> parts >>> > in my application >>> > code such as 'mpi_send_', 'mpi_recv_', 'mpi_waitany_', >>> 'mpi_get_count_', >>> > and etc. >>> > >>> > Can any one help me out? thanks in advance. >>> > >>> > Best Regards, >>> > >>> > David >>> > >>> >>> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20100322/74eaf0bf/attachment.html>
