Hi Satish, looks like simply leaving out --with-cc=gcc and --with-fc=gfortran did the trick.
thanks matt On Thursday 22 January 2009, Satish Balay wrote: > Looks like this old version of mpi has MPI_Win_create() - but not the > associated MPI_Win definition in its include. > > You can edit petscconf.h and remove the lines > > #ifndef PETSC_HAVE_MPI_WIN_CREATE > #define PETSC_HAVE_MPI_WIN_CREATE 1 > #endif > > and then recompiling petsc. > > BTW: when you specify --with-mpi-dir=/usr/mpi/mvapich-0.9.9/gcc > PETSc configure will use mpicc/mpif77 etc as compilers. > > So its best not to specify '--with-cc=gcc --with-fc=gfortran' etcc > > Satish > > On Thu, 22 Jan 2009, Matt Funk wrote: > > Hi Satish, > > > > it still doesn't work, but something changed. I configure now with: > > ./config/configure.py --with-mpi-dir=/usr/mpi/mvapich-0.9.9/gcc > > --with-debugging=0 --with-log=0 --with-cc=gcc --with-fc=gfortran > > --download-f-blas-lapack=1 --download-superlu=1 --download-superlu_dist=1 > > --download-spooles=1 --download-hypre=hypre-2.4.0b.tar.gz rather than > > specifying the include dir and lib dir specifically. > > > > Anyway, it passes through the configuration stage. After setting > > PETSC_DIR and PETSC_ARCH, i do: make all test. This one fails. > > > > The first couple error are printed to the screen as such: > > > > In file included from vinv.c:6: > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/include/private/vecimpl.h:236 > >: error: expected specifier-qualifier-list before 'MPI_Win' > > In file included from vscat.c:9: > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/include/private/vecimpl.h:236 > >: error: expected specifier-qualifier-list before 'MPI_Win' > > In file included from vpscat.c:8: > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/include/private/vecimpl.h:236 > >: error: expected specifier-qualifier-list before 'MPI_Win' > > In file included from vpscat.c:1370: > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/src/vec/vec/utils/vpscat.h: > > In function 'VecScatterBegin_1': > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/src/vec/vec/utils/vpscat.h:65 > >: warning: implicit declaration of function 'MPI_Win_fence' > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/src/vec/vec/utils/vpscat.h:65 > >: error: 'VecScatter_MPI_General' has no member named 'window' > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/src/vec/vec/utils/vpscat.h:68 > >: warning: implicit declaration of function 'MPI_Put' > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/src/vec/vec/utils/vpscat.h:68 > >: error: 'VecScatter_MPI_General' has no member named 'window' > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/src/vec/vec/utils/vpscat.h:70 > >: error: 'VecScatter_MPI_General' has no member named 'window' > > In file included from vpscat.c:1372: > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/src/vec/vec/utils/vpscat.h: > > In function 'VecScatterBegin_2': > > /home/mafunk/Packages/petsc-2.3.3-p15_Conf0/src/vec/vec/utils/vpscat.h:65 > >: error: 'VecScatter_MPI_General' has no member named 'window' > > > > > > I attached the config file. > > > > > > thanks > > matt > > > > On Wednesday 21 January 2009, you wrote: > > > Each time - do you see binary chars at the end of configure.log - > > > after the _access test? > > > > > > I don't know why this would happen. > > > > > > Perhaps you can try with the same configure options - and send in the > > > configure.log [petsc-maint at mcs.anl.gov] for us to comapre. > > > > > > Satish > > > > > > On Wed, 21 Jan 2009, Matt Funk wrote: > > > > I tried it multiple from a fresh tarball each time. Same thing > > > > everytime. Did you mean try with different configure options? If so, > > > > which ones did you have in mind? > > > > > > > > thanks > > > > matt > > > > > > > > On Wednesday 21 January 2009, Satish Balay wrote: > > > > > Not sure what the issue is. I see the configure.log [from the > > > > > rejected e-mail] - there are some binary chars at the end of the > > > > > file. > > > > > > > > > > Perhaps you can retry and see? [after rm *.log]. If you still have > > > > > issues - send configure.log to petsc-maint at mcs.anl.gov > > > > > > > > > > Satish > > > > > > > > > > On Wed, 21 Jan 2009, Matt Funk wrote: > > > > > > Hi, > > > > > > > > > > > > i was wondering if there is any issues with compiling petsc based > > > > > > on mvapich-0.9.9. I tells me that it is unable to configure with > > > > > > given options. I configure as such: > > > > > > ./config/configure.py > > > > > > --with-mpi-include=/usr/mpi/mvapich-0.9.9/gcc/include > > > > > > --with-mpi-lib=/usr/mpi/mvapich-0.9.9/gcc/lib/libmpich.a > > > > > > --with-debugging=0 --with-log=0 --with-cc=gcc --with-fc=gfortran > > > > > > --download-f-blas-lapack=1 --download-superlu=1 > > > > > > --download-superlu_dist=1 --download-spooles=1 > > > > > > --download-hypre=hypre-2.4.0b.tar.gz > > > > > > > > > > > > I can compile ?petsc against the mvapich2 libraries just fine. > > > > > > > > > > > > > > > > > > thanks > > > > > > matt
