Just use -bv_type mat and the error will go away. Note: you are highly advised to update to a new PETSc/SLEPc version.
Thanks, Pierre > On 30 Apr 2025, at 7:58 AM, hexioafeng <hexiaof...@buaa.edu.cn> wrote: > > Dear Sir, > > Thank you for your kind reply. Bellow are the full backtrace: > > [0]PETSC ERROR: --------------------- Error Message > -------------------------------------------------------------- > [0]PETSC ERROR: No support for this operation for this object type > [0]PETSC ERROR: Product of two integer 1925 4633044 overflow, you must > ./configure PETSc with --with-64-bit-indices for the case you are running > > [0]PETSC ERROR: See https://www.mcs.anl.gov/petsc/documentation/faq.html > for trouble shooting. > [0]PETSC ERROR: Petsc Release Version 3.14.3, Jan 09, 2021 > [0]PETSC ERROR: Unknown Name on a named DESKTOP-74R6I4M by ibe Sun Apr 27 > 16:10:08 2025 > [0]PETSC ERROR: Configure options --with-cc=mpicc --with-cxx=mpicxx > --with-fc=mpifort --with-scalar-type=real --with-precision=double > --prefix=/home/sd/petsc/petsc-3.14.3/..//rd CXXFLAGS=-fno-stack-protector > CFLAGS=-fno-stack-protector FFLAGS=" -O2 -fallow-argument-mismatch > -fallow-invalid-boz" --with-debugging=0 COPTFLAGS="-O3 -mtune=generic" > CXXOPTFLAGS="-O3 -mtune=generic" FOPTFLAGS="-O3 -mtune=generic" > --known-64-bit-blas-indices=0 --with-cxx-dialect=C++11 --with-ssl=0 > --with-x=0 --with-fortran-bindings=0 --with-cudac=0 --with-shared-libraries=0 > --with-mpi-lib=/c/Windows/System32/msmpi.dll > --with-mpi-include=/home/sd/petsc/thirdparty/MPI/Include > --with-mpiexec="/C/Program Files/Microsoft MPI/Bin/mpiexec" > --with-blaslapack-lib="-L/home/sd/petsc/thirdparty/openblas/lib -llibopenblas > -lopenblas" --with-metis-include=/home/sd/petsc/scalapack-mumps-dll/metis > --with-metis-lib=/home/sd/petsc/scalapack-mumps-dll/metis/libmetis.dll > --with-parmetis-include=/home/sd/petsc/scalapack-mumps-dll/metis > --with-parmetis-lib=/home/sd/petsc/scalapack-mumps-dll/metis/libparmetis.dll > --download-slepc > --with-scalapack-lib=/home/sd/petsc/scalapack-mumps-dll/scalapack/libscalapack.dll > --download-hypre --download-mumps > --download-hypre-configure-arguments="--build=x86_64-linux-gnu > --host=x86_64-linux-gnu" PETSC_ARCH=rd > [0]PETSC ERROR: #1 PetscIntMultError() line 2309 in > C:/msys64/home/sd/petsc/rd/include/petscsys.h > [0]PETSC ERROR: #2 BVCreate_Svec() line 452 in > C:/msys64/home/sd/petsc/petsc-3.14.3/rd/externalpackages/git.slepc/src/sys/classes/bv/impls/svec/svec.c > [0]PETSC ERROR: #3 BVSetSizesFromVec() line 186 in > C:/msys64/home/sd/petsc/petsc-3.14.3/rd/externalpackages/git.slepc/src/sys/classes/bv/interface/bvbasic.c[0]PETSC > ERROR: #4 EPSAllocateSolution() line 687 in > C:/msys64/home/sd/petsc/petsc-3.14.3/rd/externalpackages/git.slepc/src/eps/interface/epssetup.c > [0]PETSC ERROR: #5 EPSSetUp_KrylovSchur() line 159 in > C:/msys64/home/sd/petsc/petsc-3.14.3/rd/externalpackages/git.slepc/src/eps/impls/krylov/krylovschur/krylovschur.c > [0]PETSC ERROR: #6 EPSSetUp() line 315 in > C:/msys64/home/sd/petsc/petsc-3.14.3/rd/externalpackages/git.slepc/src/eps/interface/epssetup.c > > > > > Sincerely, > > Xiaofeng > > > >> On Apr 30, 2025, at 13:08, Pierre Jolivet <pie...@joliv.et> wrote: >> >> Could you please provide the full back trace? >> Depending on your set of options, it may be as simple as switching -bv_type >> to make your code run (if you are using svec, this would explain such an >> error but could be circumvented with something else, like mat). >> >> Thanks, >> Pierre >> >>>> On 30 Apr 2025, at 6:27 AM, Satish Balay <balay....@fastmail.org> wrote: >>> >>>> On Wed, 30 Apr 2025, hexioafeng via petsc-users wrote: >>> >>>> Dear PETSc developers, >>>> >>>> I use PETSc and SLEPC to solve generalized eigen problems. When solving an >>>> interval eigen problem with matrix size about 5 million, i got the error >>>> message: "product of two integer xx xx overflow, you must ./configure >>>> PETSc with --with-64-bit-indices for the case you are running". >>>> >>>> I use some prebuilt third-party packages when building PETSc, namely >>>> OpenBLAS, METIS, ParMETIS and SCALAPACK. I wonder should i also use 64-bit >>>> prebuilt packages when configure PETSc with the --with-64-bit-indices >>>> flag? How about the MUMPS and MPI? Do i have to also use the b4-bit >>>> version? >>> >>> Hm - metis/parmetis would need a rebuild [with -DMETIS_USE_LONGINDEX=1 >>> option]. Others should be unaffected. >>> >>> You could use petsc configure to build pkgs to ensure compatibility i.e. >>> use --download-metis --download-parmetis etc.. >>> >>> Note - there is a difference between --with-64-bit-indices (PetscInt) and >>> --with-64-bit-blas-indices (PetscBlasInt) [and ILP64 - aka fortran '-i8'] >>> >>> Satish >>> >>> ---- >>> >>> $ grep defaultIndexSize config/BuildSystem/config/packages/*.py >>> config/BuildSystem/config/packages/hypre.py: if self.defaultIndexSize == >>> 64: >>> config/BuildSystem/config/packages/metis.py: if self.defaultIndexSize >>> == 64: >>> config/BuildSystem/config/packages/mkl_cpardiso.py: elif >>> self.blasLapack.has64bitindices and not self.defaultIndexSize == 64: >>> config/BuildSystem/config/packages/mkl_cpardiso.py: elif not >>> self.blasLapack.has64bitindices and self.defaultIndexSize == 64: >>> config/BuildSystem/config/packages/mkl_pardiso.py: elif >>> self.blasLapack.has64bitindices and not self.defaultIndexSize == 64: >>> config/BuildSystem/config/packages/mkl_sparse_optimize.py: if not >>> self.blasLapack.mkl or (not self.blasLapack.has64bitindices and >>> self.defaultIndexSize == 64): >>> config/BuildSystem/config/packages/mkl_sparse.py: if not >>> self.blasLapack.mkl or (not self.blasLapack.has64bitindices and >>> self.defaultIndexSize == 64): >>> config/BuildSystem/config/packages/SuperLU_DIST.py: if >>> self.defaultIndexSize == 64: >>> >>> >>> >>>> >>>> Look forward for your reply, thanks. >>>> >>>> Xiaofeng >>> >