I get: [mae_yuexh@login01 ~]$ orte-info |grep 'MCA btl' MCA btl: smcuda (MCA v2.1, API v3.1, Component v4.1.5) MCA btl: tcp (MCA v2.1, API v3.1, Component v4.1.5) MCA btl: self (MCA v2.1, API v3.1, Component v4.1.5) MCA btl: vader (MCA v2.1, API v3.1, Component v4.1.5)
Xinhai 岳新海 南方科技大学/学生/研究生/2023级研究生 广东省深圳市南山区学苑大道1088号 ------------------ Original ------------------ From: "Satish Balay"<[email protected]>; Date: Tue, Sep 23, 2025 03:25 AM To: "岳新海"<[email protected]>; Cc: "petsc-dev"<[email protected]>; Subject: Re: [petsc-dev] Question on PETSc + CUDA configuration with MPI on cluster What do you get for (with your openmpi install) :orte-info |grep 'MCA btl' With cuda built openmpi - I get: balay@petsc-gpu-01:/scratch/balay/petsc$ ./arch-linux-c-debug/bin/orte-info |grep 'MCA btl' MCA btl: smcuda (MCA v2.1, API v3.1, Component v4.1.6) MCA btl: openib (MCA v2.1, API v3.1, Component v4.1.6) MCA btl: self (MCA v2.1, API v3.1, Component v4.1.6) MCA btl: tcp (MCA v2.1, API v3.1, Component v4.1.6) MCA btl: vader (MCA v2.1, API v3.1, Component v4.1.6) And without cuda: balay@petsc-gpu-01:/scratch/balay/petsc.x$ ./arch-test/bin/orte-info | grep 'MCA btl' MCA btl: openib (MCA v2.1, API v3.1, Component v4.1.6) MCA btl: self (MCA v2.1, API v3.1, Component v4.1.6) MCA btl: tcp (MCA v2.1, API v3.1, Component v4.1.6) MCA btl: vader (MCA v2.1, API v3.1, Component v4.1.6) i.e "smcuda" should be listed for a cuda enabled openmpi. Its not clear if GPU-aware MPI makes a difference for all MPI impls (or versions) - so good to verify. [its a performance issue anyway - so primarily useful when performing timing measurements] Satish On Mon, 22 Sep 2025, 岳新海 wrote: > Dear PETSc Team, > > I am encountering an issue when running PETSc with CUDA support on a cluster. When I set the vector type to VECCUDA, PETSc reports that my MPI is not GPU-aware. However, the MPI library (OpenMPI 4.1.5) I used to configure PETSc was built with the --with-cuda option enabled. > > > Here are some details: > PETSc version: 3.20.6 > MPI: OpenMPI 4.1.5, configured with --with-cuda > GPU: RTX3090 > CUDA version: 12.1 > I have attached both my PETSc configure command and OpenMPI configure command for reference. > > My questions are: > > > > > Even though I enabled --with-cuda in OpenMPI, why does PETSc still report that MPI is not GPU-aware? > > > > Are there additional steps or specific configuration flags required (either in OpenMPI or PETSc) to ensure GPU-aware MPI is correctly detected? > > > Any guidance or suggestions would be greatly appreciated. > > > > Best regards, > > Xinhai Yue > > > > > > > > > > > > 岳新海 > > > > 南方科技大学/学生/研究生/2023级研究生 > > > > 广东省深圳市南山区学苑大道1088号 > > > > >
