I get:
[mae_yuexh@login01 ~]$ orte-info |grep 'MCA btl'
                 MCA btl: smcuda 
(MCA v2.1, API v3.1, Component v4.1.5)
                 MCA btl: tcp (MCA 
v2.1, API v3.1, Component v4.1.5)
                 MCA btl: self 
(MCA v2.1, API v3.1, Component v4.1.5)
                 MCA btl: vader 
(MCA v2.1, API v3.1, Component v4.1.5)



Xinhai





岳新海



南方科技大学/学生/研究生/2023级研究生



广东省深圳市南山区学苑大道1088号




 
 
 
------------------ Original ------------------
From: &nbsp;"Satish&nbsp;Balay"<[email protected]&gt;;
Date: &nbsp;Tue, Sep 23, 2025 03:25 AM
To: &nbsp;"岳新海"<[email protected]&gt;; 
Cc: &nbsp;"petsc-dev"<[email protected]&gt;; 
Subject: &nbsp;Re: [petsc-dev] Question on PETSc + CUDA configuration with MPI 
on  cluster

&nbsp;

 
What&nbsp;do&nbsp;you&nbsp;get&nbsp;for&nbsp;(with&nbsp;your&nbsp;openmpi&nbsp;install)&nbsp;:orte-info&nbsp;|grep&nbsp;'MCA&nbsp;btl'

With&nbsp;cuda&nbsp;built&nbsp;openmpi&nbsp;-&nbsp;I&nbsp;get:
balay@petsc-gpu-01:/scratch/balay/petsc$&nbsp;./arch-linux-c-debug/bin/orte-info&nbsp;|grep&nbsp;'MCA&nbsp;btl'
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;smcuda&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;openib&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;self&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;tcp&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;vader&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)

And&nbsp;without&nbsp;cuda:
balay@petsc-gpu-01:/scratch/balay/petsc.x$&nbsp;./arch-test/bin/orte-info&nbsp;&nbsp;|&nbsp;grep&nbsp;'MCA&nbsp;btl'
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;openib&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;self&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;tcp&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;MCA&nbsp;btl:&nbsp;vader&nbsp;(MCA&nbsp;v2.1,&nbsp;API&nbsp;v3.1,&nbsp;Component&nbsp;v4.1.6)

i.e&nbsp;"smcuda"&nbsp;should&nbsp;be&nbsp;listed&nbsp;for&nbsp;a&nbsp;cuda&nbsp;enabled&nbsp;openmpi.

Its&nbsp;not&nbsp;clear&nbsp;if&nbsp;GPU-aware&nbsp;MPI&nbsp;makes&nbsp;a&nbsp;difference&nbsp;for&nbsp;all&nbsp;MPI&nbsp;impls&nbsp;(or&nbsp;versions)&nbsp;-&nbsp;so&nbsp;good&nbsp;to&nbsp;verify.&nbsp;[its&nbsp;a&nbsp;performance&nbsp;issue&nbsp;anyway&nbsp;-&nbsp;so&nbsp;primarily&nbsp;useful&nbsp;when&nbsp;performing&nbsp;timing&nbsp;measurements]

Satish

On&nbsp;Mon,&nbsp;22&nbsp;Sep&nbsp;2025,&nbsp;岳新海&nbsp;wrote:

&gt;&nbsp;Dear&nbsp;PETSc&nbsp;Team,
&gt;&nbsp;&nbsp;
&gt;&nbsp;I&nbsp;am&nbsp;encountering&nbsp;an&nbsp;issue&nbsp;when&nbsp;running&nbsp;PETSc&nbsp;with&nbsp;CUDA&nbsp;support&nbsp;on&nbsp;a&nbsp;cluster.&nbsp;When&nbsp;I&nbsp;set&nbsp;the&nbsp;vector&nbsp;type&nbsp;to&nbsp;VECCUDA,&nbsp;PETSc&nbsp;reports&nbsp;that&nbsp;my&nbsp;MPI&nbsp;is&nbsp;not&nbsp;GPU-aware.&nbsp;However,&nbsp;the&nbsp;MPI&nbsp;library&nbsp;(OpenMPI&nbsp;4.1.5)&nbsp;I&nbsp;used&nbsp;to&nbsp;configure&nbsp;PETSc&nbsp;was&nbsp;built&nbsp;with&nbsp;the&nbsp;--with-cuda&nbsp;option&nbsp;enabled.
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;Here&nbsp;are&nbsp;some&nbsp;details:
&gt;&nbsp;PETSc&nbsp;version:&nbsp;3.20.6
&gt;&nbsp;MPI:&nbsp;OpenMPI&nbsp;4.1.5,&nbsp;configured&nbsp;with&nbsp;--with-cuda
&gt;&nbsp;GPU:&nbsp;RTX3090
&gt;&nbsp;CUDA&nbsp;version:&nbsp;12.1&nbsp;
&gt;&nbsp;I&nbsp;have&nbsp;attached&nbsp;both&nbsp;my&nbsp;PETSc&nbsp;configure&nbsp;command&nbsp;and&nbsp;OpenMPI&nbsp;configure&nbsp;command&nbsp;for&nbsp;reference.
&gt;&nbsp;
&gt;&nbsp;My&nbsp;questions&nbsp;are:
&gt;&nbsp;
&gt;&nbsp;&nbsp;
&gt;&nbsp;&nbsp;
&gt;&nbsp;&nbsp;
&gt;&nbsp;Even&nbsp;though&nbsp;I&nbsp;enabled&nbsp;--with-cuda&nbsp;in&nbsp;OpenMPI,&nbsp;why&nbsp;does&nbsp;PETSc&nbsp;still&nbsp;report&nbsp;that&nbsp;MPI&nbsp;is&nbsp;not&nbsp;GPU-aware?
&gt;&nbsp;&nbsp;
&gt;&nbsp;&nbsp;
&gt;&nbsp;&nbsp;
&gt;&nbsp;Are&nbsp;there&nbsp;additional&nbsp;steps&nbsp;or&nbsp;specific&nbsp;configuration&nbsp;flags&nbsp;required&nbsp;(either&nbsp;in&nbsp;OpenMPI&nbsp;or&nbsp;PETSc)&nbsp;to&nbsp;ensure&nbsp;GPU-aware&nbsp;MPI&nbsp;is&nbsp;correctly&nbsp;detected?
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;Any&nbsp;guidance&nbsp;or&nbsp;suggestions&nbsp;would&nbsp;be&nbsp;greatly&nbsp;appreciated.
&gt;&nbsp;
&gt;&nbsp;&nbsp;
&gt;&nbsp;
&gt;&nbsp;Best&nbsp;regards,
&gt;&nbsp;
&gt;&nbsp;Xinhai&nbsp;Yue
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;岳新海
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;南方科技大学/学生/研究生/2023级研究生
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;广东省深圳市南山区学苑大道1088号
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;
&gt;&nbsp;&nbsp;

Reply via email to