Dear Gilles,

Great! By the way, there was only one missing value about the 
ucx device to use in the second flag suggested by you, e.g. 
I had to use:

mpirun --mca pml_ucx_tls any --mca pml_ucx_devices any --mca pml ucx ...

with the release 4.1.1 to recover the performance of the 
previous one 4.1.0.

Thank you very much for your time in responding 
and for the link:

https://github.com/open-mpi/ompi/pull/8549

where the changes were introduced.

Perhaps, it would help to add a short note (or 
links to official or unofficial pages) about the 
optional UCX environment parameters in the manual 
for the mpirun command?


Greetings. 
Jorge.

----- Mensaje original -----
> De: "Gilles Gouaillardet" 
> Para: "Jorge D'Elia" , "Open MPI Users" <users@lists.open-mpi.org>
> Enviado: Lunes, 13 de Septiembre 2021 9:35:34
> Asunto: Re: [OMPI users] (Fedora 34, x86_64-pc-linux-gnu, 
> openmpi-4.1.1.tar.gz): PML ucx cannot be selected
>
> Jorge,
> 
> I am not that familiar with UCX, but I hope that will help:
> 
> The changes I mentioned were introduced by
> https://github.com/open-mpi/ompi/pull/8549
> 
> I suspect mpirun --mca pml_ucx_tls any --mca pml_ucx_devices --mca pml ucx
> ...
> 
> will do what you expect
> 
> Cheers,
> 
> Gilles
>
> On Mon, Sep 13, 2021 at 9:05 PM Jorge D'Elia via users <
> users@lists.open-mpi.org> wrote:
> 
>> Dear Gilles,
>>
>> Despite my last answer (see below), I am noticing that
>> some tests with a coarray fortran code on a laptop show a
>> performance drop of the order of 20% using the 4.1.1 version
>> (with --mca pml ucx disabled), versus the 4.1.0 one
>> (with --mca pml ucx enabled).
>>
>> I would like to experiment with pml/ucx framework using the 4.1.0
>> version on that laptop. Then, please, how do I manually re-enable
>> those providers? (e.g. perhaps, is it during the construction
>> stage?) or where can I find out how to do it? Thanks in advance.
>>
>> Regards.
>> Jorge.
>>
>> ----- Mensaje original -----
>> > De: "Open MPI Users" <users@lists.open-mpi.org>
>> > Para: "Open MPI Users" <users@lists.open-mpi.org>
>> > CC: "Jorge D'Elia"
>> > Enviado: Sábado, 29 de Mayo 2021 7:18:23
>> > Asunto: Re: [OMPI users] (Fedora 34, x86_64-pc-linux-gnu,
>> openmpi-4.1.1.tar.gz): PML ucx cannot be selected
>> >
>> > Dear Gilles,
>> >
>> > Ahhh ... now the new behavior is better understood.
>> > The intention of using pml/ucx was simply for preliminary
>> > testing, and does not merit re-enabling these providers in
>> > this notebook.
>> >
>> > Thank you very much for the clarification.
>> >
>> > Regards,
>> > Jorge.
>> >
>> > ----- Mensaje original -----
>> >> De: "Gilles Gouaillardet"
>> >> Para: "Jorge D'Elia" , "Open MPI Users" <users@lists.open-mpi.org>
>> >> Enviado: Viernes, 28 de Mayo 2021 23:35:37
>> >> Asunto: Re: [OMPI users] (Fedora 34, x86_64-pc-linux-gnu,
>> openmpi-4.1.1.tar.gz):
>> >> PML ucx cannot be selected
>> >>
>> >> Jorge,
>> >>
>> >> pml/ucx used to be selected when no fast interconnect were detected
>> >> (since ucx provides driver for both TCP and shared memory).
>> >> These providers are now disabled by default, so unless your machine
>> >> has a supported fast interconnect (such as Infiniband),
>> >> pml/ucx cannot be used out of the box anymore.
>> >>
>> >> if you really want to use pml/ucx on your notebook, you need to
>> >> manually re-enable these providers.
>> >>
>> >> That being said, your best choice here is really not to force any pml,
>> >> and let Open MPI use pml/ob1
>> >> (that has support for both TCP and shared memory)
>> >>
>> >> Cheers,
>> >>
>> >> Gilles
>> >>
>> >> On Sat, May 29, 2021 at 11:19 AM Jorge D'Elia via users
>> >> <users@lists.open-mpi.org> wrote:
>> >>>
>> >>> Hi,
>> >>>
>> >>> We routinely build OpenMPI on x86_64-pc-linux-gnu machines from
>> >>> the sources using gcc and usually everything works fine.
>> >>>
>> >>> In one case we recently installed Fedora 34 from scratch on an
>> >>> ASUS G53SX notebook (Intel Core i7-2630QM CPU 2.00GHz ×4 cores,
>> >>> without any IB device). Next we build OpenMPI using the file
>> >>> openmpi-4.1.1.tar.gz and the GCC 12.0.0 20210524 (experimental)
>> >>> compiler.
>> >>>
>> >>> However, when trying to experiment OpenMPI using UCX
>> >>> with a simple test, we get the runtime errors:
>> >>>
>> >>>   No components were able to be opened in the btl framework.
>> >>>   PML ucx cannot be selected
>> >>>
>> >>> while the test worked fine until Fedora 33 on the same
>> >>> machine using the same OpenMPI configuration.
>> >>>
>> >>> We attach below some info about a simple test run.
>> >>>
>> >>> Please, any clues where to check or maybe something is missing?
>> >>> Thanks in advance.
>> >>>
>> >>> Regards
>> >>> Jorge.
>> >>>
>> >>> --
>> >>> $ cat /proc/version
>> >>> Linux version 5.12.7-300.fc34.x86_64
>> >>> (mockbu...@bkernel01.iad2.fedoraproject.org) (gcc (GCC) 11.1.1
>> 20210428 (Red
>> >>> Hat 11.1.1-1), GNU ld version 2.35.1-41.fc34) #1 SMP Wed May 26
>> 12:58:58 UTC
>> >>> 2021
>> >>>
>> >>> $ mpifort --version
>> >>> GNU Fortran (GCC) 12.0.0 20210524 (experimental)
>> >>> Copyright (C) 2021 Free Software Foundation, Inc.
>> >>>
>> >>> $ which mpifort
>> >>> /usr/beta/openmpi/bin/mpifort
>> >>>
>> >>> $ mpifort -o hello_usempi_f08.exe hello_usempi_f08.f90
>> >>>
>> >>> $ mpirun --mca orte_base_help_aggregate 0 --mca btl self,vader,tcp
>> --map-by node
>> >>> --report-bindings --machinefile ~/machi-openmpi.dat --np 2
>> >>> hello_usempi_f08.exe
>> >>> [verne:200650] MCW rank 0 bound to socket 0[core 0[hwt 0]]: [B/././.]
>> >>> [verne:200650] MCW rank 1 bound to socket 0[core 1[hwt 0]]: [./B/./.]
>> >>> Hello, world, I am  0 of  2: Open MPI v4.1.1, package: Open MPI
>> bigpack@verne
>> >>> Distribution, ident: 4.1.1, repo rev: v4.1.1, Apr 24, 2021
>> >>> Hello, world, I am  1 of  2: Open MPI v4.1.1, package: Open MPI
>> bigpack@verne
>> >>> Distribution, ident: 4.1.1, repo rev: v4.1.1, Apr 24, 2021
>> >>>
>> >>> $ mpirun --mca orte_base_help_aggregate 0 --mca pml ucx --mca btl
>> >>> ^self,vader,tcp --map-by node --report-bindings --machinefile
>> >>> ~/machi-openmpi.dat --np 2  hello_usempi_f08.exe
>> >>> [verne:200772] MCW rank 0 bound to socket 0[core 0[hwt 0]]: [B/././.]
>> >>> [verne:200772] MCW rank 1 bound to socket 0[core 1[hwt 0]]: [./B/./.]
>> >>>
>> --------------------------------------------------------------------------
>> >>> No components were able to be opened in the btl framework.
>> >>>
>> >>> This typically means that either no components of this type were
>> >>> installed, or none of the installed components can be loaded.
>> >>> Sometimes this means that shared libraries required by these
>> >>> components are unable to be found/loaded.
>> >>>
>> >>>   Host:      verne
>> >>>   Framework: btl
>> >>>
>> --------------------------------------------------------------------------
>> >>>
>> --------------------------------------------------------------------------
>> >>> No components were able to be opened in the btl framework.
>> >>>
>> >>> This typically means that either no components of this type were
>> >>> installed, or none of the installed components can be loaded.
>> >>> Sometimes this means that shared libraries required by these
>> >>> components are unable to be found/loaded.
>> >>>
>> >>>   Host:      verne
>> >>>   Framework: btl
>> >>>
>> --------------------------------------------------------------------------
>> >>>
>> --------------------------------------------------------------------------
>> >>> No components were able to be opened in the pml framework.
>> >>>
>> >>> This typically means that either no components of this type were
>> >>> installed, or none of the installed components can be loaded.
>> >>> Sometimes this means that shared libraries required by these
>> >>> components are unable to be found/loaded.
>> >>>
>> >>>   Host:      verne
>> >>>   Framework: pml
>> >>>
>> --------------------------------------------------------------------------
>> >>> [verne:200777] PML ucx cannot be selected
>> >>>
>> --------------------------------------------------------------------------
>> >>> No components were able to be opened in the pml framework.
>> >>>
>> >>> This typically means that either no components of this type were
>> >>> installed, or none of the installed components can be loaded.
>> >>> Sometimes this means that shared libraries required by these
>> >>> components are unable to be found/loaded.
>> >>>
>> >>>   Host:      verne
>> >>>   Framework: pml
>> >>>
>> --------------------------------------------------------------------------
>> >>> [verne:200772] PMIX ERROR: UNREACHABLE in file
>> >>>
>> ../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at
>> line
>> >>> 2198
>> >>>
>> >>>
>> >>> $ ompi_info | grep ucx
>> >>>   Configure command line: '--enable-ipv6' '--enable-sparse-groups'
>> >>>   '--enable-mpi-ext' '--enable-mpi-cxx' '--enable-oshmem'
>> >>>   '--with-libevent=internal' '--with-ucx' '--with-pmix=internal'
>> >>>   '--without-libfabric' '--prefix=/usr/beta/openmpi'
>> >>>                  MCA osc: ucx (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA pml: ucx (MCA v2.1.0, API v2.0.0, Component
>> >>>                  v4.1.1)
>> >>>
>> >>> $ ompi_info --param all all --level 9 | grep ucx
>> >>>                  MCA osc: ucx (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA pml: ucx (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>              MCA osc ucx:
>> ---------------------------------------------------
>> >>>              MCA osc ucx: parameter "osc_ucx_priority" (current value:
>> "60", data source:
>> >>>              default, level: 3 user/all, type: unsigned_int)
>> >>>                           Priority of the osc/ucx component (default:
>> 60)
>> >>>              MCA osc ucx: parameter "osc_ucx_verbose" (current value:
>> "0", data source:
>> >>>              default, level: 3 user/all, type: int, synonym of:
>> opal_common_ucx_verbose)
>> >>>              MCA osc ucx: parameter "osc_ucx_progress_iterations"
>> (current value: "100", data
>> >>>              source: default, level: 3 user/all, type: int, synonym of:
>> >>>              opal_common_ucx_progress_iterations)
>> >>>              MCA osc ucx: parameter "osc_ucx_opal_mem_hooks" (current
>> value: "false", data
>> >>>              source: default, level: 3 user/all, type: bool, synonym
>> of:
>> >>>              opal_common_ucx_opal_mem_hooks)
>> >>>              MCA osc ucx: parameter "osc_ucx_tls" (current value:
>> >>>              "rc_verbs,ud_verbs,rc_mlx5,dc_mlx5,cuda_ipc,rocm_ipc",
>> data source: default,
>> >>>              level: 3 user/all, type: string, synonym of:
>> opal_common_ucx_tls)
>> >>>              MCA osc ucx: parameter "osc_ucx_devices" (current value:
>> "mlx*", data source:
>> >>>              default, level: 3 user/all, type: string, synonym of:
>> opal_common_ucx_devices)
>> >>>              MCA pml ucx:
>> ---------------------------------------------------
>> >>>              MCA pml ucx: parameter "pml_ucx_priority" (current value:
>> "51", data source:
>> >>>              default, level: 3 user/all, type: int)
>> >>>              MCA pml ucx: parameter "pml_ucx_num_disconnect" (current
>> value: "1", data
>> >>>              source: default, level: 3 user/all, type: int)
>> >>>              MCA pml ucx: parameter "pml_ucx_verbose" (current value:
>> "0", data source:
>> >>>              default, level: 3 user/all, type: int, synonym of:
>> opal_common_ucx_verbose)
>> >>>              MCA pml ucx: parameter "pml_ucx_progress_iterations"
>> (current value: "100", data
>> >>>              source: default, level: 3 user/all, type: int, synonym of:
>> >>>              opal_common_ucx_progress_iterations)
>> >>>              MCA pml ucx: parameter "pml_ucx_opal_mem_hooks" (current
>> value: "false", data
>> >>>              source: default, level: 3 user/all, type: bool, synonym
>> of:
>> >>>              opal_common_ucx_opal_mem_hooks)
>> >>>              MCA pml ucx: parameter "pml_ucx_tls" (current value:
>> >>>              "rc_verbs,ud_verbs,rc_mlx5,dc_mlx5,cuda_ipc,rocm_ipc",
>> data source: default,
>> >>>              level: 3 user/all, type: string, synonym of:
>> opal_common_ucx_tls)
>> >>>              MCA pml ucx: parameter "pml_ucx_devices" (current value:
>> "mlx*", data source:
>> >>>              default, level: 3 user/all, type: string, synonym of:
>> opal_common_ucx_devices)
>> >>>
>> >>> $ ompi_info
>> >>>                  Package: Open MPI bigpack@verne Distribution
>> >>>                 Open MPI: 4.1.1
>> >>>   Open MPI repo revision: v4.1.1
>> >>>    Open MPI release date: Apr 24, 2021
>> >>>                 Open RTE: 4.1.1
>> >>>   Open RTE repo revision: v4.1.1
>> >>>    Open RTE release date: Apr 24, 2021
>> >>>                     OPAL: 4.1.1
>> >>>       OPAL repo revision: v4.1.1
>> >>>        OPAL release date: Apr 24, 2021
>> >>>                  MPI API: 3.1.0
>> >>>             Ident string: 4.1.1
>> >>>                   Prefix: /usr/beta/openmpi
>> >>>  Configured architecture: x86_64-pc-linux-gnu
>> >>>           Configure host: verne
>> >>>            Configured by: bigpack
>> >>>            Configured on: Tue May 25 17:16:38 UTC 2021
>> >>>           Configure host: verne
>> >>>   Configure command line: '--enable-ipv6' '--enable-sparse-groups'
>> >>>                           '--enable-mpi-ext' '--enable-mpi-cxx'
>> >>>                           '--enable-oshmem' '--with-libevent=internal'
>> >>>                           '--with-ucx' '--with-pmix=internal'
>> >>>                           '--without-libfabric'
>> '--prefix=/usr/beta/openmpi'
>> >>>                 Built by: bigpack
>> >>>                 Built on: Tue 25 May 17:57:46 UTC 2021
>> >>>               Built host: verne
>> >>>               C bindings: yes
>> >>>             C++ bindings: yes
>> >>>              Fort mpif.h: yes (all)
>> >>>             Fort use mpi: yes (full: ignore TKR)
>> >>>        Fort use mpi size: deprecated-ompi-info-value
>> >>>         Fort use mpi_f08: yes
>> >>>  Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
>> >>>                           limitations in the gfortran compiler and/or
>> Open
>> >>>                           MPI, does not support the following: array
>> >>>                           subsections, direct passthru (where
>> possible) to
>> >>>                           underlying Open MPI's C functionality
>> >>>   Fort mpi_f08 subarrays: no
>> >>>            Java bindings: no
>> >>>   Wrapper compiler rpath: runpath
>> >>>               C compiler: gcc
>> >>>      C compiler absolute: /usr/beta/gcc-trunk/bin/gcc
>> >>>   C compiler family name: GNU
>> >>>       C compiler version: 12.0.0
>> >>>             C++ compiler: g++
>> >>>    C++ compiler absolute: /usr/beta/gcc-trunk/bin/g++
>> >>>            Fort compiler: gfortran
>> >>>        Fort compiler abs: /usr/beta/gcc-trunk/bin/gfortran
>> >>>          Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::)
>> >>>    Fort 08 assumed shape: yes
>> >>>       Fort optional args: yes
>> >>>           Fort INTERFACE: yes
>> >>>     Fort ISO_FORTRAN_ENV: yes
>> >>>        Fort STORAGE_SIZE: yes
>> >>>       Fort BIND(C) (all): yes
>> >>>       Fort ISO_C_BINDING: yes
>> >>>  Fort SUBROUTINE BIND(C): yes
>> >>>        Fort TYPE,BIND(C): yes
>> >>>  Fort T,BIND(C,name="a"): yes
>> >>>             Fort PRIVATE: yes
>> >>>           Fort PROTECTED: yes
>> >>>            Fort ABSTRACT: yes
>> >>>        Fort ASYNCHRONOUS: yes
>> >>>           Fort PROCEDURE: yes
>> >>>          Fort USE...ONLY: yes
>> >>>            Fort C_FUNLOC: yes
>> >>>  Fort f08 using wrappers: yes
>> >>>          Fort MPI_SIZEOF: yes
>> >>>              C profiling: yes
>> >>>            C++ profiling: yes
>> >>>    Fort mpif.h profiling: yes
>> >>>   Fort use mpi profiling: yes
>> >>>    Fort use mpi_f08 prof: yes
>> >>>           C++ exceptions: no
>> >>>           Thread support: posix (MPI_THREAD_MULTIPLE: yes, OPAL
>> support: yes,
>> >>>                           OMPI progress: no, ORTE progress: yes, Event
>> lib:
>> >>>                           yes)
>> >>>            Sparse Groups: yes
>> >>>   Internal debug support: no
>> >>>   MPI interface warnings: yes
>> >>>      MPI parameter check: runtime
>> >>> Memory profiling support: no
>> >>> Memory debugging support: no
>> >>>               dl support: yes
>> >>>    Heterogeneous support: no
>> >>>  mpirun default --prefix: no
>> >>>        MPI_WTIME support: native
>> >>>      Symbol vis. support: yes
>> >>>    Host topology support: yes
>> >>>             IPv6 support: yes
>> >>>       MPI1 compatibility: no
>> >>>           MPI extensions: affinity, cuda, pcollreq
>> >>>    FT Checkpoint support: no (checkpoint thread: no)
>> >>>    C/R Enabled Debugging: no
>> >>>   MPI_MAX_PROCESSOR_NAME: 256
>> >>>     MPI_MAX_ERROR_STRING: 256
>> >>>      MPI_MAX_OBJECT_NAME: 64
>> >>>         MPI_MAX_INFO_KEY: 36
>> >>>         MPI_MAX_INFO_VAL: 256
>> >>>        MPI_MAX_PORT_NAME: 1024
>> >>>   MPI_MAX_DATAREP_STRING: 128
>> >>>            MCA allocator: basic (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>            MCA allocator: bucket (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>            MCA backtrace: execinfo (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA btl: self (MCA v2.1.0, API v3.1.0, Component
>> v4.1.1)
>> >>>                  MCA btl: openib (MCA v2.1.0, API v3.1.0, Component
>> v4.1.1)
>> >>>                  MCA btl: tcp (MCA v2.1.0, API v3.1.0, Component
>> v4.1.1)
>> >>>                  MCA btl: vader (MCA v2.1.0, API v3.1.0, Component
>> v4.1.1)
>> >>>             MCA compress: bzip (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>             MCA compress: gzip (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA crs: none (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                   MCA dl: dlopen (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                MCA event: libevent2022 (MCA v2.1.0, API v2.0.0,
>> Component
>> >>>                           v4.1.1)
>> >>>                MCA hwloc: external (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                   MCA if: linux_ipv6 (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>                   MCA if: posix_ipv4 (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>          MCA installdirs: env (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>          MCA installdirs: config (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>               MCA memory: patcher (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA mpool: hugepage (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>              MCA patcher: overwrite (MCA v2.1.0, API v1.0.0, Component
>> >>>                           v4.1.1)
>> >>>                 MCA pmix: isolated (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA pmix: flux (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA pmix: pmix3x (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA pstat: linux (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>               MCA rcache: grdma (MCA v2.1.0, API v3.3.0, Component
>> v4.1.1)
>> >>>            MCA reachable: weighted (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA shmem: mmap (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA shmem: posix (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA shmem: sysv (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA timer: linux (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>               MCA errmgr: default_app (MCA v2.1.0, API v3.0.0,
>> Component
>> >>>                           v4.1.1)
>> >>>               MCA errmgr: default_hnp (MCA v2.1.0, API v3.0.0,
>> Component
>> >>>                           v4.1.1)
>> >>>               MCA errmgr: default_orted (MCA v2.1.0, API v3.0.0,
>> Component
>> >>>                           v4.1.1)
>> >>>               MCA errmgr: default_tool (MCA v2.1.0, API v3.0.0,
>> Component
>> >>>                           v4.1.1)
>> >>>                  MCA ess: env (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA ess: hnp (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA ess: pmi (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA ess: singleton (MCA v2.1.0, API v3.0.0, Component
>> >>>                           v4.1.1)
>> >>>                  MCA ess: tool (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA ess: slurm (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                MCA filem: raw (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>              MCA grpcomm: direct (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA iof: hnp (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA iof: orted (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA iof: tool (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA odls: default (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA odls: pspawn (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA oob: tcp (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA plm: isolated (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA plm: rsh (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA plm: slurm (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA ras: simulator (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>                  MCA ras: slurm (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA regx: fwd (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                 MCA regx: naive (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                 MCA regx: reverse (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                MCA rmaps: mindist (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA rmaps: ppr (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA rmaps: rank_file (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>                MCA rmaps: resilient (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>                MCA rmaps: round_robin (MCA v2.1.0, API v2.0.0,
>> Component
>> >>>                           v4.1.1)
>> >>>                MCA rmaps: seq (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA rml: oob (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>               MCA routed: binomial (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>               MCA routed: direct (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>               MCA routed: radix (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA rtc: hwloc (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>               MCA schizo: flux (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>               MCA schizo: ompi (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>               MCA schizo: orte (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>               MCA schizo: jsm (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>               MCA schizo: slurm (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                MCA state: app (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                MCA state: hnp (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                MCA state: novm (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                MCA state: orted (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                MCA state: tool (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                  MCA bml: r2 (MCA v2.1.0, API v2.0.0, Component v4.1.1)
>> >>>                 MCA coll: adapt (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA coll: basic (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA coll: han (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA coll: inter (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA coll: libnbc (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA coll: self (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA coll: sm (MCA v2.1.0, API v2.0.0, Component v4.1.1)
>> >>>                 MCA coll: sync (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA coll: tuned (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                 MCA coll: monitoring (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>                 MCA fbtl: posix (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA fcoll: dynamic (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                MCA fcoll: dynamic_gen2 (MCA v2.1.0, API v2.0.0,
>> Component
>> >>>                           v4.1.1)
>> >>>                MCA fcoll: individual (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>                MCA fcoll: two_phase (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>                MCA fcoll: vulcan (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                   MCA fs: ufs (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                   MCA io: ompio (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                   MCA io: romio321 (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                   MCA op: avx (MCA v2.1.0, API v1.0.0, Component
>> v4.1.1)
>> >>>                  MCA osc: sm (MCA v2.1.0, API v3.0.0, Component v4.1.1)
>> >>>                  MCA osc: monitoring (MCA v2.1.0, API v3.0.0, Component
>> >>>                           v4.1.1)
>> >>>                  MCA osc: pt2pt (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA osc: rdma (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA osc: ucx (MCA v2.1.0, API v3.0.0, Component
>> v4.1.1)
>> >>>                  MCA pml: v (MCA v2.1.0, API v2.0.0, Component v4.1.1)
>> >>>                  MCA pml: cm (MCA v2.1.0, API v2.0.0, Component v4.1.1)
>> >>>                  MCA pml: monitoring (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>                  MCA pml: ob1 (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA pml: ucx (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>                  MCA rte: orte (MCA v2.1.0, API v2.0.0, Component
>> v4.1.1)
>> >>>             MCA sharedfp: individual (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>             MCA sharedfp: lockedfile (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>             MCA sharedfp: sm (MCA v2.1.0, API v2.0.0, Component v4.1.1)
>> >>>                 MCA topo: basic (MCA v2.1.0, API v2.2.0, Component
>> v4.1.1)
>> >>>                 MCA topo: treematch (MCA v2.1.0, API v2.2.0, Component
>> >>>                           v4.1.1)
>> >>>            MCA vprotocol: pessimist (MCA v2.1.0, API v2.0.0, Component
>> >>>                           v4.1.1)
>> >>>
>> > > > #end

Reply via email to