Re: [OMPI users] psec warning when launching with srun

2023-05-23 Thread Zhéxué M. Krawutschke via users
Hello all,

yes, I have noticed that with the individual development libraries for a very 
long time, because I rather compile the entire products by HAND to my needs 
myself and do not want to rely on the libraries of the individual Linux - 
variants.

I have done very well with this at the moment. What I would like to do now is 
to test and implement this on a Rocky Linux system.
I can gladly report the results here

Best regards from Berlin

Z. Matthias

> On Samstag, Mai 20, 2023 at 1:29 PM, christof.koehler--- via users 
> mailto:users@lists.open-mpi.org)> wrote:
> Hello Gilles,
>
> thank you very much for the prompt patch.
>
> I can confirm that configure now prefers the external PMIx. I can
> confirm that the munge warnings and PMIx errors we observed are gone. An
> mpi hello world runs successfully with srun --mpi=pmix and --mpi=pmi2.
>
> I noticed that configure complained loudly about a missing external
> libevent (i.e. libevent-devel package), but did not complain at all
> that an external hwloc-devel was also missing.
>
> Best Regards
>
> Christof
>
>
>
> On Sat, May 20, 2023 at 06:54:54PM +0900, Gilles Gouaillardet wrote:
> > Christof,
> >
> > Open MPI switching to the internal PMIx is a bug I addressed in
> > https://github.com/open-mpi/ompi/pull/11704
> >
> > Feel free to manually download and apply the patch, you will then need
> > recent autotools and run
> > ./autogen.pl --force
> >
> > An other option is to manually edit the configure file
> >
> > Look for the following snippet
> >
> > # Final check - if they didn't point us explicitly at an
> > external version
> >
> > # but we found one anyway, use the internal version if it is
> > higher
> >
> > if test "$opal_external_pmix_version" != "internal" && (test -z
> > "$with_pmix" || test "$with_pmix" = "yes")
> >
> > then :
> >
> > if test "$opal_external_pmix_version" != "3x"
> >
> >
> > and replace the last line with
> >
> > if test $opal_external_pmix_version_major -lt 3
> >
> >
> > Cheers,
> >
> > Gilles
> >
> > On Sat, May 20, 2023 at 6:13 PM christof.koehler--- via users <
> > users@lists.open-mpi.org> wrote:
> >
> > > Hello Z. Matthias Krawutschke,
> > >
> > > On Fri, May 19, 2023 at 09:08:08PM +0200, Zhéxué M. Krawutschke wrote:
> > > > Hello Christoph,
> > > > what exactly is your problem with OpenMPI and Slurm?
> > > > Do you compile the products yourself? Which LINUX distribution and
> > > version are you using?
> > > >
> > > > If you compile the software yourself, could you please tell me what the
> > > "configure" command looks like and which MUNGE version is in use? From the
> > > distribution or compiled by yourself?
> > > >
> > > > I would be very happy to take on this topic and help you. You can also
> > > reach me at +49 176 67270992.
> > > > Best regards from Berlin
> > >
> > > please refer to (especially the end) of my first mail in this thread
> > > which is available here
> > > https://www.mail-archive.com/users@lists.open-mpi.org/msg35141.html
> > >
> > > I believe this contains the relevant information you are requesting. The
> > > second mail which you are replying to was just additional information.
> > > My apologies if this led to confusion.
> > >
> > > Please let me know if any relevant information is missing from my first
> > > email. At the bottom of this email I include the ompi_info output as
> > > further addendum.
> > >
> > > To summarize: I would like to understand where the munge warning
> > > and PMIx error described in the first email (and the github link
> > > included) come from. The explanation in the github issue
> > > does not appear to be correct as all munge libraries are
> > > available everywhere. To me, it appears at the moment that OpenMPIs
> > > configure decides erroneously to build and use the internal pmix
> > > instead of using the (presumably) newer externally available PMIx,
> > > leading to launcher problems with srun.
> > >
> > >
> > > Best Regards
> > >
> > > Christof
> > >
> > > Package: Open MPI root@admin.service Distribution
> > > Open MPI: 4.1.5
> > > Open MPI repo revision: v4.1.5
> > > Open MPI release date: Feb 23, 2023
> > > Open RTE: 4.1.5
> > > Open RTE repo revision: v4.1.5
> > > Open RTE release date: Feb 23, 2023
> > > OPAL: 4.1.5
> > > OPAL repo revision: v4.1.5
> > > OPAL release date: Feb 23, 2023
> > > MPI API: 3.1.0
> > > Ident string: 4.1.5
> > > Prefix: /cluster/mpi/openmpi/4.1.5/gcc-11.3.1
> > > Configured architecture: x86_64-pc-linux-gnu
> > > Configure host: admin.service
> > > Configured by: root
> > > Configured on: Wed May 17 18:45:42 UTC 2023
> > > Configure host: admin.service
> > > Configure command line: '--enable-mpi1-compatibility'
> > > '--enable-orterun-prefix-by-default'
> > > '--with-ofi=/cluster/libraries/libfabric/1.18.0/' '--with-slurm'
> > > '--with-pmix' '--with-pmix-libdir=/usr/lib64' '--with-pmi'
> > > '--with-pmi-libdir=/usr/lib64'
> > > '--prefix=/cluster/mpi/openmpi/4.1.5/gcc-11.3.1'
> > > Built by: root
> > > Built on: Wed May 

Re: [OMPI users] psec warning when launching with srun

2023-05-20 Thread christof.koehler--- via users
Hello Gilles,

thank you very much for the prompt patch. 

I can confirm that configure now prefers the external PMIx. I can
confirm that the munge warnings and PMIx errors we observed are gone. An
mpi hello world runs successfully with srun --mpi=pmix and --mpi=pmi2.

I noticed that configure complained loudly about a missing external
libevent (i.e. libevent-devel package), but did not complain at all
that an external hwloc-devel was also missing.

Best Regards

Christof



On Sat, May 20, 2023 at 06:54:54PM +0900, Gilles Gouaillardet wrote:
> Christof,
> 
> Open MPI switching to the internal PMIx is a bug I addressed in
> https://github.com/open-mpi/ompi/pull/11704
> 
> Feel free to manually download and apply the patch, you will then need
> recent autotools and run
> ./autogen.pl --force
> 
> An other option is to manually edit the configure file
> 
> Look for the following snippet
> 
># Final check - if they didn't point us explicitly at an
> external version
> 
># but we found one anyway, use the internal version if it is
> higher
> 
>if test "$opal_external_pmix_version" != "internal" && (test -z
> "$with_pmix" || test "$with_pmix" = "yes")
> 
> then :
> 
>   if test "$opal_external_pmix_version" != "3x"
> 
> 
> and replace the last line with
> 
>   if test $opal_external_pmix_version_major -lt 3
> 
> 
> Cheers,
> 
> Gilles
> 
> On Sat, May 20, 2023 at 6:13 PM christof.koehler--- via users <
> users@lists.open-mpi.org> wrote:
> 
> > Hello Z. Matthias Krawutschke,
> >
> > On Fri, May 19, 2023 at 09:08:08PM +0200, Zhéxué M. Krawutschke wrote:
> > > Hello Christoph,
> > > what exactly is your problem with OpenMPI and Slurm?
> > > Do you compile the products yourself? Which LINUX distribution and
> > version are you using?
> > >
> > > If you compile the software yourself, could you please tell me what the
> > "configure" command looks like and which MUNGE version is in use? From the
> > distribution or compiled by yourself?
> > >
> > > I would be very happy to take on this topic and help you. You can also
> > reach me at +49 176 67270992.
> > > Best regards from Berlin
> >
> > please refer to (especially the end) of my first mail in this thread
> > which is available here
> > https://www.mail-archive.com/users@lists.open-mpi.org/msg35141.html
> >
> > I believe this contains the relevant information you are requesting. The
> > second mail which you are replying to was just additional information.
> > My apologies if this led to confusion.
> >
> > Please let me know if any relevant information is missing from my first
> > email. At the bottom of this email I include the ompi_info output as
> > further addendum.
> >
> > To summarize: I would like to understand where the munge warning
> > and PMIx error described in the first email (and the github link
> > included) come from. The explanation in the github issue
> > does not appear to be correct as all munge libraries are
> > available everywhere. To me, it appears at the moment that OpenMPIs
> > configure decides erroneously to build and use the internal pmix
> > instead of using the (presumably) newer externally available PMIx,
> > leading to launcher problems with srun.
> >
> >
> > Best Regards
> >
> > Christof
> >
> >  Package: Open MPI root@admin.service Distribution
> > Open MPI: 4.1.5
> >   Open MPI repo revision: v4.1.5
> >Open MPI release date: Feb 23, 2023
> > Open RTE: 4.1.5
> >   Open RTE repo revision: v4.1.5
> >Open RTE release date: Feb 23, 2023
> > OPAL: 4.1.5
> >   OPAL repo revision: v4.1.5
> >OPAL release date: Feb 23, 2023
> >  MPI API: 3.1.0
> > Ident string: 4.1.5
> >   Prefix: /cluster/mpi/openmpi/4.1.5/gcc-11.3.1
> >  Configured architecture: x86_64-pc-linux-gnu
> >   Configure host: admin.service
> >Configured by: root
> >Configured on: Wed May 17 18:45:42 UTC 2023
> >   Configure host: admin.service
> >   Configure command line: '--enable-mpi1-compatibility'
> > '--enable-orterun-prefix-by-default'
> > '--with-ofi=/cluster/libraries/libfabric/1.18.0/' '--with-slurm'
> > '--with-pmix' '--with-pmix-libdir=/usr/lib64' '--with-pmi'
> > '--with-pmi-libdir=/usr/lib64'
> > '--prefix=/cluster/mpi/openmpi/4.1.5/gcc-11.3.1'
> > Built by: root
> > Built on: Wed May 17 06:48:36 PM UTC 2023
> >   Built host: admin.service
> >   C bindings: yes
> > C++ bindings: no
> >  Fort mpif.h: yes (all)
> > Fort use mpi: yes (full: ignore TKR)
> >Fort use mpi size: deprecated-ompi-info-value
> > Fort use mpi_f08: yes
> >  Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
> > limitations in the gfortran compiler and/or Open MPI, does not support
> > the following: array subsections, direct passthru (where possible) to
> > 

Re: [OMPI users] psec warning when launching with srun

2023-05-20 Thread Gilles Gouaillardet via users
Christof,

Open MPI switching to the internal PMIx is a bug I addressed in
https://github.com/open-mpi/ompi/pull/11704

Feel free to manually download and apply the patch, you will then need
recent autotools and run
./autogen.pl --force

An other option is to manually edit the configure file

Look for the following snippet

   # Final check - if they didn't point us explicitly at an
external version

   # but we found one anyway, use the internal version if it is
higher

   if test "$opal_external_pmix_version" != "internal" && (test -z
"$with_pmix" || test "$with_pmix" = "yes")

then :

  if test "$opal_external_pmix_version" != "3x"


and replace the last line with

  if test $opal_external_pmix_version_major -lt 3


Cheers,

Gilles

On Sat, May 20, 2023 at 6:13 PM christof.koehler--- via users <
users@lists.open-mpi.org> wrote:

> Hello Z. Matthias Krawutschke,
>
> On Fri, May 19, 2023 at 09:08:08PM +0200, Zhéxué M. Krawutschke wrote:
> > Hello Christoph,
> > what exactly is your problem with OpenMPI and Slurm?
> > Do you compile the products yourself? Which LINUX distribution and
> version are you using?
> >
> > If you compile the software yourself, could you please tell me what the
> "configure" command looks like and which MUNGE version is in use? From the
> distribution or compiled by yourself?
> >
> > I would be very happy to take on this topic and help you. You can also
> reach me at +49 176 67270992.
> > Best regards from Berlin
>
> please refer to (especially the end) of my first mail in this thread
> which is available here
> https://www.mail-archive.com/users@lists.open-mpi.org/msg35141.html
>
> I believe this contains the relevant information you are requesting. The
> second mail which you are replying to was just additional information.
> My apologies if this led to confusion.
>
> Please let me know if any relevant information is missing from my first
> email. At the bottom of this email I include the ompi_info output as
> further addendum.
>
> To summarize: I would like to understand where the munge warning
> and PMIx error described in the first email (and the github link
> included) come from. The explanation in the github issue
> does not appear to be correct as all munge libraries are
> available everywhere. To me, it appears at the moment that OpenMPIs
> configure decides erroneously to build and use the internal pmix
> instead of using the (presumably) newer externally available PMIx,
> leading to launcher problems with srun.
>
>
> Best Regards
>
> Christof
>
>  Package: Open MPI root@admin.service Distribution
> Open MPI: 4.1.5
>   Open MPI repo revision: v4.1.5
>Open MPI release date: Feb 23, 2023
> Open RTE: 4.1.5
>   Open RTE repo revision: v4.1.5
>Open RTE release date: Feb 23, 2023
> OPAL: 4.1.5
>   OPAL repo revision: v4.1.5
>OPAL release date: Feb 23, 2023
>  MPI API: 3.1.0
> Ident string: 4.1.5
>   Prefix: /cluster/mpi/openmpi/4.1.5/gcc-11.3.1
>  Configured architecture: x86_64-pc-linux-gnu
>   Configure host: admin.service
>Configured by: root
>Configured on: Wed May 17 18:45:42 UTC 2023
>   Configure host: admin.service
>   Configure command line: '--enable-mpi1-compatibility'
> '--enable-orterun-prefix-by-default'
> '--with-ofi=/cluster/libraries/libfabric/1.18.0/' '--with-slurm'
> '--with-pmix' '--with-pmix-libdir=/usr/lib64' '--with-pmi'
> '--with-pmi-libdir=/usr/lib64'
> '--prefix=/cluster/mpi/openmpi/4.1.5/gcc-11.3.1'
> Built by: root
> Built on: Wed May 17 06:48:36 PM UTC 2023
>   Built host: admin.service
>   C bindings: yes
> C++ bindings: no
>  Fort mpif.h: yes (all)
> Fort use mpi: yes (full: ignore TKR)
>Fort use mpi size: deprecated-ompi-info-value
> Fort use mpi_f08: yes
>  Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
> limitations in the gfortran compiler and/or Open MPI, does not support
> the following: array subsections, direct passthru (where possible) to
> underlying Open MPI's C functionality
>   Fort mpi_f08 subarrays: no
>Java bindings: no
>   Wrapper compiler rpath: runpath
>   C compiler: gcc
>  C compiler absolute: /usr/bin/gcc
>   C compiler family name: GNU
>   C compiler version: 11.3.1
> C++ compiler: g++
>C++ compiler absolute: /usr/bin/g++
>Fort compiler: gfortran
>Fort compiler abs: /usr/bin/gfortran
>  Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::)
>Fort 08 assumed shape: yes
>   Fort optional args: yes
>   Fort INTERFACE: yes
> Fort ISO_FORTRAN_ENV: yes
>Fort STORAGE_SIZE: yes
>   Fort BIND(C) (all): yes
>   Fort ISO_C_BINDING: yes
>  Fort SUBROUTINE BIND(C): yes
>Fort TYPE,BIND(C): yes
>  

Re: [OMPI users] psec warning when launching with srun

2023-05-20 Thread christof.koehler--- via users
Hello Z. Matthias Krawutschke,

On Fri, May 19, 2023 at 09:08:08PM +0200, Zhéxué M. Krawutschke wrote:
> Hello Christoph,
> what exactly is your problem with OpenMPI and Slurm?
> Do you compile the products yourself? Which LINUX distribution and version 
> are you using?
> 
> If you compile the software yourself, could you please tell me what the 
> "configure" command looks like and which MUNGE version is in use? From the 
> distribution or compiled by yourself?
> 
> I would be very happy to take on this topic and help you. You can also reach 
> me at +49 176 67270992.
> Best regards from Berlin

please refer to (especially the end) of my first mail in this thread
which is available here
https://www.mail-archive.com/users@lists.open-mpi.org/msg35141.html

I believe this contains the relevant information you are requesting. The
second mail which you are replying to was just additional information.
My apologies if this led to confusion.

Please let me know if any relevant information is missing from my first
email. At the bottom of this email I include the ompi_info output as
further addendum.

To summarize: I would like to understand where the munge warning 
and PMIx error described in the first email (and the github link
included) come from. The explanation in the github issue 
does not appear to be correct as all munge libraries are
available everywhere. To me, it appears at the moment that OpenMPIs 
configure decides erroneously to build and use the internal pmix 
instead of using the (presumably) newer externally available PMIx, 
leading to launcher problems with srun.


Best Regards

Christof

 Package: Open MPI root@admin.service Distribution
Open MPI: 4.1.5
  Open MPI repo revision: v4.1.5
   Open MPI release date: Feb 23, 2023
Open RTE: 4.1.5
  Open RTE repo revision: v4.1.5
   Open RTE release date: Feb 23, 2023
OPAL: 4.1.5
  OPAL repo revision: v4.1.5
   OPAL release date: Feb 23, 2023
 MPI API: 3.1.0
Ident string: 4.1.5
  Prefix: /cluster/mpi/openmpi/4.1.5/gcc-11.3.1
 Configured architecture: x86_64-pc-linux-gnu
  Configure host: admin.service
   Configured by: root
   Configured on: Wed May 17 18:45:42 UTC 2023
  Configure host: admin.service
  Configure command line: '--enable-mpi1-compatibility'
'--enable-orterun-prefix-by-default'
'--with-ofi=/cluster/libraries/libfabric/1.18.0/' '--with-slurm'
'--with-pmix' '--with-pmix-libdir=/usr/lib64' '--with-pmi'
'--with-pmi-libdir=/usr/lib64'
'--prefix=/cluster/mpi/openmpi/4.1.5/gcc-11.3.1'
Built by: root
Built on: Wed May 17 06:48:36 PM UTC 2023
  Built host: admin.service
  C bindings: yes
C++ bindings: no
 Fort mpif.h: yes (all)
Fort use mpi: yes (full: ignore TKR)
   Fort use mpi size: deprecated-ompi-info-value
Fort use mpi_f08: yes
 Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
limitations in the gfortran compiler and/or Open MPI, does not support
the following: array subsections, direct passthru (where possible) to
underlying Open MPI's C functionality
  Fort mpi_f08 subarrays: no
   Java bindings: no
  Wrapper compiler rpath: runpath
  C compiler: gcc
 C compiler absolute: /usr/bin/gcc
  C compiler family name: GNU
  C compiler version: 11.3.1
C++ compiler: g++
   C++ compiler absolute: /usr/bin/g++
   Fort compiler: gfortran
   Fort compiler abs: /usr/bin/gfortran
 Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::)
   Fort 08 assumed shape: yes
  Fort optional args: yes
  Fort INTERFACE: yes
Fort ISO_FORTRAN_ENV: yes
   Fort STORAGE_SIZE: yes
  Fort BIND(C) (all): yes
  Fort ISO_C_BINDING: yes
 Fort SUBROUTINE BIND(C): yes
   Fort TYPE,BIND(C): yes
 Fort T,BIND(C,name="a"): yes
Fort PRIVATE: yes
  Fort PROTECTED: yes
   Fort ABSTRACT: yes
   Fort ASYNCHRONOUS: yes
  Fort PROCEDURE: yes
 Fort USE...ONLY: yes
   Fort C_FUNLOC: yes
 Fort f08 using wrappers: yes
 Fort MPI_SIZEOF: yes
 C profiling: yes
   C++ profiling: no
   Fort mpif.h profiling: yes
  Fort use mpi profiling: yes
   Fort use mpi_f08 prof: yes
  C++ exceptions: no
  Thread support: posix (MPI_THREAD_MULTIPLE: yes, OPAL support:
yes, OMPI progress: no, ORTE progress: yes, Event lib: yes)
   Sparse Groups: no
  Internal debug support: no
  MPI interface warnings: yes
 MPI parameter check: runtime
Memory profiling support: no
Memory debugging support: no
  dl support: yes
   Heterogeneous support: no
 mpirun default --prefix: yes
   MPI_WTIME support: native
 Symbol vis. support: yes
   Host topology support: yes
IPv6 support: no
  MPI1 compatibility: yes
  

Re: [OMPI users] psec warning when launching with srun

2023-05-19 Thread Zhéxué M. Krawutschke via users
Hello Christoph,
what exactly is your problem with OpenMPI and Slurm?
Do you compile the products yourself? Which LINUX distribution and version are 
you using?

If you compile the software yourself, could you please tell me what the 
"configure" command looks like and which MUNGE version is in use? From the 
distribution or compiled by yourself?

I would be very happy to take on this topic and help you. You can also reach me 
at +49 176 67270992.
Best regards from Berlin

Z. Matthias Krawutschke

> On Donnerstag, Mai 18, 2023 at 5:47 PM, christof.koehler--- via users 
> mailto:users@lists.open-mpi.org)> wrote:
> Hello again,
>
> I should add that the openmpi configure decides to use the internal pmix
>
> configure: WARNING: discovered external PMIx version is less than internal 
> version 3.x
> configure: WARNING: using internal PMIx
> ...
> ...
> checking if user requested PMI support... yes
> checking for pmi.h in /usr/include... not found
> checking for pmi.h in /usr/include/slurm... found
> checking pmi.h usability... yes
> checking pmi.h presence... yes
> checking for pmi.h... yes
> checking for libpmi in /usr/lib64... found
> checking for PMI_Init in -lpmi... yes
> checking for pmi2.h in /usr/include... not found
> checking for pmi2.h in /usr/include/slurm... found
> checking pmi2.h usability... yes
> checking pmi2.h presence... yes
> checking for pmi2.h... yes
> checking for libpmi2 in /usr/lib64... found
> checking for PMI2_Init in -lpmi2... yes
> checking for pmix.h in ... not found
> checking for pmix.h in /include... not found
> checking can PMI support be built... yes
> checking if user requested internal PMIx support(yes)... no
> checking for pmix.h in /usr... not found
> checking for pmix.h in /usr/include... found
> checking libpmix.* in /usr/lib64... found
> checking PMIx version... version file found
> checking version 4x... found
> checking PMIx version to be used... internal
>
> I am not sure how it decides that, the external one is already a quite
> new version.
>
> # srun --mpi=list
> MPI plugin types are...
> pmix
> cray_shasta
> none
> pmi2
> specific pmix plugin versions available: pmix_v4
>
>
> Best Regards
>
> Christof
>
> --
> Dr. rer. nat. Christof Köhler email: c.koeh...@uni-bremen.de
> Universitaet Bremen/FB1/BCCMS phone: +49-(0)421-218-62334
> Am Fallturm 1/ TAB/ Raum 3.06 fax: +49-(0)421-218-62770
> 28359 Bremen
>


Re: [OMPI users] psec warning when launching with srun

2023-05-18 Thread christof.koehler--- via users
Hello again,

I should add that the openmpi configure decides to use the internal pmix

configure: WARNING: discovered external PMIx version is less than internal 
version 3.x
configure: WARNING: using internal PMIx
...
...
checking if user requested PMI support... yes
checking for pmi.h in /usr/include... not found
checking for pmi.h in /usr/include/slurm... found
checking pmi.h usability... yes
checking pmi.h presence... yes
checking for pmi.h... yes
checking for libpmi in /usr/lib64... found
checking for PMI_Init in -lpmi... yes
checking for pmi2.h in /usr/include... not found
checking for pmi2.h in /usr/include/slurm... found
checking pmi2.h usability... yes
checking pmi2.h presence... yes
checking for pmi2.h... yes
checking for libpmi2 in /usr/lib64... found
checking for PMI2_Init in -lpmi2... yes
checking for pmix.h in ... not found
checking for pmix.h in /include... not found
checking can PMI support be built... yes
checking if user requested internal PMIx support(yes)... no
checking for pmix.h in /usr... not found
checking for pmix.h in /usr/include... found
checking libpmix.* in /usr/lib64... found
checking PMIx version... version file found
checking version 4x... found
checking PMIx version to be used... internal

I am not sure how it decides that, the external one is already a quite
new version.

# srun --mpi=list
MPI plugin types are...
pmix
cray_shasta
none
pmi2
specific pmix plugin versions available: pmix_v4


Best Regards

Christof

-- 
Dr. rer. nat. Christof Köhler   email: c.koeh...@uni-bremen.de
Universitaet Bremen/FB1/BCCMS   phone:  +49-(0)421-218-62334
Am Fallturm 1/ TAB/ Raum 3.06   fax: +49-(0)421-218-62770
28359 Bremen  


[OMPI users] psec warning when launching with srun

2023-05-18 Thread christof.koehler--- via users
Hello everybody,

we are seeing the sypmptoms described in
https://github.com/open-mpi/ompi/issues/11557

However, according to the systems package manager (dnf) all munge
related packages on the build node and the execution node are identical,
see details at the bottom. So, the explanation given by Ralph Castain in 
the git issue, which I read to refer to a missing munge library, does 
not appear to explain the warning. I would like to note that a 
PMIX_MCA_psec=native is sufficient to make the warning go away, it is 
not necessary to disable munge completely. Launching with 
srun --mpi=pmi2 instead works fine, anyway.

Related to that: pmix has a configure switch to disable munge support.
Would it be possible and/or adviseable to disable munge in the pmix
build ?

I have to note that we are seeing also 
PMIX ERROR: ERROR in file gds_ds12_lock_pthread.c at line 168
for each rank launched. This might be a separate issue or related.
Setting PMIX_MCA_gds=^ds12 makes this go away.

With or without setting PMIX_MCA_psec and PMIX_MCA_gds the job launches
and is executed. Still, I would like to understand this better in case
we have a broken slurm or mpi setup which would need to be corrected.

Best Regards

Christof

System Details:

Rocky Linux 9.1, slurm 23.02.2, open pmix 4.2.4rc1, openmpi 4.1.5.

slurm and pmix were compiled in the same step to rpms in a mock 
build environment using the spec files included with them. I can 
provide the build logs of the rpms. Both are not from distribution
repositories.

openmpi was configured with 
./configure --enable-mpi1-compatibility
--enable-orterun-prefix-by-default
--with-ofi=/cluster/libraries/libfabric/1.18.0/  --with-slurm
--with-pmix --with-pmix-libdir=/usr/lib64 --with-pmi
--with-pmi-libdir=/usr/lib64

Installed munge packages from distribution according to dnf

on build node
 munge.x86_64   0.5.13-13.el9
 munge-devel.x86_64 0.5.13-13.el9
 munge-libs.x86_64  0.5.13-13.el9

on execution node
 munge.x86_640.5.13-13.el9
 munge-devel.x86_64  0.5.13-13.el9
 munge-libs.x86_64   0.5.13-13.el9

-- 
Dr. rer. nat. Christof Köhler   email: c.koeh...@uni-bremen.de
Universitaet Bremen/FB1/BCCMS   phone:  +49-(0)421-218-62334
Am Fallturm 1/ TAB/ Raum 3.06   fax: +49-(0)421-218-62770
28359 Bremen