The Second release candidate for the Open MPI v4.0.0 release has been built and will be available tonight at:
https://www.open-mpi.org/software/ompi/v4.0/
https://www.open-mpi.org/software/ompi/v4.0/
Major differences from v4.0.0rc1 include:
------------------------
All Major differences from v3.1x include:
- Removed support for SCIF.
- Enable use of CUDA allocated buffers for OMPIO.
- Fix a problem with ORTE not reporting error messages if an application
terminated normally but exited with non-zero error code.
Thanks to Emre Brookes for reporting.
- Enable use of CUDA allocated buffers for OMPIO.
- Fix a problem with ORTE not reporting error messages if an application
terminated normally but exited with non-zero error code.
Thanks to Emre Brookes for reporting.
------------------------
- OSHMEM updated to the OpenSHMEM 1.4 API.
- Do not build Open SHMEM layer when there are no SPMLs available.
Currently, this means the Open SHMEM layer will only build if
a MXM or UCX library is found.
- A UCX BTL was added for enhanced MPI RMA support using UCX
- With this release, OpenIB BTL now only supports iWarp and RoCE by default.
- Updated internal HWLOC to 2.0.1
- Updated internal PMIx to 3.0.1
- Change the priority for selecting external verses internal HWLOC
and PMIx packages to build. Starting with this release, configure
by default selects available external HWLOC and PMIx packages over
the internal ones.
- Updated internal ROMIO to 3.2.1.
- Removed support for the MXM MTL.
- Removed support for SCIF.
- Improved CUDA support when using UCX.
- Enable use of CUDA allocated buffers for OMPIO.
- Improved support for two phase MPI I/O operations when using OMPIO.
- Added support for Software-based Performance Counters, see
https://github.com/davideberius/ompi/wiki/How-to-Use-Software-Based-Performance-Counters-(SPCs)-in-Open-MPI
- Various improvements to MPI RMA performance when using RDMA
capable interconnects.
- Update memkind component to use the memkind 1.6 public API.
- Fix problems with use of newer map-by mpirun options. Thanks to
Tony Reina for reporting.
- Fix rank-by algorithms to properly rank by object and span
- Allow for running as root of two environment variables are set.
Requested by Axel Huebl.
- Fix a problem with building the Java bindings when using Java 10.
Thanks to Bryce Glover for reporting.
- Fix a problem with ORTE not reporting error messages if an application
terminated normally but exited with non-zero error code. Thanks to
Emre Brookes for reporting.
- Do not build Open SHMEM layer when there are no SPMLs available.
Currently, this means the Open SHMEM layer will only build if
a MXM or UCX library is found.
- A UCX BTL was added for enhanced MPI RMA support using UCX
- With this release, OpenIB BTL now only supports iWarp and RoCE by default.
- Updated internal HWLOC to 2.0.1
- Updated internal PMIx to 3.0.1
- Change the priority for selecting external verses internal HWLOC
and PMIx packages to build. Starting with this release, configure
by default selects available external HWLOC and PMIx packages over
the internal ones.
- Updated internal ROMIO to 3.2.1.
- Removed support for the MXM MTL.
- Removed support for SCIF.
- Improved CUDA support when using UCX.
- Enable use of CUDA allocated buffers for OMPIO.
- Improved support for two phase MPI I/O operations when using OMPIO.
- Added support for Software-based Performance Counters, see
https://github.com/davideberius/ompi/wiki/How-to-Use-Software-Based-Performance-Counters-(SPCs)-in-Open-MPI
- Various improvements to MPI RMA performance when using RDMA
capable interconnects.
- Update memkind component to use the memkind 1.6 public API.
- Fix problems with use of newer map-by mpirun options. Thanks to
Tony Reina for reporting.
- Fix rank-by algorithms to properly rank by object and span
- Allow for running as root of two environment variables are set.
Requested by Axel Huebl.
- Fix a problem with building the Java bindings when using Java 10.
Thanks to Bryce Glover for reporting.
- Fix a problem with ORTE not reporting error messages if an application
terminated normally but exited with non-zero error code. Thanks to
Emre Brookes for reporting.
_______________________________________________ ompi-packagers mailing list ompi-packagers@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/ompi-packagers