Hi, I tested on Summitdev here at ORNL and here are my comments (but I only have a limited set of data for summitdev so my feedback is somewhat limited): - netpipe/mpi is showing a slightly lower bandwidth than the 3.x series (I do not believe it is a problem). - I am facing a problem with UCX, it is unclear to me that it is relevant since I am using UCX master and I do not know whether it is expected to work with OMPI v2.1.x. Note that I am using the same tool for testing all other releases of Open MPI and I never had that problem before, having in mind that I only tested the 3.x series so far.
make[2]: Entering directory `/autofs/nccs-svm1_home1/gvh/.ompi-release-tester/scratch/summitdev/2.1.4rc1/scratch/UCX/ompi_build/ompi/mca/pml/ucx' /bin/sh ../../../../libtool --tag=CC --mode=link gcc -std=gnu99 -O3 -DNDEBUG -finline-functions -fno-strict-aliasing -pthread -module -avoid-version -o mca_pml_ucx.la -rpath /ccs/home/gvh/.ompi-release-tester/scratch/summitdev/2.1.4rc1/scratch/UCX/ompi_install/lib/openmpi pml_ucx.lo pml_ucx_request.lo pml_ucx_datatype.lo pml_ucx_component.lo -lucp -lrt -lm -lutil libtool: link: gcc -std=gnu99 -shared -fPIC -DPIC .libs/pml_ucx.o .libs/pml_ucx_request.o .libs/pml_ucx_datatype.o .libs/pml_ucx_component.o -lucp -lrt -lm -lutil -O3 -pthread -pthread -Wl,-soname -Wl,mca_pml_ucx.so -o .libs/mca_pml_ucx.so /usr/bin/ld: cannot find -lucp collect2: error: ld returned 1 exit status make[2]: *** [mca_pml_ucx.la] Error 1 make[2]: Leaving directory `/autofs/nccs-svm1_home1/gvh/.ompi-release-tester/scratch/summitdev/2.1.4rc1/scratch/UCX/ompi_build/ompi/mca/pml/ucx' make[1]: *** [all-recursive] Error 1 make[1]: Leaving directory `/autofs/nccs-svm1_home1/gvh/.ompi-release-tester/scratch/summitdev/2.1.4rc1/scratch/UCX/ompi_build/ompi' make: *** [all-recursive] Error 1 My 2 cents, > On Aug 6, 2018, at 5:04 PM, Jeff Squyres (jsquyres) via devel > <devel@lists.open-mpi.org> wrote: > > Open MPI v2.1.4rc1 has been pushed. It is likely going to be the last in the > v2.1.x series (since v4.0.0 is now visible on the horizon). It is just a > bunch of bug fixes that have accumulated since v2.1.3; nothing huge. We'll > encourage users who are still using the v2.1.x series to upgrade to this > release; it should be a non-event for anyone who has already upgraded to the > v3.0.x or v3.1.x series. > > https://www.open-mpi.org/software/ompi/v2.1/ > > If no serious-enough issues are found, we plan to release 2.1.4 this Friday, > August 10, 2018. > > Please test! > > Bug fixes/minor improvements: > - Disable the POWER 7/BE block in configure. Note that POWER 7/BE is > still not a supported platform, but it is no longer automatically > disabled. See > https://github.com/open-mpi/ompi/issues/4349#issuecomment-374970982 > for more information. > - Fix bug with request-based one-sided MPI operations when using the > "rdma" component. > - Fix issue with large data structure in the TCP BTL causing problems > in some environments. Thanks to @lgarithm for reporting the issue. > - Minor Cygwin build fixes. > - Minor fixes for the openib BTL: > - Support for the QLogic RoCE HCA > - Support for the Boradcom Cumulus RoCE HCA > - Enable support for HDR link speeds > - Fix MPI_FINALIZED hang if invoked from an attribute destructor > during the MPI_COMM_SELF destruction in MPI_FINALIZE. Thanks to > @AndrewGaspar for reporting the issue. > - Java fixes: > - Modernize Java framework detection, especially on OS X/MacOS. > Thanks to Bryce Glover for reporting and submitting the fixes. > - Prefer "javac -h" to "javah" to support newer Java frameworks. > - Fortran fixes: > - Use conformant dummy parameter names for Fortran bindings. Thanks > to Themos Tsikas for reporting and submitting the fixes. > - Build the MPI_SIZEOF() interfaces in the "TKR"-style "mpi" module > whenever possible. Thanks to Themos Tsikas for reporting the > issue. > - Fix array of argv handling for the Fortran bindings of > MPI_COMM_SPAWN_MULTIPLE (and its associated man page). > - Make NAG Fortran compiler support more robust in configure. > - Disable the "pt2pt" one-sided MPI component when MPI_THREAD_MULTIPLE > is used. This component is simply not safe in MPI_THREAD_MULTIPLE > scenarios, and will not be fixed in the v2.1.x series. > - Make the "external" hwloc component fail gracefully if it is tries > to use an hwloc v2.x.y installation. hwloc v2.x.y will not be > supported in the Open MPI v2.1.x series. > - Fix "vader" shared memory support for messages larger than 2GB. > Thanks to Heiko Bauke for the bug report. > - Configure fixes for external PMI directory detection. Thanks to > Davide Vanzo for the report. > > > -- > Jeff Squyres > jsquy...@cisco.com > > _______________________________________________ > devel mailing list > devel@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/devel _______________________________________________ devel mailing list devel@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/devel