Thanks Geoffroy. I don't think I'm worried about this for v2.1.4, and the UCX community hasn't responded. So I'm going to release 2.1.4 as-is.
> On Aug 9, 2018, at 3:33 PM, Vallee, Geoffroy R. <valle...@ornl.gov> wrote: > > Hi, > > I tested on Summitdev here at ORNL and here are my comments (but I only have > a limited set of data for summitdev so my feedback is somewhat limited): > - netpipe/mpi is showing a slightly lower bandwidth than the 3.x series (I do > not believe it is a problem). > - I am facing a problem with UCX, it is unclear to me that it is relevant > since I am using UCX master and I do not know whether it is expected to work > with OMPI v2.1.x. Note that I am using the same tool for testing all other > releases of Open MPI and I never had that problem before, having in mind that > I only tested the 3.x series so far. > > make: Entering directory > `/autofs/nccs-svm1_home1/gvh/.ompi-release-tester/scratch/summitdev/2.1.4rc1/scratch/UCX/ompi_build/ompi/mca/pml/ucx' > /bin/sh ../../../../libtool --tag=CC --mode=link gcc -std=gnu99 -O3 > -DNDEBUG -finline-functions -fno-strict-aliasing -pthread -module > -avoid-version -o mca_pml_ucx.la -rpath > /ccs/home/gvh/.ompi-release-tester/scratch/summitdev/2.1.4rc1/scratch/UCX/ompi_install/lib/openmpi > pml_ucx.lo pml_ucx_request.lo pml_ucx_datatype.lo pml_ucx_component.lo -lucp > -lrt -lm -lutil > libtool: link: gcc -std=gnu99 -shared -fPIC -DPIC .libs/pml_ucx.o > .libs/pml_ucx_request.o .libs/pml_ucx_datatype.o .libs/pml_ucx_component.o > -lucp -lrt -lm -lutil -O3 -pthread -pthread -Wl,-soname -Wl,mca_pml_ucx.so > -o .libs/mca_pml_ucx.so > /usr/bin/ld: cannot find -lucp > collect2: error: ld returned 1 exit status > make: *** [mca_pml_ucx.la] Error 1 > make: Leaving directory > `/autofs/nccs-svm1_home1/gvh/.ompi-release-tester/scratch/summitdev/2.1.4rc1/scratch/UCX/ompi_build/ompi/mca/pml/ucx' > make: *** [all-recursive] Error 1 > make: Leaving directory > `/autofs/nccs-svm1_home1/gvh/.ompi-release-tester/scratch/summitdev/2.1.4rc1/scratch/UCX/ompi_build/ompi' > make: *** [all-recursive] Error 1 > > My 2 cents, > >> On Aug 6, 2018, at 5:04 PM, Jeff Squyres (jsquyres) via devel >> <firstname.lastname@example.org> wrote: >> >> Open MPI v2.1.4rc1 has been pushed. It is likely going to be the last in >> the v2.1.x series (since v4.0.0 is now visible on the horizon). It is just >> a bunch of bug fixes that have accumulated since v2.1.3; nothing huge. >> We'll encourage users who are still using the v2.1.x series to upgrade to >> this release; it should be a non-event for anyone who has already upgraded >> to the v3.0.x or v3.1.x series. >> >> https://www.open-mpi.org/software/ompi/v2.1/ >> >> If no serious-enough issues are found, we plan to release 2.1.4 this Friday, >> August 10, 2018. >> >> Please test! >> >> Bug fixes/minor improvements: >> - Disable the POWER 7/BE block in configure. Note that POWER 7/BE is >> still not a supported platform, but it is no longer automatically >> disabled. See >> https://github.com/open-mpi/ompi/issues/4349#issuecomment-374970982 >> for more information. >> - Fix bug with request-based one-sided MPI operations when using the >> "rdma" component. >> - Fix issue with large data structure in the TCP BTL causing problems >> in some environments. Thanks to @lgarithm for reporting the issue. >> - Minor Cygwin build fixes. >> - Minor fixes for the openib BTL: >> - Support for the QLogic RoCE HCA >> - Support for the Boradcom Cumulus RoCE HCA >> - Enable support for HDR link speeds >> - Fix MPI_FINALIZED hang if invoked from an attribute destructor >> during the MPI_COMM_SELF destruction in MPI_FINALIZE. Thanks to >> @AndrewGaspar for reporting the issue. >> - Java fixes: >> - Modernize Java framework detection, especially on OS X/MacOS. >> Thanks to Bryce Glover for reporting and submitting the fixes. >> - Prefer "javac -h" to "javah" to support newer Java frameworks. >> - Fortran fixes: >> - Use conformant dummy parameter names for Fortran bindings. Thanks >> to Themos Tsikas for reporting and submitting the fixes. >> - Build the MPI_SIZEOF() interfaces in the "TKR"-style "mpi" module >> whenever possible. Thanks to Themos Tsikas for reporting the >> issue. >> - Fix array of argv handling for the Fortran bindings of >> MPI_COMM_SPAWN_MULTIPLE (and its associated man page). >> - Make NAG Fortran compiler support more robust in configure. >> - Disable the "pt2pt" one-sided MPI component when MPI_THREAD_MULTIPLE >> is used. This component is simply not safe in MPI_THREAD_MULTIPLE >> scenarios, and will not be fixed in the v2.1.x series. >> - Make the "external" hwloc component fail gracefully if it is tries >> to use an hwloc v2.x.y installation. hwloc v2.x.y will not be >> supported in the Open MPI v2.1.x series. >> - Fix "vader" shared memory support for messages larger than 2GB. >> Thanks to Heiko Bauke for the bug report. >> - Configure fixes for external PMI directory detection. Thanks to >> Davide Vanzo for the report. >> >> >> -- >> Jeff Squyres >> jsquy...@cisco.com >> >> _______________________________________________ >> devel mailing list >> email@example.com >> https://lists.open-mpi.org/mailman/listinfo/devel > -- Jeff Squyres jsquy...@cisco.com _______________________________________________ devel mailing list firstname.lastname@example.org https://lists.open-mpi.org/mailman/listinfo/devel