btw, I have iomkl/2017.01 with OpenMPI/2.0.1-iccifort-2017.1.132-GCC-5.4.0-2.26 and I had to apply this patch to be able to compile it with slurm support https://www.mail-archive.com/[email protected]/msg30048.html
I suppose this is fixed in 2.0.2 2017-02-01 15:01 GMT+01:00 Kenneth Hoste <[email protected]>: > Hi Joachim, > > On 01/02/2017 14:59, Joachim Hein wrote: > > HI Kenneth et al. > > Can’t make if for todays telcon, but got the below this morning. You > might want to considerer for foss/2017a, which is on the agenda. We’ll > live with foss/2017a until August or so. > > > Thanks for the heads-up, I was already aware of it. > > The PR for foss/2017a has already been updated accordingly, see > https://github.com/hpcugent/easybuild-easyconfigs/pull/3968 . > > I also bumped FFTW to 3.3.6 which was released two weeks ago. > > > regards, > > Kenneth > > > Joachim > > Begin forwarded message: > > *From: *"Jeff Squyres (jsquyres)" <[email protected]> > *Subject: **[Open MPI Announce] Open MPI v2.0.2 released* > *Date: *1 February 2017 at 01:40:35 GMT+1 > *To: *Open MPI Announcements <[email protected]> > *Reply-To: *<[email protected]> > > The Open MPI Team, representing a consortium of research, academic, and > industry partners, is pleased to announce the release of Open MPI version > 2.0.2. > > v2.0.2 is a bug fix release that includes a variety of bug fixes and some > performance fixes. All users are encouraged to upgrade to v2.0.2 when > possible. > > Version 2.0.2 can be downloaded from the main Open MPI web site: > > https://www.open-mpi.org/software/ompi/v2.0/ > > NEWS > > 2.0.2 -- 31 January 2017 > ------------------------- > > Bug fixes/minor improvements: > > - Fix a problem with MPI_FILE_WRITE_SHARED when using MPI_MODE_APPEND and > Open MPI's native MPI-IO implementation. Thanks to Nicolas Joly for > reporting. > - Fix a typo in the MPI_WIN_GET_NAME man page. Thanks to Nicolas Joly > for reporting. > - Fix a race condition with ORTE's session directory setup. Thanks to > @tbj900 for reporting this issue. > - Fix a deadlock issue arising from Open MPI's approach to catching calls > to > munmap. Thanks to Paul Hargrove for reporting and helping to analyze this > problem. > - Fix a problem with PPC atomics which caused make check to fail unless > builtin > atomics configure option was enabled. Thanks to Orion Poplawski for > reporting. > - Fix a problem with use of x86_64 cpuid instruction which led to > segmentation > faults when Open MPI was configured with -O3 optimization. Thanks to Mark > Santcroos for reporting this problem. > - Fix a problem when using built in atomics configure options on PPC > platforms > when building 32 bit applications. Thanks to Paul Hargrove for reporting. > - Fix a problem with building Open MPI against an external hwloc > installation. > Thanks to Orion Poplawski for reporting this issue. > - Remove use of DATE in the message queue version string reported to > debuggers to > insure bit-wise reproducibility of binaries. Thanks to Alastair McKinstry > for help in fixing this problem. > - Fix a problem with early exit of a MPI process without calling > MPI_FINALIZE > or MPI_ABORT that could lead to job hangs. Thanks to Christof Koehler for > reporting. > - Fix a problem with forwarding of SIGTERM signal from mpirun to MPI > processes > in a job. Thanks to Noel Rycroft for reporting this problem > - Plug some memory leaks in MPI_WIN_FREE discovered using Valgrind. Thanks > to Joseph Schuchart for reporting. > - Fix a problems MPI_NEIGHOR_ALLTOALL when using a communicator with an > empty topology > graph. Thanks to Daniel Ibanez for reporting. > - Fix a typo in a PMIx component help file. Thanks to @njoly for > reporting this. > - Fix a problem with Valgrind false positives when using Open MPI's > internal memchecker. > Thanks to Yvan Fournier for reporting. > - Fix a problem with MPI_FILE_DELETE returning MPI_SUCCESS when > deleting a non-existent file. Thanks to Wei-keng Liao for reporting. > - Fix a problem with MPI_IMPROBE that could lead to hangs in subsequent MPI > point to point or collective calls. Thanks to Chris Pattison for > reporting. > - Fix a problem when configure Open MPI for powerpc with --enable-mpi-cxx > enabled. Thanks to Alastair McKinstry for reporting. > - Fix a problem using MPI_IALLTOALL with MPI_IN_PLACE argument. Thanks to > Chris Ward for reporting. > - Fix a problem using MPI_RACCUMULATE with the Portals4 transport. Thanks > to > @PDeveze for reporting. > - Fix an issue with static linking and duplicate symbols arising from PMIx > Slurm components. Thanks to Limin Gu for reporting. > - Fix a problem when using MPI dynamic memory windows. Thanks to > Christoph Niethammer for reporting. > - Fix a problem with Open MPI's pkgconfig files. Thanks to Alastair > McKinstry > for reporting. > - Fix a problem with MPI_IREDUCE when the same buffer is supplied for the > send and recv buffer arguments. Thanks to Valentin Petrov for reporting. > - Fix a problem with atomic operations on PowerPC. Thanks to Paul > Hargrove for reporting. > > Known issues (to be addressed in v2.0.3): > > - See the list of fixes slated for v2.0.3 here: > https://github.com/open-mpi/ompi/milestone/23 > > -- > Jeff Squyres > [email protected] > > _______________________________________________ > announce mailing list > [email protected] > https://rfd.newmexicoconsortium.org/mailman/listinfo/announce > > > > -- Pablo Escobar López HPC systems engineer sciCORE, University of Basel SIB Swiss Institute of Bioinformatics http://scicore.unibas.ch

