1.6.1rc1 is a bust because of a compile error.  :(

It wasn't caught on the build machine because it's a bug in the openib BTL, and 
the build machine doesn't have OpenFabrics support.

1.6.1rc2 will be posted later today.



On Jul 27, 2012, at 10:20 PM, Jeff Squyres wrote:

> Finally!  It's in the usual place:
> 
>    http://www.open-mpi.org/software/ompi/v1.6/
> 
> Please test, especially with low-registered-memory-available scenarios with 
> Mellanox OpenFabrics devices.
> 
> Here's a list of changes since 1.6:
> 
> - A bunch of changes to eliminate hangs on OpenFabrics-based networks.
>  Users with Mellanox hardware are ***STRONGLY ENCOURAGED*** to check
>  their registered memory kernel module settings to ensure that the OS
>  will allow registering more than 8GB of memory.  See this FAQ item
>  for details:
> 
>  http://www.open-mpi.org/faq/?category=openfabrics#ib-low-reg-mem
> 
>  - Fall back to send/receive semantics if registered memory is
>    unavilable for RDMA.
>  - Fix two fragment leaks when registered memory is exhausted.
>  - Hueristically determine how much registered memory is available
>    and warn if it's significantly less than all of RAM.
>  - Artifically limit the amount of registered memory each MPI process
>    can use to about 1/Nth to total registered memory available.
>  - Improve error messages when events occur that are likely due to
>    unexpected registered memory exhaustion.
>  - Remove the last openib default per-peer receive queue
>    specification (and make it an SRQ).
> 
> - Switch the MPI_ALLTOALLV default algorithm to a pairwise exchange.
> - Increase the openib BTL default CQ length to handle more types of
>  OpenFabrics devices.
> - Lots of VampirTrace fixes; upgrade to v5.13.0.4.
> - Map MPI_2INTEGER to underlying MPI_INTEGERs, not MPI_INTs.
> - Ensure that the OMPI version number is toleant of handling spaces.
>  Thanks to dragonboy for identifying the issue.
> - Fixed IN parameter marking on Fortran "mpi" module
>  MPI_COMM_TEST_INTER interface.
> - Various MXM improvements.
> - Make the output of "mpirun --report-bindings" much more friendly /
>  human-readable.
> - Properly handle MPI_COMPLEX8|16|32.
> - More fixes for mpirun's processor affinity options (--bind-to-core
>  and friends).
> - Use aligned memory for OpenFabrics registered memory.
> - Multiple fixes for parameter checking in MPI_ALLGATHERV,
>  MPI_REDUCE_SCATTER, MPI_SCATTERV, and MPI_GATHERV.  Thanks to the
>  mpi4py community (Bennet Fauber, Lisandro Dalcin, Jonathan Dursi).
> - Fixed file positioning overflows in MPI_FILE_GET_POSITION,
>  MPI_FILE_GET_POSITION_SHARED, FILE_GET_SIZE, FILE_GET_VIEW.
> - Removed the broken --cpu-set mpirun option.
> - Fix cleanup of MPI errorcodes.  Thanks to Alexey Bayduraev for the
>  patch.
> - Fix default hostfile location.  Thanks to Götz Waschk for noticing
>  the issue.
> - Improve several error messages.
> 
> 
> -- 
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to: 
> http://www.cisco.com/web/about/doing_business/legal/cri/
> 


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/


Reply via email to