On Feb 12, 2012, at 4:52 AM, Paul Hargrove wrote:
> I just tried tonight's nightly tarball for the 1.5 branch (1.5.5rc2r25906).
> I found the following issues, which I wad previously reported against
> 1.5.5rc1, for which I did NOT find a corresponding ticket in "report/15". My
> apologies is
Also, double check that you have an optimized build, not a debugging build.
SVN and HG checkouts default to debugging builds, which add in lots of latency.
On Feb 13, 2012, at 10:22 AM, Ralph Castain wrote:
> Few thoughts
>
> 1. Bind to socket is broken in 1.5.4 - fixed in next release
>
>
After much off-list back-n-forth with Paul Hargrove (many thanks, Paul!), we
have 1.3.2rc2 up in the usual place:
http://www.open-mpi.org/software/hwloc/v1.3/
Due to the volume of off-list emails, I'm kinda expecting this rc to be good /
final. However, please do at least some cursory
Few thoughts
1. Bind to socket is broken in 1.5.4 - fixed in next release
2. Add --report-bindings to cmd line and see where it thinks the procs are bound
3. Sounds lime memory may not be local - might be worth checking mem binding.
Sent from my iPad
On Feb 13, 2012, at 7:07 AM, Matthias
Hi Sylvain,
thanks for the quick response!
Here some results with enabled process binding. I hope I used the parameters
correctly...
bind two ranks to one socket:
$ mpirun -np 2 --bind-to-core ./all2all
$ mpirun -np 2 -mca mpi_paffinity_alone 1 ./all2all
bind two ranks to two different
Hi Matthias,
You might want to play with process binding to see if your problem is
related to bad memory affinity.
Try to launch pingpong on two CPUs of the same socket, then on different
sockets (i.e. bind each process to a core, and try different
configurations).
Sylvain
De :
Hello all,
on our new AMD cluster (AMD Opteron 6274, 2,2GHz) we get very bad latencies
(~1.5us) when performing 0-byte p2p communication on one single node using the
Open MPI sm BTL. When using Platform MPI we get ~0.5us latencies which is
pretty good. The bandwidth results are similar for