Re: [OMPI devel] 1.5.5rc2r25906 test results

2012-02-13 Thread Jeff Squyres
On Feb 12, 2012, at 4:52 AM, Paul Hargrove wrote: > I just tried tonight's nightly tarball for the 1.5 branch (1.5.5rc2r25906). > I found the following issues, which I wad previously reported against > 1.5.5rc1, for which I did NOT find a corresponding ticket in "report/15". My > apologies is

Re: [OMPI devel] poor btl sm latency

2012-02-13 Thread Jeff Squyres
Also, double check that you have an optimized build, not a debugging build. SVN and HG checkouts default to debugging builds, which add in lots of latency. On Feb 13, 2012, at 10:22 AM, Ralph Castain wrote: > Few thoughts > > 1. Bind to socket is broken in 1.5.4 - fixed in next release > >

[hwloc-devel] hwloc 1.3.2rc2 released

2012-02-13 Thread Jeff Squyres
After much off-list back-n-forth with Paul Hargrove (many thanks, Paul!), we have 1.3.2rc2 up in the usual place: http://www.open-mpi.org/software/hwloc/v1.3/ Due to the volume of off-list emails, I'm kinda expecting this rc to be good / final. However, please do at least some cursory

Re: [OMPI devel] poor btl sm latency

2012-02-13 Thread Ralph Castain
Few thoughts 1. Bind to socket is broken in 1.5.4 - fixed in next release 2. Add --report-bindings to cmd line and see where it thinks the procs are bound 3. Sounds lime memory may not be local - might be worth checking mem binding. Sent from my iPad On Feb 13, 2012, at 7:07 AM, Matthias

Re: [OMPI devel] poor btl sm latency

2012-02-13 Thread Matthias Jurenz
Hi Sylvain, thanks for the quick response! Here some results with enabled process binding. I hope I used the parameters correctly... bind two ranks to one socket: $ mpirun -np 2 --bind-to-core ./all2all $ mpirun -np 2 -mca mpi_paffinity_alone 1 ./all2all bind two ranks to two different

Re: [OMPI devel] poor btl sm latency

2012-02-13 Thread sylvain . jeaugey
Hi Matthias, You might want to play with process binding to see if your problem is related to bad memory affinity. Try to launch pingpong on two CPUs of the same socket, then on different sockets (i.e. bind each process to a core, and try different configurations). Sylvain De :

[OMPI devel] poor btl sm latency

2012-02-13 Thread Matthias Jurenz
Hello all, on our new AMD cluster (AMD Opteron 6274, 2,2GHz) we get very bad latencies (~1.5us) when performing 0-byte p2p communication on one single node using the Open MPI sm BTL. When using Platform MPI we get ~0.5us latencies which is pretty good. The bandwidth results are similar for