Something is definitely wrong -- 1.4us is way too high for a 0 or 1 byte HRT 
ping pong.  What is this all2all benchmark, btw?  Is it measuring an 
MPI_ALLTOALL, or a pingpong?

FWIW, on an older Nehalem machine running NetPIPE/MPI, I'm getting about .27us 
latencies for short messages over sm and binding to socket.


On Feb 14, 2012, at 7:20 AM, Matthias Jurenz wrote:

> I've built Open MPI 1.5.5rc1 (tarball from Web) with CFLAGS=-O3. 
> Unfortunately, also without any effect.
> 
> Here some results with enabled binding reports:
> 
> $ mpirun *--bind-to-core* --report-bindings -np 2 ./all2all_ompi1.5.5
> [n043:61313] [[56788,0],0] odls:default:fork binding child [[56788,1],1] to 
> cpus 0002
> [n043:61313] [[56788,0],0] odls:default:fork binding child [[56788,1],0] to 
> cpus 0001
> latency: 1.415us
> 
> $ mpirun *-mca maffinity hwloc --bind-to-core* --report-bindings -np 2 
> ./all2all_ompi1.5.5
> [n043:61469] [[49736,0],0] odls:default:fork binding child [[49736,1],1] to 
> cpus 0002
> [n043:61469] [[49736,0],0] odls:default:fork binding child [[49736,1],0] to 
> cpus 0001
> latency: 1.4us
> 
> $ mpirun *-mca maffinity first_use --bind-to-core* --report-bindings -np 2 
> ./all2all_ompi1.5.5
> [n043:61508] [[49681,0],0] odls:default:fork binding child [[49681,1],1] to 
> cpus 0002
> [n043:61508] [[49681,0],0] odls:default:fork binding child [[49681,1],0] to 
> cpus 0001
> latency: 1.4us
> 
> 
> $ mpirun *--bind-to-socket* --report-bindings -np 2 ./all2all_ompi1.5.5
> [n043:61337] [[56780,0],0] odls:default:fork binding child [[56780,1],1] to 
> socket 0 cpus 0001
> [n043:61337] [[56780,0],0] odls:default:fork binding child [[56780,1],0] to 
> socket 0 cpus 0001
> latency: 4.0us
> 
> $ mpirun *-mca maffinity hwloc --bind-to-socket* --report-bindings -np 2 
> ./all2all_ompi1.5.5 
> [n043:61615] [[49914,0],0] odls:default:fork binding child [[49914,1],1] to 
> socket 0 cpus 0001
> [n043:61615] [[49914,0],0] odls:default:fork binding child [[49914,1],0] to 
> socket 0 cpus 0001
> latency: 4.0us
> 
> $ mpirun *-mca maffinity first_use --bind-to-socket* --report-bindings -np 2 
> ./all2all_ompi1.5.5 
> [n043:61639] [[49810,0],0] odls:default:fork binding child [[49810,1],1] to 
> socket 0 cpus 0001
> [n043:61639] [[49810,0],0] odls:default:fork binding child [[49810,1],0] to 
> socket 0 cpus 0001
> latency: 4.0us
> 
> 
> If socket-binding is enabled it seems that all ranks are bind to the very 
> first 
> core of one and the same socket. Is it intended? I expected that each rank 
> gets its own socket (i.e. 2 ranks -> 2 sockets)...
> 
> Matthias
> 
> On Monday 13 February 2012 22:36:50 Jeff Squyres wrote:
>> Also, double check that you have an optimized build, not a debugging build.
>> 
>> SVN and HG checkouts default to debugging builds, which add in lots of
>> latency.
>> 
>> On Feb 13, 2012, at 10:22 AM, Ralph Castain wrote:
>>> Few thoughts
>>> 
>>> 1. Bind to socket is broken in 1.5.4 - fixed in next release
>>> 
>>> 2. Add --report-bindings to cmd line and see where it thinks the procs
>>> are bound
>>> 
>>> 3. Sounds lime memory may not be local - might be worth checking mem
>>> binding.
>>> 
>>> Sent from my iPad
>>> 
>>> On Feb 13, 2012, at 7:07 AM, Matthias Jurenz <matthias.jurenz@tu-
> dresden.de> wrote:
>>>> Hi Sylvain,
>>>> 
>>>> thanks for the quick response!
>>>> 
>>>> Here some results with enabled process binding. I hope I used the
>>>> parameters correctly...
>>>> 
>>>> bind two ranks to one socket:
>>>> $ mpirun -np 2 --bind-to-core ./all2all
>>>> $ mpirun -np 2 -mca mpi_paffinity_alone 1 ./all2all
>>>> 
>>>> bind two ranks to two different sockets:
>>>> $ mpirun -np 2 --bind-to-socket ./all2all
>>>> 
>>>> All three runs resulted in similar bad latencies (~1.4us).
>>>> 
>>>> :-(
>>>> 
>>>> Matthias
>>>> 
>>>> On Monday 13 February 2012 12:43:22 sylvain.jeau...@bull.net wrote:
>>>>> Hi Matthias,
>>>>> 
>>>>> You might want to play with process binding to see if your problem is
>>>>> related to bad memory affinity.
>>>>> 
>>>>> Try to launch pingpong on two CPUs of the same socket, then on
>>>>> different sockets (i.e. bind each process to a core, and try different
>>>>> configurations).
>>>>> 
>>>>> Sylvain
>>>>> 
>>>>> 
>>>>> 
>>>>> De :    Matthias Jurenz <matthias.jur...@tu-dresden.de>
>>>>> A :     Open MPI Developers <de...@open-mpi.org>
>>>>> Date :  13/02/2012 12:12
>>>>> Objet : [OMPI devel] poor btl sm latency
>>>>> Envoyé par :    devel-boun...@open-mpi.org
>>>>> 
>>>>> 
>>>>> 
>>>>> Hello all,
>>>>> 
>>>>> on our new AMD cluster (AMD Opteron 6274, 2,2GHz) we get very bad
>>>>> latencies
>>>>> (~1.5us) when performing 0-byte p2p communication on one single node
>>>>> using the
>>>>> Open MPI sm BTL. When using Platform MPI we get ~0.5us latencies which
>>>>> is pretty good. The bandwidth results are similar for both MPI
>>>>> implementations
>>>>> (~3,3GB/s) - this is okay.
>>>>> 
>>>>> One node has 64 cores and 64Gb RAM where it doesn't matter how many
>>>>> ranks allocated by the application. We get similar results with
>>>>> different number of
>>>>> ranks.
>>>>> 
>>>>> We are using Open MPI 1.5.4 which is built by gcc 4.3.4 without any
>>>>> special
>>>>> configure options except the installation prefix and the location of
>>>>> the LSF
>>>>> stuff.
>>>>> 
>>>>> As mentioned at http://www.open-mpi.org/faq/?category=sm we tried to
>>>>> use /dev/shm instead of /tmp for the session directory, but it had no
>>>>> effect. Furthermore, we tried the current release candidate 1.5.5rc1
>>>>> of Open MPI which
>>>>> provides an option to use the SysV shared memory (-mca shmem sysv) -
>>>>> also this
>>>>> results in similar poor latencies.
>>>>> 
>>>>> Do you have any idea? Please help!
>>>>> 
>>>>> Thanks,
>>>>> Matthias
>>>>> _______________________________________________
>>>>> devel mailing list
>>>>> de...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>>> 
>>>> _______________________________________________
>>>> devel mailing list
>>>> de...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> 
>>> _______________________________________________
>>> devel mailing list
>>> de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/


Reply via email to