Hello,Testing performance of OpenMPI over Infiniband I have the following result :
1) Hardware is : SilversStorm interface 2) Openmpi version is : (from ompi_info) Open MPI: 1.0.2a9r9159 Open MPI SVN revision: r9159 Open RTE: 1.0.2a9r9159 Open RTE SVN revision: r9159 OPAL: 1.0.2a9r9159 OPAL SVN revision: r9159 3) Cluster with Bi-processors Opteron 248 2.2 GHz Configure has been run with option --with-mvapi=path-to-mvapi 4) a C coded pinpong gives the following values : LOOPS: 1000 BYTES: 4096 SECONDS: 0.085557 MBytes/sec: 95.749051 LOOPS: 1000 BYTES: 8192 SECONDS: 0.050657 MBytes/sec: 323.429912 LOOPS: 1000 BYTES: 16384 SECONDS: 0.084038 MBytes/sec: 389.918757 LOOPS: 1000 BYTES: 32768 SECONDS: 0.163161 MBytes/sec: 401.665104 LOOPS: 1000 BYTES: 65536 SECONDS: 0.306694 MBytes/sec: 427.370561 LOOPS: 1000 BYTES: 131072 SECONDS: 0.529589 MBytes/sec: 494.995011 LOOPS: 1000 BYTES: 262144 SECONDS: 0.952616 MBytes/sec: 550.366583 LOOPS: 1000 BYTES: 524288 SECONDS: 1.927987 MBytes/sec: 543.870859 LOOPS: 1000 BYTES: 1048576 SECONDS: 3.673732 MBytes/sec: 570.850562 LOOPS: 1000 BYTES: 2097152 SECONDS: 9.993185 MBytes/sec: 419.716435 LOOPS: 1000 BYTES: 4194304 SECONDS: 18.211958 MBytes/sec: 460.609893 LOOPS: 1000 BYTES: 8388608 SECONDS: 35.421490 MBytes/sec: 473.645124 My questions are : a) Is OpenMPI doing in this case TCP/IP over IB ? (I guess so)b) Is it possible to improve significantly these values by changing the defaults ?
I have used several mca btl parameters but without improving the maximum bandwith.
For example : --mca btl mvapi --mca btl_mvapi_max_send_size 8388608 c) Is it possible that other IB hardware implementations have better performances with OpenMPI ?d) Is it possible to use specific IB drivers for optimal performance ? (should reach almost 800 MB/sec)
Thank you very much for your help Best Regards, Jean Latour
<<attachment: latour.vcf>>