Hi Christoph, I applied the commits - pull/#1250 as Nathan told me and added "-mca btl_openib_flags 311" to the mpirun command line option, then it worked for me. I don't know the reason, but it looks ATOMIC_FOP in the btl_openib_flags degrades the sm/vader perfomance.
Regards, Tetsuya Mishima 2016/08/05 22:10:37、"devel"さんは「Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0」で書きました > Hello, > > We see the same problem here on various machines with Open MPI 2.0.0. > To us it seems that enabling the openib btl triggers bad performance for the sm AND vader btls! > --mca btl_base_verbose 10 reports in both cases the correct use of sm and vader between MPI ranks - only performance differs?! > > One irritating thing I see in the log output is the following: > openib BTL: rdmacm CPC unavailable for use on mlx4_0:1; skipped > [rank=1] openib: using port mlx4_0:1 > select: init of component openib returned success > > Did not look into the "Skipped" code part yet, ... > > Results see below. > > Best regards > Christoph Niethammer > > -- > > Christoph Niethammer > High Performance Computing Center Stuttgart (HLRS) > Nobelstrasse 19 > 70569 Stuttgart > > Tel: ++49(0)711-685-87203 > email: nietham...@hlrs.de > http://www.hlrs.de/people/niethammer > > > > mpirun -np 2 --mca btl self,vader osu_bw > # OSU MPI Bandwidth Test > # Size Bandwidth (MB/s) > 1 4.83 > 2 10.30 > 4 24.68 > 8 49.27 > 16 95.80 > 32 187.52 > 64 270.82 > 128 405.00 > 256 659.26 > 512 1165.14 > 1024 2372.83 > 2048 3592.85 > 4096 4283.51 > 8192 5523.55 > 16384 7388.92 > 32768 7024.37 > 65536 7353.79 > 131072 7465.96 > 262144 8597.56 > 524288 9292.86 > 1048576 9168.01 > 2097152 9009.62 > 4194304 9013.02 > > mpirun -np 2 --mca btl self,vader,openib osu_bw > # OSU MPI Bandwidth Test > # Size Bandwidth (MB/s) > 1 5.32 > 2 11.14 > 4 20.88 > 8 49.26 > 16 99.11 > 32 197.42 > 64 301.08 > 128 413.64 > 256 651.15 > 512 1161.12 > 1024 2460.99 > 2048 3627.36 > 4096 2191.06 > 8192 3118.36 > 16384 3428.45 > 32768 3676.96 > 65536 3709.65 > 131072 3748.64 > 262144 3764.88 > 524288 3764.61 > 1048576 3772.45 > 2097152 3757.37 > 4194304 3746.45 > > mpirun -np 2 --mca btl self,sm osu_bw > # OSU MPI Bandwidth Test > # Size Bandwidth (MB/s) > 1 2.98 > 2 5.97 > 4 11.99 > 8 23.47 > 16 50.64 > 32 99.91 > 64 197.87 > 128 343.32 > 256 667.48 > 512 1200.86 > 1024 2050.05 > 2048 3578.52 > 4096 3966.92 > 8192 5687.96 > 16384 7395.88 > 32768 7101.41 > 65536 7619.49 > 131072 7978.09 > 262144 8648.87 > 524288 9129.18 > 1048576 10525.31 > 2097152 10511.63 > 4194304 10489.66 > > mpirun -np 2 --mca btl self,sm,openib osu_bw > # OSU MPI Bandwidth Test > # Size Bandwidth (MB/s) > 1 2.02 > 2 3.00 > 4 9.99 > 8 19.96 > 16 40.10 > 32 70.63 > 64 144.08 > 128 282.21 > 256 543.55 > 512 1032.61 > 1024 1871.09 > 2048 3294.07 > 4096 2336.48 > 8192 3142.22 > 16384 3419.93 > 32768 3647.30 > 65536 3725.40 > 131072 3749.43 > 262144 3765.31 > 524288 3771.06 > 1048576 3772.54 > 2097152 3760.93 > 4194304 3745.37 > > ----- Original Message ----- > From: tmish...@jcity.maeda.co.jp > To: "Open MPI Developers" <de...@open-mpi.org> > Sent: Wednesday, July 27, 2016 6:04:48 AM > Subject: Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0 > > HiNathan, > > I applied those commits and ran again without any BTL specified. > > Then, although it says "mca: bml: Using vader btl for send to [[18993,1],1] > on node manage", > the osu_bw still shows it's very slow as shown below: > > [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca btl_base_verbose > 10 -bind-to core -report-bindings osu_bw > [manage.cluster:17482] MCW rank 0 bound to socket 0[core 0[hwt 0]]: > [B/././././.][./././././.] > [manage.cluster:17482] MCW rank 1 bound to socket 0[core 1[hwt 0]]: > [./B/./././.][./././././.] > [manage.cluster:17487] mca: base: components_register: registering > framework btl components > [manage.cluster:17487] mca: base: components_register: found loaded > component self > [manage.cluster:17487] mca: base: components_register: component self > register function successful > [manage.cluster:17487] mca: base: components_register: found loaded > component vader > [manage.cluster:17488] mca: base: components_register: registering > framework btl components > [manage.cluster:17488] mca: base: components_register: found loaded > component self > [manage.cluster:17487] mca: base: components_register: component vader > register function successful > [manage.cluster:17488] mca: base: components_register: component self > register function successful > [manage.cluster:17488] mca: base: components_register: found loaded > component vader > [manage.cluster:17487] mca: base: components_register: found loaded > component tcp > [manage.cluster:17488] mca: base: components_register: component vader > register function successful > [manage.cluster:17488] mca: base: components_register: found loaded > component tcp > [manage.cluster:17487] mca: base: components_register: component tcp > register function successful > [manage.cluster:17487] mca: base: components_register: found loaded > component sm > [manage.cluster:17488] mca: base: components_register: component tcp > register function successful > [manage.cluster:17488] mca: base: components_register: found loaded > component sm > [manage.cluster:17487] mca: base: components_register: component sm > register function successful > [manage.cluster:17488] mca: base: components_register: component sm > register function successful > [manage.cluster:17488] mca: base: components_register: found loaded > component openib > [manage.cluster:17487] mca: base: components_register: found loaded > component openib > [manage.cluster:17488] mca: base: components_register: component openib > register function successful > [manage.cluster:17488] mca: base: components_open: opening btl components > [manage.cluster:17488] mca: base: components_open: found loaded component > self > [manage.cluster:17488] mca: base: components_open: component self open > function successful > [manage.cluster:17488] mca: base: components_open: found loaded component > vader > [manage.cluster:17488] mca: base: components_open: component vader open > function successful > [manage.cluster:17488] mca: base: components_open: found loaded component > tcp > [manage.cluster:17488] mca: base: components_open: component tcp open > function successful > [manage.cluster:17488] mca: base: components_open: found loaded component > sm > [manage.cluster:17488] mca: base: components_open: component sm open > function successful > [manage.cluster:17488] mca: base: components_open: found loaded component > openib > [manage.cluster:17488] mca: base: components_open: component openib open > function successful > [manage.cluster:17488] select: initializing btl component self > [manage.cluster:17488] select: init of component self returned success > [manage.cluster:17488] select: initializing btl component vader > [manage.cluster:17487] mca: base: components_register: component openib > register function successful > [manage.cluster:17487] mca: base: components_open: opening btl components > [manage.cluster:17487] mca: base: components_open: found loaded component > self > [manage.cluster:17487] mca: base: components_open: component self open > function successful > [manage.cluster:17487] mca: base: components_open: found loaded component > vader > [manage.cluster:17487] mca: base: components_open: component vader open > function successful > [manage.cluster:17487] mca: base: components_open: found loaded component > tcp > [manage.cluster:17487] mca: base: components_open: component tcp open > function successful > [manage.cluster:17487] mca: base: components_open: found loaded component > sm > [manage.cluster:17487] mca: base: components_open: component sm open > function successful > [manage.cluster:17487] mca: base: components_open: found loaded component > openib > [manage.cluster:17488] select: init of component vader returned success > [manage.cluster:17488] select: initializing btl component tcp > [manage.cluster:17487] mca: base: components_open: component openib open > function successful > [manage.cluster:17487] select: initializing btl component self > [manage.cluster:17487] select: init of component self returned success > [manage.cluster:17487] select: initializing btl component vader > [manage.cluster:17488] select: init of component tcp returned success > [manage.cluster:17488] select: initializing btl component sm > [manage.cluster:17488] select: init of component sm returned success > [manage.cluster:17488] select: initializing btl component openib > [manage.cluster:17487] select: init of component vader returned success > [manage.cluster:17487] select: initializing btl component tcp > [manage.cluster:17487] select: init of component tcp returned success > [manage.cluster:17487] select: initializing btl component sm > [manage.cluster:17488] Checking distance from this process to device=mthca0 > [manage.cluster:17488] hwloc_distances->nbobjs=2 > [manage.cluster:17488] hwloc_distances->latency[0]=1.000000 > [manage.cluster:17488] hwloc_distances->latency[1]=1.600000 > [manage.cluster:17488] hwloc_distances->latency[2]=1.600000 > [manage.cluster:17488] hwloc_distances->latency[3]=1.000000 > [manage.cluster:17488] ibv_obj->type set to NULL > [manage.cluster:17488] Process is bound: distance to device is 0.000000 > [manage.cluster:17487] select: init of component sm returned success > [manage.cluster:17487] select: initializing btl component openib > [manage.cluster:17488] openib BTL: rdmacm CPC unavailable for use on > mthca0:1; skipped > [manage.cluster:17487] Checking distance from this process to device=mthca0 > [manage.cluster:17487] hwloc_distances->nbobjs=2 > [manage.cluster:17487] hwloc_distances->latency[0]=1.000000 > [manage.cluster:17487] hwloc_distances->latency[1]=1.600000 > [manage.cluster:17487] hwloc_distances->latency[2]=1.600000 > [manage.cluster:17487] hwloc_distances->latency[3]=1.000000 > [manage.cluster:17487] ibv_obj->type set to NULL > [manage.cluster:17487] Process is bound: distance to device is 0.000000 > [manage.cluster:17488] [rank=1] openib: using port mthca0:1 > [manage.cluster:17488] select: init of component openib returned success > [manage.cluster:17487] openib BTL: rdmacm CPC unavailable for use on > mthca0:1; skipped > [manage.cluster:17487] [rank=0] openib: using port mthca0:1 > [manage.cluster:17487] select: init of component openib returned success > [manage.cluster:17488] mca: bml: Using self btl for send to [[18993,1],1] > on node manage > [manage.cluster:17487] mca: bml: Using self btl for send to [[18993,1],0] > on node manage > [manage.cluster:17488] mca: bml: Using vader btl for send to [[18993,1],0] > on node manage > [manage.cluster:17487] mca: bml: Using vader btl for send to [[18993,1],1] > on node manage > # OSU MPI Bandwidth Test v3.1.1 > # Size Bandwidth (MB/s) > 1 1.76 > 2 3.53 > 4 7.06 > 8 14.46 > 16 29.12 > 32 57.54 > 64 100.12 > 128 157.78 > 256 277.32 > 512 477.53 > 1024 894.81 > 2048 1330.68 > 4096 278.58 > 8192 516.00 > 16384 762.99 > 32768 1037.19 > 65536 1181.66 > 131072 1261.91 > 262144 1237.39 > 524288 1247.86 > 1048576 1252.04 > 2097152 1273.46 > 4194304 1281.21 > [manage.cluster:17488] mca: base: close: component self closed > [manage.cluster:17488] mca: base: close: unloading component self > [manage.cluster:17487] mca: base: close: component self closed > [manage.cluster:17487] mca: base: close: unloading component self > [manage.cluster:17488] mca: base: close: component vader closed > [manage.cluster:17488] mca: base: close: unloading component vader > [manage.cluster:17487] mca: base: close: component vader closed > [manage.cluster:17487] mca: base: close: unloading component vader > [manage.cluster:17488] mca: base: close: component tcp closed > [manage.cluster:17488] mca: base: close: unloading component tcp > [manage.cluster:17487] mca: base: close: component tcp closed > [manage.cluster:17487] mca: base: close: unloading component tcp > [manage.cluster:17488] mca: base: close: component sm closed > [manage.cluster:17488] mca: base: close: unloading component sm > [manage.cluster:17487] mca: base: close: component sm closed > [manage.cluster:17487] mca: base: close: unloading component sm > [manage.cluster:17488] mca: base: close: component openib closed > [manage.cluster:17488] mca: base: close: unloading component openib > [manage.cluster:17487] mca: base: close: component openib closed > [manage.cluster:17487] mca: base: close: unloading component openib > > Tetsuya Mishima > > 2016/07/27 9:20:28、"devel"さんは「Re: [OMPI devel] sm BTL performace of > the openmpi-2.0.0」で書きました > > sm is deprecated in 2.0.0 and will likely be removed in favor of vader in > 2.1.0. > > > > This issue is probably this known issue: > https://github.com/open-mpi/ompi-release/pull/1250 > > > > Please apply those commits and see if it fixes the issue for you. > > > > -Nathan > > > > > On Jul 26, 2016, at 6:17 PM, tmish...@jcity.maeda.co.jp wrote: > > > > > > Hi Gilles, > > > > > > Thanks. I ran again with --mca pml ob1 but I've got the same results as > > > below: > > > > > > [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca pml ob1 > -bind-to > > > core -report-bindings osu_bw > > > [manage.cluster:18142] MCW rank 0 bound to socket 0[core 0[hwt 0]]: > > > [B/././././.][./././././.] > > > [manage.cluster:18142] MCW rank 1 bound to socket 0[core 1[hwt 0]]: > > > [./B/./././.][./././././.] > > > # OSU MPI Bandwidth Test v3.1.1 > > > # Size Bandwidth (MB/s) > > > 1 1.48 > > > 2 3.07 > > > 4 6.26 > > > 8 12.53 > > > 16 24.33 > > > 32 49.03 > > > 64 83.46 > > > 128 132.60 > > > 256 234.96 > > > 512 420.86 > > > 1024 842.37 > > > 2048 1231.65 > > > 4096 264.67 > > > 8192 472.16 > > > 16384 740.42 > > > 32768 1030.39 > > > 65536 1191.16 > > > 131072 1269.45 > > > 262144 1238.33> > 524288 1247.97 > > > 1048576 1257.96 > > > 2097152 1274.74 > > > 4194304 1280.94 > > > [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca pml ob1 -mca > btl > > > self,sm -bind-to core -report-bindings osu_b > > > w > > > [manage.cluster:18204] MCW rank 0 bound to socket 0[core 0[hwt 0]]: > > > [B/././././.][./././././.] > > > [manage.cluster:18204] MCW rank 1 bound to socket 0[core 1[hwt 0]]: > > > [./B/./././.][./././././.] > > > # OSU MPI Bandwidth Test v3.1.1 > > > # Size Bandwidth (MB/s) > > > 1 0.52 > > > 2 1.05 > > > 4 2.08 > > > 8 4.18 > > > 16 8.21 > > > 32 16.65 > > > 64 32.60 > > > 128 66.70 > > > 256 132.45 > > > 512 269.27 > > > 1024 504.63 > > > 2048 819.76 > > > 4096 874.54 > > > 8192 1447.11 > > > 16384 2263.28 > > > 32768 3236.85 > > > 65536 3567.34 > > > 131072 3555.17 > > > 262144 3455.76 > > > 524288 3441.80 > > > 1048576 3505.30 > > > 2097152 3534.01 > > > 4194304 3546.94 > > > [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca pml ob1 -mca > btl > > > self,sm,openib -bind-to core -report-binding > > > s osu_bw > > > [manage.cluster:18218] MCW rank 0 bound to socket 0[core 0[hwt 0]]: > > > [B/././././.][./././././.] > > > [manage.cluster:18218] MCW rank 1 bound to socket 0[core 1[hwt 0]]: > > > [./B/./././.][./././././.] > > > # OSU MPI Bandwidth Test v3.1.1 > > > # Size Bandwidth (MB/s) > > > 1 0.51 > > > 2 1.03 > > > 4 2.05 > > > 8 4.07 > > > 16 8.14 > > > 32 16.32 > > > 64 32.98 > > > 128 63.70 > > > 256 126.66 > > > 512 252.61 > > > 1024 480.22 > > > 2048 810.54 > > > 4096 290.61 > > > 8192 512.49 > > > 16384 764.60 > > > 32768 1036.81 > > > 65536 1182.81 > > > 131072 1264.48 > > > 262144 1235.82 > > > 524288 1246.70 > > > 1048576 1254.66 > > > 2097152 1274.64 > > > 4194304 1280.65 > > > [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca pml ob1 -mca > btl > > > self,openib -bind-to core -report-bindings o > > > su_bw > > > [manage.cluster:18276] MCW rank 0 bound to socket 0[core 0[hwt 0]]: > > > [B/././././.][./././././.] > > > [manage.cluster:18276] MCW rank 1 bound to socket 0[core 1[hwt 0]]: > > > [./B/./././.][./././././.] > > > # OSU MPI Bandwidth Test v3.1.1 > > > # Size Bandwidth (MB/s) > > > 1 0.54 > > > 2 1.08 > > > 4 2.18 > > > 8 4.33 > > > 16 8.69 > > > 32 17.39 > > > 64 34.34 > > > 128 66.28 > > > 256 130.36 > > > 512 241.81 > > > 1024 429.86 > > > 2048 553.44 > > > 4096 707.14 > > > 8192 879.60 > > > 16384 763.02 > > > 32768 1042.89 > > > 65536 1185.45 > > > 131072 1267.56 > > > 262144 1227.41 > > > 524288 1244.61 > > > 1048576 1255.66 > > > 2097152 1273.55 > > > 4194304 1281.05 > > > > > > > > > 2016/07/27 9:02:49、"devel"さんは「Re: [OMPI devel] sm BTL performace > of > > > the openmpi-2.0.0」で書きました > > >> Hi, > > >> > > >> > > >> can you please run again with > > >> > > >> --mca pml ob1 > > >> > > >> > > >> if Open MPI was built with mxm support, pml/cm and mtl/mxm are used > > >> instead of pml/ob1 and btl/openib > > >> > > >> > > >> Cheers, > > >> > > >> > > >> Gilles > > >> > > >> > > >> On 7/27/2016 8:56 AM, tmish...@jcity.maeda.co.jp wrote: > > >>> Hi folks, > > >>> > > >>> I saw a performance degradation of openmpi-2.0.0 when I ran our > > > application > > >>> on a node (12cores). So I did 4 tests using osu_bw as below: > > >>> > > >>> 1: mpirun –np 2 osu_bw bad(30% of test2) > > >>> 2: mpirun –np 2 –mca btl self,sm osu_bw good(same as > > > openmpi1.10.3) > > >>> 3: mpirun –np 2 –mca btl self,sm,openib osu_bw bad(30% of test2) > > >>> 4: mpirun –np 2 –mca btl self,openib osu_bw bad(30% of test2) > > >>> > > >>> I guess openib BTL was used in the test 1 and 3, because these > results > > > are > > >>> almost same as test 4. I believe that sm BTL should be used even > in > > > the > > >>> test 1 and 3, because its priority is higher than openib. > > > Unfortunately, at > > >>> the moment, I couldn’t figure out the root cause. So please someone > > > would > > >>> take care of it. > > >>> > > >>> Regards, > > >>> Tetsuya Mishima > > >>> > > >>> P.S. Here I attached these test results. > > >>> > > >>> [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -bind-to > > > core > > >>> -report-bindings osu_bw > > >>> [manage.cluster:13389] MCW rank 0 bound to socket 0[core 0 > [hwt > > > 0]]: > > >>> [B/././././.][./././././.] > > >>> [manage.cluster:13389] MCW rank 1 bound to socket 0[core 1 > [hwt > > > 0]]: > > >>> [./B/./././.][./././././.] > > >>> # OSU MPI Bandwidth Test v3.1.1 > > >>> # Size Bandwidth (MB/s) > > >>> 1 1.49 > > >>> 2 3.04 > > >>> 4 6.13 > > >>> 8 12.23 > > >>> 16 25.01 > > >>> 32 49.96 > > >>> 64 87.07 > > >>> 128 138.87 > > >>> 256 245.97 > > >>> 512 423.30 > > >>> 1024 865.85 > > >>> 2048 1279.63 > > >>> 4096 264.79 > > >>> 8192 473.92 > > >>> 16384 739.27 > > >>> 32768 1030.49 > > >>> 65536 1190.21 > > >>> 131072 1270.77 > > >>> 262144 1238.74 > > >>> 524288 1245.97 > > >>> 1048576 1260.09 > > >>> 2097152 1274.53 > > >>> 4194304 1285.07 > > >>> [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca btl > > > self,sm > > >>> -bind-to core -report-bindings osu_bw > > >>> [manage.cluster:13448] MCW rank 0 bound to socket 0[core 0 > [hwt > > > 0]]: > > >>> [B/././././.][./././././.] > > >>> [manage.cluster:13448] MCW rank 1 bound to socket 0[core 1 > [hwt > > > 0]]: > > >>> [./B/./././.][./././././.] > > >>> # OSU MPI Bandwidth Test v3.1.1 > > >>> # Size Bandwidth (MB/s) > > >>> 1 0.51 > > >>> 2 1.01 > > >>> 4 2.03 > > >>> 8 4.08 > > >>> 16 7.92 > > >>> 32 16.16 > > >>> 64 32.53 > > >>> 128 64.30 > > >>> 256 128.19 > > >>> 512 256.48 > > >>> 1024 468.62 > > >>> 2048 785.29 > > >>> 4096 854.78 > > >>> 8192 1404.51 > > >>> 16384 2249.20 > > >>> 32768 3136.40 > > >>> 65536 3495.84 > > >>> 131072 3436.69 > > >>> 262144 3392.11 > > >>> 524288 3400.07 > > >>> 1048576 3460.60 > > >>> 2097152 3488.09 > > >>> 4194304 3498.45 > > >>> [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca > > > btl > > >>> self,sm,openib -bind-to core -report-bindings osu_bw > > >>> [manage.cluster:13462] MCW rank 0 bound to socket 0[core 0 > [hwt > > > 0]]: > > >>> [B/././././.][./././././.] > > >>> [manage.cluster:13462] MCW rank 1 bound to socket 0[core 1 > [hwt > > > 0]]: > > >>> [./B/./././.][./././././.] > > >>> # OSU MPI Bandwidth Test v3.1.1 > > >>> # Size Bandwidth (MB/s) > > >>> 1 0.54 > > >>> 2 1.09 > > >>> 4 2.18 > > >>> 8 4.37 > > >>> 16 8.75 > > >>> 32 17.37 > > >>> 64 34.67 > > >>> 128 66.66 > > >>> 256 132.55 > > >>> 512 261.52 > > >>> 1024 489.51 > > >>> 2048 818.38 > > >>> 4096 290.48 > > >>> 8192 511.64 > > >>> 16384 765.24 > > >>> 32768 1043.28 > > >>> 65536 1180.48 > > >>> 131072 1261.41 > > >>> 262144 1232.86 > > >>> 524288 1245.70 > > >>> 1048576 1245.69 > > >>> 2097152 1268.67 > > >>> 4194304 1281.33 > > >>> [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca btl > > > self,openib > > >>> -bind-to core -report-bindings osu_bw > > >>> [manage.cluster:13521] MCW rank 0 bound to socket 0[core 0 > [hwt > > > 0]]: > > >>> [B/././././.][./././././.] > > >>> [manage.cluster:13521] MCW rank 1 bound to socket 0[core 1 > [hwt > > > 0]]: > > >>> [./B/./././.][./././././.] > > >>> # OSU MPI Bandwidth Test v3.1.1 > > >>> # Size Bandwidth (MB/s) > > >>> 1 0.54 > > >>> 2 1.08 > > >>> 4 2.16 > > >>> 8 4.34 > > >>> 16 8.64 > > >>> 32 17.25 > > >>> 64 34.30 > > >>> 128 66.13 > > >>> 256 129.99 > > >>> 512 242.26 > > >>> 1024 429.24 > > >>> 2048 556.00 > > >>> 4096 706.80 > > >>> 8192 874.35 > > >>> 16384 762.60 > > >>> 32768 1039.61 > > >>> 65536 1184.03 > > >>> 131072 1267.09 > > >>> 262144 1230.76 > > >>> 524288 1246.92 > > >>> 1048576 1255.88 > > >>> 2097152 1274.54 > > >>> 4194304 > > > 1281.63 > > >>> _______________________________________________ > > >>> devel mailing list > > >>> de...@open-mpi.org > > >>> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel > > >>> Link to this post: > > > http://www.open-mpi.org/community/lists/devel/2016/07/19288.php > > >> > > >> _______________________________________________ > > >> devel mailing list > > >> de...@open-mpi.org > > >> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/develLink > to > > > this post: > http://www.open-mpi.org/community/lists/devel/2016/07/19289.php > > > _______________________________________________ > > > devel mailing list > > > de...@open-mpi.org > > > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel > > > Link to this post: > http://www.open-mpi.org/community/lists/devel/2016/07/19290.php > > > > _______________________________________________ > > devel mailing list > > de...@open-mpi.org > > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/develLink to > this post: http://www.open-mpi.org/community/lists/devel/2016/07/19291.php > _______________________________________________ > devel mailing list > de...@open-mpi.org > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel > Link to this post: http://www.open-mpi.org/community/lists/devel/2016/07/19295.php > _______________________________________________ > devel mailing list > de...@lists.open-mpi.orghttps://rfd.newmexicoconsortium.org/mailman/listinfo/devel _______________________________________________ devel mailing list devel@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/devel