Does UCT support iWARP?
> On Apr 6, 2018, at 9:45 AM, Jeff Squyres (jsquyres) <jsquy...@cisco.com> > wrote: > > That would be sweet. Are you aiming for v4.0.0, perchance? > > I.e., should we have "remove openib BTL / add uct BTL" to the target feature > list for v4.0.0? > > (I did laugh at the Kylo name, but I really, really don't want to keep the > propagate the idea of meaningless names for BTLs -- especially since BTL > names, more so than most other component names, are visible to the user. > ...unless someone wants to finally finish the ideas and implement the whole > network-transport-name system that we've talked about for a few years... :-) ) > > >> On Apr 5, 2018, at 1:49 PM, Thananon Patinyasakdikul <tpati...@vols.utk.edu> >> wrote: >> >> Just more information to help with the decision: >> >> I am working on Nathan’s uct btl to make it work with ob1 and infiniband. So >> this could be a replacement for openib and honestly we should totally call >> this new uct btl Kylo. >> >> Arm >> >>> On Apr 5, 2018, at 1:37 PM, Jeff Squyres (jsquyres) <jsquy...@cisco.com> >>> wrote: >>> >>> Below is an email exchange from the users mailing list. >>> >>> I'm moving this over to devel to talk among the developer community. >>> >>> Multiple times recently on the users list, we've told people with problems >>> with the openib BTL that they should be using UCX (per Mellanox's >>> publicly-stated support positions). >>> >>> Is it time to deprecate / print warning messages / remove the openib BTL? >>> >>> >>> >>>> Begin forwarded message: >>>> >>>> From: Nathan Hjelm <hje...@me.com> >>>> Subject: Re: [OMPI users] Eager RDMA causing slow osu_bibw with 3.0.0 >>>> Date: April 5, 2018 at 12:48:08 PM EDT >>>> To: Open MPI Users <us...@lists.open-mpi.org> >>>> Cc: Open MPI Users <us...@lists.open-mpi.org> >>>> Reply-To: Open MPI Users <us...@lists.open-mpi.org> >>>> >>>> >>>> Honestly, this is a configuration issue with the openib btl. There is no >>>> reason to keep either eager RDMA nor is there a reason to pipeline RDMA. I >>>> haven't found an app where either of these "features" helps you with >>>> infiniband. You have the right idea with the parameter changes but Howard >>>> is correct, for Mellanox the future is UCX not verbs. I would try it and >>>> see if it works for you but if it doesn't I would set those two parameters >>>> in your /etc/openmpi-mca-params.conf and run like that. >>>> >>>> -Nathan >>>> >>>> On Apr 05, 2018, at 01:18 AM, Ben Menadue <ben.mena...@nci.org.au> wrote: >>>> >>>>> Hi, >>>>> >>>>> Another interesting point. I noticed that the last two message sizes >>>>> tested (2MB and 4MB) are lower than expected for both osu_bw and >>>>> osu_bibw. Increasing the minimum size to use the RDMA pipeline to above >>>>> these sizes brings those two data-points up to scratch for both >>>>> benchmarks: >>>>> >>>>> 3.0.0, osu_bw, no rdma for large messages >>>>> >>>>>> mpirun -mca btl_openib_min_rdma_pipeline_size 4194304 -map-by ppr:1:node >>>>>> -np 2 -H r6,r7 ./osu_bw -m 2097152:4194304 >>>>> # OSU MPI Bi-Directional Bandwidth Test v5.4.0 >>>>> # Size Bandwidth (MB/s) >>>>> 2097152 6133.22 >>>>> 4194304 6054.06 >>>>> >>>>> 3.0.0, osu_bibw, eager rdma disabled, no rdma for large messages >>>>> >>>>>> mpirun -mca btl_openib_min_rdma_pipeline_size 4194304 -mca >>>>>> btl_openib_use_eager_rdma 0 -map-by ppr:1:node -np 2 -H r6,r7 ./osu_bibw >>>>>> -m 2097152:4194304 >>>>> # OSU MPI Bi-Directional Bandwidth Test v5.4.0 >>>>> # Size Bandwidth (MB/s) >>>>> 2097152 11397.85 >>>>> 4194304 11389.64 >>>>> >>>>> This makes me think something odd is going on in the RDMA pipeline. >>>>> >>>>> Cheers, >>>>> Ben >>>>> >>>>> >>>>> >>>>>> On 5 Apr 2018, at 5:03 pm, Ben Menadue <ben.mena...@nci.org.au> wrote: >>>>>> Hi, >>>>>> >>>>>> We’ve just been running some OSU benchmarks with OpenMPI 3.0.0 and >>>>>> noticed that osu_bibw gives nowhere near the bandwidth I’d expect (this >>>>>> is on FDR IB). However, osu_bw is fine. >>>>>> >>>>>> If I disable eager RDMA, then osu_bibw gives the expected numbers. >>>>>> Similarly, if I increase the number of eager RDMA buffers, it gives the >>>>>> expected results. >>>>>> >>>>>> OpenMPI 1.10.7 gives consistent, reasonable numbers with default >>>>>> settings, but they’re not as good as 3.0.0 (when tuned) for large >>>>>> buffers. The same option changes produce no different in the performance >>>>>> for 1.10.7. >>>>>> >>>>>> I was wondering if anyone else has noticed anything similar, and if this >>>>>> is unexpected, if anyone has a suggestion on how to investigate further? >>>>>> >>>>>> Thanks, >>>>>> Ben >>>>>> >>>>>> >>>>>> Here’s are the numbers: >>>>>> >>>>>> 3.0.0, osu_bw, default settings >>>>>> >>>>>>> mpirun -map-by ppr:1:node -np 2 -H r6,r7 ./osu_bw >>>>>> # OSU MPI Bandwidth Test v5.4.0 >>>>>> # Size Bandwidth (MB/s) >>>>>> 1 1.13 >>>>>> 2 2.29 >>>>>> 4 4.63 >>>>>> 8 9.21 >>>>>> 16 18.18 >>>>>> 32 36.46 >>>>>> 64 69.95 >>>>>> 128 128.55 >>>>>> 256 250.74 >>>>>> 512 451.54 >>>>>> 1024 829.44 >>>>>> 2048 1475.87 >>>>>> 4096 2119.99 >>>>>> 8192 3452.37 >>>>>> 16384 2866.51 >>>>>> 32768 4048.17 >>>>>> 65536 5030.54 >>>>>> 131072 5573.81 >>>>>> 262144 5861.61 >>>>>> 524288 6015.15 >>>>>> 1048576 6099.46 >>>>>> 2097152 989.82 >>>>>> 4194304 989.81 >>>>>> >>>>>> 3.0.0, osu_bibw, default settings >>>>>> >>>>>>> mpirun -map-by ppr:1:node -np 2 -H r6,r7 ./osu_bibw >>>>>> # OSU MPI Bi-Directional Bandwidth Test v5.4.0 >>>>>> # Size Bandwidth (MB/s) >>>>>> 1 0.00 >>>>>> 2 0.01 >>>>>> 4 0.01 >>>>>> 8 0.02 >>>>>> 16 0.04 >>>>>> 32 0.09 >>>>>> 64 0.16 >>>>>> 128 135.30 >>>>>> 256 265.35 >>>>>> 512 499.92 >>>>>> 1024 949.22 >>>>>> 2048 1440.27 >>>>>> 4096 1960.09 >>>>>> 8192 3166.97 >>>>>> 16384 127.62 >>>>>> 32768 165.12 >>>>>> 65536 312.80 >>>>>> 131072 1120.03 >>>>>> 262144 4724.01 >>>>>> 524288 4545.93 >>>>>> 1048576 5186.51 >>>>>> 2097152 989.84 >>>>>> 4194304 989.88 >>>>>> >>>>>> 3.0.0, osu_bibw, eager RDMA disabled >>>>>> >>>>>>> mpirun -mca btl_openib_use_eager_rdma 0 -map-by ppr:1:node -np 2 -H >>>>>>> r6,r7 ./osu_bibw >>>>>> # OSU MPI Bi-Directional Bandwidth Test v5.4.0 >>>>>> # Size Bandwidth (MB/s) >>>>>> 1 1.49 >>>>>> 2 2.97 >>>>>> 4 5.96 >>>>>> 8 11.98 >>>>>> 16 23.95 >>>>>> 32 47.39 >>>>>> 64 93.57 >>>>>> 128 153.82 >>>>>> 256 304.69 >>>>>> 512 572.30 >>>>>> 1024 1003.52 >>>>>> 2048 1083.89 >>>>>> 4096 1879.32 >>>>>> 8192 2785.18 >>>>>> 16384 3535.77 >>>>>> 32768 5614.72 >>>>>> 65536 8113.69 >>>>>> 131072 9666.74 >>>>>> 262144 10738.97 >>>>>> 524288 11247.02 >>>>>> 1048576 11416.50 >>>>>> 2097152 989.88 >>>>>> 4194304 989.88 >>>>>> >>>>>> 3.0.0, osu_bibw, increased eager RDMA buffer count >>>>>> >>>>>>> mpirun -mca btl_openib_eager_rdma_num 32768 -map-by ppr:1:node -np 2 -H >>>>>>> r6,r7 ./osu_bibw >>>>>> # OSU MPI Bi-Directional Bandwidth Test v5.4.0 >>>>>> # Size Bandwidth (MB/s) >>>>>> 1 1.42 >>>>>> 2 2.84 >>>>>> 4 5.67 >>>>>> 8 11.18 >>>>>> 16 22.46 >>>>>> 32 44.65 >>>>>> 64 83.10 >>>>>> 128 154.00 >>>>>> 256 291.63 >>>>>> 512 537.66 >>>>>> 1024 942.35 >>>>>> 2048 1433.09 >>>>>> 4096 2356.40 >>>>>> 8192 1998.54 >>>>>> 16384 3584.82 >>>>>> 32768 5523.08 >>>>>> 65536 7717.63 >>>>>> 131072 9419.50 >>>>>> 262144 10564.77 >>>>>> 524288 11104.71 >>>>>> 1048576 11130.75 >>>>>> 2097152 7943.89 >>>>>> 4194304 5270.00 >>>>>> >>>>>> 1.10.7, osu_bibw, default settings >>>>>> >>>>>>> mpirun -map-by ppr:1:node -np 2 -H r6,r7 ./osu_bibw >>>>>> # OSU MPI Bi-Directional Bandwidth Test v5.4.0 >>>>>> # Size Bandwidth (MB/s) >>>>>> 1 1.70 >>>>>> 2 3.45 >>>>>> 4 6.95 >>>>>> 8 13.68 >>>>>> 16 27.41 >>>>>> 32 53.80 >>>>>> 64 105.34 >>>>>> 128 164.40 >>>>>> 256 324.63 >>>>>> 512 623.95 >>>>>> 1024 1127.35 >>>>>> 2048 1784.58 >>>>>> 4096 3305.45 >>>>>> 8192 3697.55 >>>>>> 16384 4935.75 >>>>>> 32768 7186.28 >>>>>> 65536 8996.94 >>>>>> 131072 9301.78 >>>>>> 262144 4691.36 >>>>>> 524288 7039.18 >>>>>> 1048576 7213.33 >>>>>> 2097152 9601.41 >>>>>> 4194304 9281.31 >>>>>> >>>>>> >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@lists.open-mpi.org >>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>> _______________________________________________ >>>> users mailing list >>>> us...@lists.open-mpi.org >>>> https://lists.open-mpi.org/mailman/listinfo/users >>> >>> >>> -- >>> Jeff Squyres >>> jsquy...@cisco.com >>> >>> _______________________________________________ >>> devel mailing list >>> devel@lists.open-mpi.org >>> https://lists.open-mpi.org/mailman/listinfo/devel >> >> _______________________________________________ >> devel mailing list >> devel@lists.open-mpi.org >> https://lists.open-mpi.org/mailman/listinfo/devel > > > -- > Jeff Squyres > jsquy...@cisco.com > > _______________________________________________ > devel mailing list > devel@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/devel -- Jeff Squyres jsquy...@cisco.com _______________________________________________ devel mailing list devel@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/devel