Gus, Gilles, Russell, John: Thanks very much for the replies and the help. I got confirmation from the "root" that it is indeed RoCE with 100G.
I'll go over the info in the link Russell provided, but have a quick question: if I run the "*mpiexec*" with "*-mca btl tcp,self*" do I get the benefit of *RoCE *(the fastest speed)? I'll go over the details of all reply and post useful feedback. Thanks very much all! Best, --Boris On Mon, Jul 17, 2017 at 6:31 AM, Russell Dekema <deke...@umich.edu> wrote: > It looks like you have two dual-port Mellanox VPI cards in this > machine. These cards can be set to run InfiniBand or Ethernet on a > port-by-port basis, and all four of your ports are set to Ethernet > mode. Two of your ports have active 100 gigabit Ethernet links, and > the other two have no link up at all. > > With no InfiniBand links on the machine, you will, of course, not be > able to run your OpenMPI job over InfiniBand. > > If your machines and network are set up for it, you might be able to > run your job over RoCE (RDMA Over Converged Ethernet) using one or > both of those 100 GbE links. I have never used RoCE myself, but one > starting point for gathering more information on it might be the > following section of the OpenMPI FAQ: > > https://www.open-mpi.org/faq/?category=openfabrics#ompi-over-roce > > Sincerely, > Rusty Dekema > University of Michigan > Advanced Research Computing - Technology Services > > > On Fri, Jul 14, 2017 at 12:34 PM, Boris M. Vulovic > <boris.m.vulo...@gmail.com> wrote: > > Gus, Gilles and John, > > > > Thanks for the help. Let me first post (below) the output from checkouts > of > > the IB network: > > ibdiagnet > > ibhosts > > ibstat (for login node, for now) > > > > What do you think? > > Thanks > > --Boris > > > > > > %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% > %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% > > > > -bash-4.1$ ibdiagnet > > ---------- > > Load Plugins from: > > /usr/share/ibdiagnet2.1.1/plugins/ > > (You can specify more paths to be looked in with "IBDIAGNET_PLUGINS_PATH" > > env variable) > > > > Plugin Name Result Comment > > libibdiagnet_cable_diag_plugin-2.1.1 Succeeded Plugin loaded > > libibdiagnet_phy_diag_plugin-2.1.1 Succeeded Plugin loaded > > > > --------------------------------------------- > > Discovery > > -E- Failed to initialize > > > > -E- Fabric Discover failed, err=IBDiag initialize wasn't done > > -E- Fabric Discover failed, MAD err=Failed to register SMI class > > > > --------------------------------------------- > > Summary > > -I- Stage Warnings Errors Comment > > -I- Discovery NA > > -I- Lids Check NA > > -I- Links Check NA > > -I- Subnet Manager NA > > -I- Port Counters NA > > -I- Nodes Information NA > > -I- Speed / Width checks NA > > -I- Partition Keys NA > > -I- Alias GUIDs NA > > -I- Temperature Sensing NA > > > > -I- You can find detailed errors/warnings in: > > /var/tmp/ibdiagnet2/ibdiagnet2.log > > > > -E- A fatal error occurred, exiting... > > -bash-4.1$ > > %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% > %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% > > > > -bash-4.1$ ibhosts > > ibwarn: [168221] mad_rpc_open_port: client_register for mgmt 1 failed > > src/ibnetdisc.c:766; can't open MAD port ((null):0) > > /usr/sbin/ibnetdiscover: iberror: failed: discover failed > > -bash-4.1$ > > > > %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% > %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% > > -bash-4.1$ ibstat > > CA 'mlx5_0' > > CA type: MT4115 > > Number of ports: 1 > > Firmware version: 12.17.2020 > > Hardware version: 0 > > Node GUID: 0x248a0703005abb1c > > System image GUID: 0x248a0703005abb1c > > Port 1: > > State: Active > > Physical state: LinkUp > > Rate: 100 > > Base lid: 0 > > LMC: 0 > > SM lid: 0 > > Capability mask: 0x3c010000 > > Port GUID: 0x268a07fffe5abb1c > > Link layer: Ethernet > > CA 'mlx5_1' > > CA type: MT4115 > > Number of ports: 1 > > Firmware version: 12.17.2020 > > Hardware version: 0 > > Node GUID: 0x248a0703005abb1d > > System image GUID: 0x248a0703005abb1c > > Port 1: > > State: Active > > Physical state: LinkUp > > Rate: 100 > > Base lid: 0 > > LMC: 0 > > SM lid: 0 > > Capability mask: 0x3c010000 > > Port GUID: 0x0000000000000000 > > Link layer: Ethernet > > CA 'mlx5_2' > > CA type: MT4115 > > Number of ports: 1 > > Firmware version: 12.17.2020 > > Hardware version: 0 > > Node GUID: 0x248a0703005abb30 > > System image GUID: 0x248a0703005abb30 > > Port 1: > > State: Down > > Physical state: Disabled > > Rate: 100 > > Base lid: 0 > > LMC: 0 > > SM lid: 0 > > Capability mask: 0x3c010000 > > Port GUID: 0x268a07fffe5abb30 > > Link layer: Ethernet > > CA 'mlx5_3' > > CA type: MT4115 > > Number of ports: 1 > > Firmware version: 12.17.2020 > > Hardware version: 0 > > Node GUID: 0x248a0703005abb31 > > System image GUID: 0x248a0703005abb30 > > Port 1: > > State: Down > > Physical state: Disabled > > Rate: 100 > > Base lid: 0 > > LMC: 0 > > SM lid: 0 > > Capability mask: 0x3c010000 > > Port GUID: 0x268a07fffe5abb31 > > Link layer: Ethernet > > -bash-4.1$ > > %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% > %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% > > > > On Fri, Jul 14, 2017 at 12:37 AM, John Hearns via users > > <users@lists.open-mpi.org> wrote: > >> > >> ABoris, as Gilles says - first do som elower level checkouts of your > >> Infiniband network. > >> I suggest running: > >> ibdiagnet > >> ibhosts > >> and then as Gilles says 'ibstat' on each node > >> > >> > >> > >> On 14 July 2017 at 03:58, Gilles Gouaillardet <gil...@rist.or.jp> > wrote: > >>> > >>> Boris, > >>> > >>> > >>> Open MPI should automatically detect the infiniband hardware, and use > >>> openib (and *not* tcp) for inter node communications > >>> > >>> and a shared memory optimized btl (e.g. sm or vader) for intra node > >>> communications. > >>> > >>> > >>> note if you "-mca btl openib,self", you tell Open MPI to use the openib > >>> btl between any tasks, > >>> > >>> including tasks running on the same node (which is less efficient than > >>> using sm or vader) > >>> > >>> > >>> at first, i suggest you make sure infiniband is up and running on all > >>> your nodes. > >>> > >>> (just run ibstat, at least one port should be listed, state should be > >>> Active, and all nodes should have the same SM lid) > >>> > >>> > >>> then try to run two tasks on two nodes. > >>> > >>> > >>> if this does not work, you can > >>> > >>> mpirun --mca btl_base_verbose 100 ... > >>> > >>> and post the logs so we can investigate from there. > >>> > >>> > >>> Cheers, > >>> > >>> > >>> Gilles > >>> > >>> > >>> > >>> On 7/14/2017 6:43 AM, Boris M. Vulovic wrote: > >>>> > >>>> > >>>> I would like to know how to invoke InfiniBand hardware on CentOS 6x > >>>> cluster with OpenMPI (static libs.) for running my C++ code. This is > how I > >>>> compile and run: > >>>> > >>>> /usr/local/open-mpi/1.10.7/bin/mpic++ -L/usr/local/open-mpi/1.10.7/ > lib > >>>> -Bstatic main.cpp -o DoWork > >>>> > >>>> usr/local/open-mpi/1.10.7/bin/mpiexec -mca btl tcp,self --hostfile > >>>> hostfile5 -host node01,node02,node03,node04,node05 -n 200 DoWork > >>>> > >>>> Here, "*-mca btl tcp,self*" reveals that *TCP* is used, and the > cluster > >>>> has InfiniBand. > >>>> > >>>> What should be changed in compiling and running commands for > InfiniBand > >>>> to be invoked? If I just replace "*-mca btl tcp,self*" with "*-mca btl > >>>> openib,self*" then I get plenty of errors with relevant one saying: > >>>> > >>>> /At least one pair of MPI processes are unable to reach each other for > >>>> MPI communications. This means that no Open MPI device has indicated > that it > >>>> can be used to communicate between these processes. This is an error; > Open > >>>> MPI requires that all MPI processes be able to reach each other. This > error > >>>> can sometimes be the result of forgetting to specify the "self" BTL./ > >>>> > >>>> Thanks very much!!! > >>>> > >>>> > >>>> *Boris * > >>>> > >>>> > >>>> > >>>> > >>>> _______________________________________________ > >>>> users mailing list > >>>> users@lists.open-mpi.org > >>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users > >>> > >>> > >>> _______________________________________________ > >>> users mailing list > >>> users@lists.open-mpi.org > >>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users > >> > >> > >> > >> _______________________________________________ > >> users mailing list > >> users@lists.open-mpi.org > >> https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > > > > > > > > > -- > > > > Boris M. Vulovic > > > > > > > > _______________________________________________ > > users mailing list > > users@lists.open-mpi.org > > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > -- *Boris M. Vulovic*
_______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users