The openib btl should be able to run alongside cm/mxm or yalla. If I
have time this weekend I will get on the mustang and see what the
problem is. The best answer is to change the openmpi-mca-params.conf in
the install to have pml = ob1. I have seen little to no benefit with
using MXM on mustang. In fact, the default configuration (which uses UD)
gets terrible bandwidth.

-Nathan

On Thu, Apr 21, 2016 at 01:48:46PM +0300, Alina Sklarevich wrote:
>    David, thanks for the info you provided.
>    I will try to dig in further to see what might be causing this issue.
>    In the meantime, maybe Nathan can please comment about the openib btl
>    behavior here?
>    Thanks,
>    Alina.
>    On Wed, Apr 20, 2016 at 8:01 PM, David Shrader <dshra...@lanl.gov> wrote:
> 
>      Hello Alina,
> 
>      Thank you for the information about how the pml components work. I knew
>      that the other components were being opened and ultimately closed in
>      favor of yalla, but I didn't realize that initial open would cause a
>      persistent change in the ompi runtime.
> 
>      Here's the information you requested about the ib network:
> 
>      - MOFED version:
>      We are using the Open Fabrics Software as bundled by RedHat, and my ib
>      network folks say we're running something close to v1.5.4
>      - ibv_devinfo:
>      [dshrader@mu0001 examples]$ ibv_devinfo
>      hca_id: mlx4_0
>              transport:                      InfiniBand (0)
>              fw_ver:                         2.9.1000
>              node_guid:                      0025:90ff:ff16:78d8
>              sys_image_guid:                 0025:90ff:ff16:78db
>              vendor_id:                      0x02c9
>              vendor_part_id:                 26428
>              hw_ver:                         0xB0
>              board_id:                       SM_2121000001000
>              phys_port_cnt:                  1
>                      port:   1
>                              state:                  PORT_ACTIVE (4)
>                              max_mtu:                4096 (5)
>                              active_mtu:             4096 (5)
>                              sm_lid:                 250
>                              port_lid:               366
>                              port_lmc:               0x00
>                              link_layer:             InfiniBand
> 
>      I still get the seg fault when specifying the hca:
> 
>      $> mpirun -n 1 -mca btl_openib_receive_queues
>      X,4096,1024:X,12288,512:X,65536,512 -mca btl_openib_if_include mlx4_0
>      ./hello_c.x
>      Hello, world, I am 0 of 1, (Open MPI v1.10.2, package: Open MPI
>      dshra...@mu-fey.lanl.gov Distribution, ident: 1.10.2, repo rev:
>      v1.10.1-145-g799148f, Jan 21, 2016, 135)
>      
> --------------------------------------------------------------------------
>      mpirun noticed that process rank 0 with PID 10045 on node mu0001 exited
>      on signal 11 (Segmentation fault).
>      
> --------------------------------------------------------------------------
> 
>      I don't know if this helps, but the first time I tried the command I
>      mistyped the hca name. This got me a warning, but no seg fault:
> 
>      $> mpirun -n 1 -mca btl_openib_receive_queues
>      X,4096,1024:X,12288,512:X,65536,512 -mca btl_openib_if_include ml4_0
>      ./hello_c.x
>      
> --------------------------------------------------------------------------
>      WARNING: One or more nonexistent OpenFabrics devices/ports were
>      specified:
> 
>        Host:                 mu0001
>        MCA parameter:        mca_btl_if_include
>        Nonexistent entities: ml4_0
> 
>      These entities will be ignored.  You can disable this warning by
>      setting the btl_openib_warn_nonexistent_if MCA parameter to 0.
>      
> --------------------------------------------------------------------------
>      Hello, world, I am 0 of 1, (Open MPI v1.10.2, package: Open MPI
>      dshra...@mu-fey.lanl.gov Distribution, ident: 1.10.2, repo rev:
>      v1.10.1-145-g799148f, Jan 21, 2016, 135)
> 
>      So, telling the openib btl to use the actual hca didn't get the seg
>      fault to go away, but giving it a dummy value did.
> 
>      Thanks,
>      David
> 
>      On 04/20/2016 08:13 AM, Alina Sklarevich wrote:
> 
>        Hi David,
>        I was able to reproduce the issue you reported. 
>        When the command line doesn't specify the components to use, ompi will
>        try to load/open all the ones available (and close them in the end)
>        and then choose the components according to their priority and whether
>        or not they were opened successfully.
>        This means that even if pml yalla was the one running, other
>        components were opened and closed as well.
>        The parameter you are using - btl_openib_receive_queues, doesn't have
>        an effect on pml yalla. It only affects the openib btl which is used
>        by pml ob1.
>        Using the verbosity of btl_base_verbose I see that when the
>        segmentation fault happens, the code doesn't reach the phase of
>        unloading the openib btl so perhaps the problem originates there
>        (since pml yalla was already unloaded).
>        Can you please try adding this mca parameter to your command line to
>        specify the HCA you are using?
>        -mca btl_openib_if_include <hca>
>        It made the segv go away for me.
>        Can you please attach the output of ibv_devinfo and write the MOFED
>        version you are using?
>        Thank you,
>        Alina.
>        On Wed, Apr 20, 2016 at 2:53 PM, Joshua Ladd <jladd.m...@gmail.com>
>        wrote:
> 
>          Hi, David
> 
>          We are looking into your report.
> 
>          Best,
> 
>          Josh
>          On Tue, Apr 19, 2016 at 4:41 PM, David Shrader <dshra...@lanl.gov>
>          wrote:
> 
>            Hello,
> 
>            I have been investigating using XRC on a cluster with a mellanox
>            interconnect. I have found that in a certain situation I get a seg
>            fault. I am using 1.10.2 compiled with gcc 5.3.0, and the simplest
>            configure line that I have found that still results in the seg
>            fault is as follows:
> 
>            $> ./configure --with-hcoll --with-mxm --prefix=...
> 
>            I do have mxm 3.4.3065 and hcoll 3.3.768 installed in to system
>            space (/usr/lib64). If I use '--without-hcoll --without-mxm,' the
>            seg fault does not happen.
> 
>            The seg fault happens even when using examples/hello_c.c, so here
>            is an example of the seg fault using it:
> 
>            $> mpicc hello_c.c -o hello_c.x
>            $> mpirun -n 1 ./hello_c.x
>            Hello, world, I am 0 of 1, (Open MPI v1.10.2, package: Open MPI
>            dshra...@mu-fey.lanl.gov Distribution, ident: 1.10.2, repo rev:
>            v1.10.1-145-g799148f, Jan 21, 2016, 135)
>            $> mpirun -n 1 -mca btl_openib_receive_queues
>            X,4096,1024:X,12288,512:X,65536,512
>            Hello, world, I am 0 of 1, (Open MPI v1.10.2, package: Open MPI
>            dshra...@mu-fey.lanl.gov Distribution, ident: 1.10.2, repo rev:
>            v1.10.1-145-g799148f, Jan 21, 2016, 135)
>            
> --------------------------------------------------------------------------
>            mpirun noticed that process rank 0 with PID 22819 on node mu0001
>            exited on signal 11 (Segmentation fault).
>            
> --------------------------------------------------------------------------
> 
>            The seg fault happens no matter the number of ranks. I have tried
>            the above command with '-mca pml_base_verbose,' and it shows that
>            I am using the yalla pml:
> 
>            $> mpirun -n 1 -mca btl_openib_receive_queues
>            X,4096,1024:X,12288,512:X,65536,512 -mca pml_base_verbose 100
>            ./hello_c.x
>            ...output snipped...
>            [mu0001.localdomain:22825] select: component cm not selected /
>            finalized
>            [mu0001.localdomain:22825] select: component ob1 not selected /
>            finalized
>            [mu0001.localdomain:22825] select: component yalla selected
>            ...output snipped...
>            
> --------------------------------------------------------------------------
>            mpirun noticed that process rank 0 with PID 22825 on node mu0001
>            exited on signal 11 (Segmentation fault).
>            
> --------------------------------------------------------------------------
> 
>            Interestingly enough, if I tell mpirun what pml to use, the seg
>            fault goes away. The following command does not get the seg fault:
> 
>            $> mpirun -n 1 -mca btl_openib_receive_queues
>            X,4096,1024:X,12288,512:X,65536,512 -mca pml yalla ./hello_c.x
> 
>            Passing either ob1 or cm to '-mca pml' also works. So it seems
>            that the seg fault comes about when the yalla pml is chosen by
>            default, when mxm/hcoll is involved, and using XRC. I'm not sure
>            if mxm is to blame, however, as using '-mca pml cm -mca mtl mxm'
>            with the XRC parameters doesn't throw the seg fault.
> 
>            Other information...
>            OS: RHEL 6.7-based (TOSS)
>            OpenFabrics: RedHat provided
>            Kernel: 2.6.32-573.8.1.2chaos.ch5.4.x86_64
>            Config.log and 'ompi_info --all' are in the tarball ompi.tar.bz2
>            which is attached.
> 
>            Is there something else I should be doing with the yalla pml when
>            using XRC? Regardless, I hope reporting the seg fault is useful.
> 
>            Thanks,
>            David
> 
>            --
>            David Shrader
>            HPC-ENV High Performance Computer Systems
>            Los Alamos National Lab
>            Email: dshrader <at> lanl.gov
> 
>            _______________________________________________
>            devel mailing list
>            de...@open-mpi.org
>            Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>            Link to this post:
>            http://www.open-mpi.org/community/lists/devel/2016/04/18786.php
> 
>          _______________________________________________
>          devel mailing list
>          de...@open-mpi.org
>          Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>          Link to this post:
>          http://www.open-mpi.org/community/lists/devel/2016/04/18788.php
> 
>  _______________________________________________
>  devel mailing list
>  de...@open-mpi.org
>  Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>  Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2016/04/18789.php
> 
>  --
>  David Shrader
>  HPC-ENV High Performance Computer Systems
>  Los Alamos National Lab
>  Email: dshrader <at> lanl.gov
> 
>      _______________________________________________
>      devel mailing list
>      de...@open-mpi.org
>      Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>      Link to this post:
>      http://www.open-mpi.org/community/lists/devel/2016/04/18793.php

> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2016/04/18801.php

Attachment: pgpFswJzLCfh6.pgp
Description: PGP signature

Reply via email to