If I might interject here before lots of time is wasted. Spectrum MPI is an IBM 
-product- and is not free. What you are likely running into is that their 
license manager is blocking you from running, albeit without a really nice 
error message. I’m sure that’s something they are working on.

If you really want to use Spectrum MPI, I suggest you contact them about 
purchasing it.


> On May 19, 2017, at 1:16 AM, Gabriele Fatigati <g.fatig...@cineca.it> wrote:
> 
> Hi Gilles, in attach the outpuf of:
> 
> mpirun --mca btl_base_verbose 100 -np 2 ...
> 
> 2017-05-19 9:43 GMT+02:00 Gilles Gouaillardet <gil...@rist.or.jp 
> <mailto:gil...@rist.or.jp>>:
> Gabriele,
> 
> 
> can you
> 
> mpirun --mca btl_base_verbose 100 -np 2 ...
> 
> 
> so we can figure out why nor sm nor vader is used ?
> 
> 
> Cheers,
> 
> 
> Gilles
> 
> 
> 
> On 5/19/2017 4:23 PM, Gabriele Fatigati wrote:
> Oh no, by using two procs:
> 
> 
> findActiveDevices Error
> We found no active IB device ports
> findActiveDevices Error
> We found no active IB device ports
> --------------------------------------------------------------------------
> At least one pair of MPI processes are unable to reach each other for
> MPI communications.  This means that no Open MPI device has indicated
> that it can be used to communicate between these processes.  This is
> an error; Open MPI requires that all MPI processes be able to reach
> each other.  This error can sometimes be the result of forgetting to
> specify the "self" BTL.
> 
>   Process 1 ([[12380,1],0]) is on host: openpower
>   Process 2 ([[12380,1],1]) is on host: openpower
>   BTLs attempted: self
> 
> Your MPI job is now going to abort; sorry.
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> --------------------------------------------------------------------------
> MPI_INIT has failed because at least one MPI process is unreachable
> from another.  This *usually* means that an underlying communication
> plugin -- such as a BTL or an MTL -- has either not loaded or not
> allowed itself to be used.  Your MPI job will now abort.
> 
> You may wish to try to narrow down the problem;
>  * Check the output of ompi_info to see which BTL/MTL plugins are
>    available.
>  * Run your application with MPI_THREAD_SINGLE.
>  * Set the MCA parameter btl_base_verbose to 100 (or mtl_base_verbose,
>    if using MTL-based communications) to see exactly which
>    communication plugins were considered and/or discarded.
> --------------------------------------------------------------------------
> [openpower:88867] 1 more process has sent help message help-mca-bml-r2.txt / 
> unreachable proc
> [openpower:88867] Set MCA parameter "orte_base_help_aggregate" to 0 to see 
> all help / error messages
> [openpower:88867] 1 more process has sent help message help-mpi-runtime.txt / 
> mpi_init:startup:pml-add-procs-fail
> 
> 
> 
> 
> 
> 2017-05-19 9:22 GMT+02:00 Gabriele Fatigati <g.fatig...@cineca.it 
> <mailto:g.fatig...@cineca.it> <mailto:g.fatig...@cineca.it 
> <mailto:g.fatig...@cineca.it>>>:
> 
>     Hi GIlles,
> 
>     using your command with one MPI procs I get:
> 
>     findActiveDevices Error
>     We found no active IB device ports
>     Hello world from rank 0  out of 1 processors
> 
>     So it seems to work apart the error message.
> 
> 
>     2017-05-19 9:10 GMT+02:00 Gilles Gouaillardet <gil...@rist.or.jp 
> <mailto:gil...@rist.or.jp>
>     <mailto:gil...@rist.or.jp <mailto:gil...@rist.or.jp>>>:
> 
>         Gabriele,
> 
> 
>         so it seems pml/pami assumes there is an infiniband card
>         available (!)
> 
>         i guess IBM folks will comment on that shortly.
> 
> 
>         meanwhile, you do not need pami since you are running on a
>         single node
> 
>         mpirun --mca pml ^pami ...
> 
>         should do the trick
> 
>         (if it does not work, can run and post the logs)
> 
>         mpirun --mca pml ^pami --mca pml_base_verbose 100 ...
> 
> 
>         Cheers,
> 
> 
>         Gilles
> 
> 
>         On 5/19/2017 4:01 PM, Gabriele Fatigati wrote:
> 
>             Hi John,
>             Infiniband is not used, there is a single node on this
>             machine.
> 
>             2017-05-19 8:50 GMT+02:00 John Hearns via users
>             <users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>>
>             <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>>>>:
> 
>                 Gabriele,   pleae run  'ibv_devinfo'
>                 It looks to me like you may have the physical
>             interface cards in
>                 these systems, but you do not have the correct drivers or
>                 libraries loaded.
> 
>                 I have had similar messages when using Infiniband on
>             x86 systems -
>                 which did not have libibverbs installed.
> 
> 
>                 On 19 May 2017 at 08:41, Gabriele Fatigati
>             <g.fatig...@cineca.it <mailto:g.fatig...@cineca.it> 
> <mailto:g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>>
>                 <mailto:g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>
>             <mailto:g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>>>> 
> wrote:
> 
>                     Hi Gilles, using your command:
> 
>                     [openpower:88536] mca: base: components_register:
>             registering
>                     framework pml components
>                     [openpower:88536] mca: base: components_register:
>             found loaded
>                     component pami
>                     [openpower:88536] mca: base: components_register:
>             component
>                     pami register function successful
>                     [openpower:88536] mca: base: components_open:
>             opening pml
>                     components
>                     [openpower:88536] mca: base: components_open:
>             found loaded
>                     component pami
>                     [openpower:88536] mca: base: components_open:
>             component pami
>                     open function successful
>                     [openpower:88536] select: initializing pml
>             component pami
>                     findActiveDevices Error
>                     We found no active IB device ports
>                     [openpower:88536] select: init returned failure
>             for component pami
>                     [openpower:88536] PML pami cannot be selected
>                                
> --------------------------------------------------------------------------
>                     No components were able to be opened in the pml
>             framework.
> 
>                     This typically means that either no components of
>             this type were
>                     installed, or none of the installed componnets can
>             be loaded.
>                     Sometimes this means that shared libraries
>             required by these
>                     components are unable to be found/loaded.
> 
>                       Host:      openpower
>                       Framework: pml
>                                
> --------------------------------------------------------------------------
> 
> 
>                     2017-05-19 7:03 GMT+02:00 Gilles Gouaillardet
>                     <gil...@rist.or.jp <mailto:gil...@rist.or.jp> 
> <mailto:gil...@rist.or.jp <mailto:gil...@rist.or.jp>>
>             <mailto:gil...@rist.or.jp <mailto:gil...@rist.or.jp> 
> <mailto:gil...@rist.or.jp <mailto:gil...@rist.or.jp>>>>:
> 
> 
>                         Gabriele,
> 
> 
>                         pml/pami is here, at least according to ompi_info
> 
> 
>                         can you update your mpirun command like this
> 
>                         mpirun --mca pml_base_verbose 100 ..
> 
> 
>                         and post the output ?
> 
> 
>                         Cheers,
> 
>                         Gilles
> 
>                         On 5/18/2017 10:41 PM, Gabriele Fatigati wrote:
> 
>                             Hi Gilles, attached the requested info
> 
>                             2017-05-18 15:04 GMT+02:00 Gilles Gouaillardet
>                             <gilles.gouaillar...@gmail.com 
> <mailto:gilles.gouaillar...@gmail.com>
>             <mailto:gilles.gouaillar...@gmail.com 
> <mailto:gilles.gouaillar...@gmail.com>>
>                             <mailto:gilles.gouaillar...@gmail.com 
> <mailto:gilles.gouaillar...@gmail.com>
>             <mailto:gilles.gouaillar...@gmail.com 
> <mailto:gilles.gouaillar...@gmail.com>>>
>                             <mailto:gilles.gouaillar...@gmail.com 
> <mailto:gilles.gouaillar...@gmail.com>
>             <mailto:gilles.gouaillar...@gmail.com 
> <mailto:gilles.gouaillar...@gmail.com>>
>                             <mailto:gilles.gouaillar...@gmail.com 
> <mailto:gilles.gouaillar...@gmail.com>
>             <mailto:gilles.gouaillar...@gmail.com 
> <mailto:gilles.gouaillar...@gmail.com>>>>>:
> 
>                                 Gabriele,
> 
>                                 can you
>                                 ompi_info --all | grep pml
> 
>                                 also, make sure there is nothing in your
>                             environment pointing to
>                                 an other Open MPI install
>                                 for example
>                                 ldd a.out
>                                 should only point to IBM libraries
> 
>                                 Cheers,
> 
>                                 Gilles
> 
> 
>                                 On Thursday, May 18, 2017, Gabriele
>             Fatigati
>                             <g.fatig...@cineca.it 
> <mailto:g.fatig...@cineca.it>
>             <mailto:g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>> 
> <mailto:g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>
>             <mailto:g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>>>
>                                 <mailto:g.fatig...@cineca.it 
> <mailto:g.fatig...@cineca.it>
>             <mailto:g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>>
> 
>                             <mailto:g.fatig...@cineca.it 
> <mailto:g.fatig...@cineca.it>
>             <mailto:g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>>>>> 
> wrote:
> 
>                                     Dear OpenMPI users and developers,
>             I'm using
>                             IBM Spectrum MPI
>                                     10.1.0 based on OpenMPI, so I hope
>             there are
>                             some MPI expert
>                                     can help me to solve the problem.
> 
>                                     When I run a simple Hello World
>             MPI program, I
>                             get the follow
>                                     error message:
> 
> 
>                                     A requested component was not
>             found, or was
>                             unable to be
>                                     opened.  This
>                                     means that this component is
>             either not
>                             installed or is unable
>                                     to be
>                                     used on your system (e.g.,
>             sometimes this
>                             means that shared
>                                     libraries
>                                     that the component requires are
>             unable to be
>                             found/loaded).         Note that
>                                     Open MPI stopped checking at the first
>                             component that it did
>                                     not find.
> 
>                                     Host:      openpower
>                                     Framework: pml
>                                     Component: pami
>             
> --------------------------------------------------------------------------
>             
> --------------------------------------------------------------------------
>                                     It looks like MPI_INIT failed for
>             some reason;
>                             your parallel
>                                     process is
>                                     likely to abort. There are many
>             reasons that a
>                             parallel
>                                     process can
>                                     fail during MPI_INIT; some of
>             which are due to
>                             configuration
>                                     or environment
>                                     problems.  This failure appears to
>             be an
>                             internal failure;
>                                     here's some
>                                     additional information (which may
>             only be
>                             relevant to an Open MPI
>                                     developer):
> 
>                                     mca_pml_base_open() failed
>                                       --> Returned "Not found" (-13)
>             instead of
>                             "Success" (0)
>             
> --------------------------------------------------------------------------
>                                     *** An error occurred in MPI_Init
>                                     *** on a NULL communicator
>                                     *** MPI_ERRORS_ARE_FATAL
>             (processes in this
>                             communicator will
>                                     now abort,
>                                     ***    and potentially your MPI job)
> 
>                                     My sysadmin used official IBM Spectrum
>                             packages to install
>                                     MPI, so It's quite strange that
>             there are some
>                             components
>                                     missing (pami). Any help? Thanks
> 
> 
>                                     --         Ing. Gabriele Fatigati
> 
>                                     HPC specialist
> 
>                                     SuperComputing Applications and
>             Innovation
>                             Department
> 
>                                     Via Magnanelli 6/3, Casalecchio di
>             Reno (BO) Italy
> 
>             www.cineca.it <http://www.cineca.it/> <http://www.cineca.it 
> <http://www.cineca.it/>> <http://www.cineca.it <http://www.cineca.it/>>
>                             <http://www.cineca.it <http://www.cineca.it/>>    
>          Tel: +39
>             051 6171722 <tel:051%206171722> <tel:051%206171722> 
> <tel:051%206171722>
> 
>             <tel:051%20617%201722>
> 
>                                     g.fatigati [AT] cineca.it 
> <http://cineca.it/>
>             <http://cineca.it <http://cineca.it/>> <http://cineca.it 
> <http://cineca.it/>>
>                             <http://cineca.it <http://cineca.it/>>
> 
> 
>             _______________________________________________
>                                 users mailing list
>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 
> <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>>
>                             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>
>             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>>>
>                             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>
>             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>>
>                             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>
>             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>>>>
>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>                                        
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>>
>                                                           
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>                                        
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>>>
> 
> 
> 
> 
>                             --                 Ing. Gabriele Fatigati
> 
>                             HPC specialist
> 
>                             SuperComputing Applications and Innovation
>             Department
> 
>                             Via Magnanelli 6/3, Casalecchio di Reno
>             (BO) Italy
> 
>             www.cineca.it <http://www.cineca.it/> <http://www.cineca.it 
> <http://www.cineca.it/>> <http://www.cineca.it <http://www.cineca.it/>>
>                             <http://www.cineca.it <http://www.cineca.it/>> 
> Tel: +39 051
>             6171722 <tel:%2B39%20051%206171722>
>             <tel:%2B39%20051%206171722>
> 
>                             g.fatigati [AT] cineca.it <http://cineca.it/>
>             <http://cineca.it <http://cineca.it/>> <http://cineca.it 
> <http://cineca.it/>>
>                             <http://cineca.it <http://cineca.it/>>
> 
> 
>                                        
> _______________________________________________
>                             users mailing list
>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 
> <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>>
>             <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>>>
>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>                                        
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>>
> 
> 
> 
> 
> 
>                     --         Ing. Gabriele Fatigati
> 
>                     HPC specialist
> 
>                     SuperComputing Applications and Innovation Department
> 
>                     Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
> 
>             www.cineca.it <http://www.cineca.it/> <http://www.cineca.it 
> <http://www.cineca.it/>>
>             <http://www.cineca.it <http://www.cineca.it/>>                  
> Tel:
>             +39 051 6171722 <tel:%2B39%20051%206171722> 
> <tel:%2B39%20051%206171722>
>             <tel:+39%20051%20617%201722>
> 
> 
>                     g.fatigati [AT] cineca.it <http://cineca.it/> 
> <http://cineca.it <http://cineca.it/>>
>             <http://cineca.it <http://cineca.it/>>
> 
>                     _______________________________________________
>                     users mailing list
>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 
> <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>>
>             <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>>>
>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>                                
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>>
> 
> 
> 
>                 _______________________________________________
>                 users mailing list
>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 
> <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>>
>             <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>             <mailto:users@lists.open-mpi.org 
> <mailto:users@lists.open-mpi.org>>>
>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>                            
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>>
> 
> 
> 
> 
>             --             Ing. Gabriele Fatigati
> 
>             HPC specialist
> 
>             SuperComputing Applications and Innovation Department
> 
>             Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
> 
>             www.cineca.it <http://www.cineca.it/> <http://www.cineca.it 
> <http://www.cineca.it/>>
>             <http://www.cineca.it <http://www.cineca.it/>> Tel: +39 051 
> 6171722 <tel:%2B39%20051%206171722>
>             <tel:%2B39%20051%206171722>
> 
>             g.fatigati [AT] cineca.it <http://cineca.it/> <http://cineca.it 
> <http://cineca.it/>>
>             <http://cineca.it <http://cineca.it/>>
> 
> 
>             _______________________________________________
>             users mailing list
>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 
> <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>>
>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
> 
> 
>         _______________________________________________
>         users mailing list
>         users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 
> <mailto:users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>>
>         https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>         <https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
> 
> 
> 
> 
>     --     Ing. Gabriele Fatigati
> 
>     HPC specialist
> 
>     SuperComputing Applications and Innovation Department
> 
>     Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
> 
>     www.cineca.it <http://www.cineca.it/> <http://www.cineca.it 
> <http://www.cineca.it/>>       Tel: +39 051 6171722 
> <tel:%2B39%20051%206171722>
>     <tel:051%20617%201722>
> 
>     g.fatigati [AT] cineca.it <http://cineca.it/> <http://cineca.it 
> <http://cineca.it/>>
> 
> 
> 
> 
> -- 
> Ing. Gabriele Fatigati
> 
> HPC specialist
> 
> SuperComputing Applications and Innovation Department
> 
> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
> 
> www.cineca.it <http://www.cineca.it/> <http://www.cineca.it 
> <http://www.cineca.it/>> Tel:   +39 051 6171722 <tel:%2B39%20051%206171722>
> 
> g.fatigati [AT] cineca.it <http://cineca.it/> <http://cineca.it 
> <http://cineca.it/>>
> 
> 
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
> 
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
> <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
> 
> 
> 
> -- 
> Ing. Gabriele Fatigati
> 
> HPC specialist
> 
> SuperComputing Applications and Innovation Department
> 
> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
> 
> www.cineca.it <http://www.cineca.it/>                    Tel:   +39 051 
> 6171722
> 
> g.fatigati [AT] cineca.it <http://cineca.it/>          
> <output_mpirun>_______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to