Yes, using " -pami_noib" solve the problem, I lost the previous message.
Thanks you so much for the support.

2017-05-19 11:12 GMT+02:00 John Hearns via users <users@lists.open-mpi.org>:

> I am not sure I agree with that.
> (a) the original error message from Gabriele was quite clear - the MPI
> could not find an interface card which was up, so it would not run.
> (b) Nysal actually pointed out the solution which looks good - after
> reaidng the documentation.. use  pami_noib
> (c) Having discussions like this helps us all to learn. I have made many
> stupid replies on this list, and looking at problems like this has helped
> me to learn.
>
>
>
>
> On 19 May 2017 at 11:01, r...@open-mpi.org <r...@open-mpi.org> wrote:
>
>> If I might interject here before lots of time is wasted. Spectrum MPI is
>> an IBM -product- and is not free. What you are likely running into is that
>> their license manager is blocking you from running, albeit without a really
>> nice error message. I’m sure that’s something they are working on.
>>
>> If you really want to use Spectrum MPI, I suggest you contact them about
>> purchasing it.
>>
>>
>> On May 19, 2017, at 1:16 AM, Gabriele Fatigati <g.fatig...@cineca.it>
>> wrote:
>>
>> Hi Gilles, in attach the outpuf of:
>>
>> mpirun --mca btl_base_verbose 100 -np 2 ...
>>
>> 2017-05-19 9:43 GMT+02:00 Gilles Gouaillardet <gil...@rist.or.jp>:
>>
>>> Gabriele,
>>>
>>>
>>> can you
>>>
>>> mpirun --mca btl_base_verbose 100 -np 2 ...
>>>
>>>
>>> so we can figure out why nor sm nor vader is used ?
>>>
>>>
>>> Cheers,
>>>
>>>
>>> Gilles
>>>
>>>
>>>
>>> On 5/19/2017 4:23 PM, Gabriele Fatigati wrote:
>>>
>>>> Oh no, by using two procs:
>>>>
>>>>
>>>> findActiveDevices Error
>>>> We found no active IB device ports
>>>> findActiveDevices Error
>>>> We found no active IB device ports
>>>> ------------------------------------------------------------
>>>> --------------
>>>> At least one pair of MPI processes are unable to reach each other for
>>>> MPI communications.  This means that no Open MPI device has indicated
>>>> that it can be used to communicate between these processes.  This is
>>>> an error; Open MPI requires that all MPI processes be able to reach
>>>> each other.  This error can sometimes be the result of forgetting to
>>>> specify the "self" BTL.
>>>>
>>>>   Process 1 ([[12380,1],0]) is on host: openpower
>>>>   Process 2 ([[12380,1],1]) is on host: openpower
>>>>   BTLs attempted: self
>>>>
>>>> Your MPI job is now going to abort; sorry.
>>>> ------------------------------------------------------------
>>>> --------------
>>>> *** An error occurred in MPI_Init
>>>> *** on a NULL communicator
>>>> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
>>>> ***    and potentially your MPI job)
>>>> *** An error occurred in MPI_Init
>>>> *** on a NULL communicator
>>>> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
>>>> ***    and potentially your MPI job)
>>>> ------------------------------------------------------------
>>>> --------------
>>>> MPI_INIT has failed because at least one MPI process is unreachable
>>>> from another.  This *usually* means that an underlying communication
>>>> plugin -- such as a BTL or an MTL -- has either not loaded or not
>>>> allowed itself to be used.  Your MPI job will now abort.
>>>>
>>>> You may wish to try to narrow down the problem;
>>>>  * Check the output of ompi_info to see which BTL/MTL plugins are
>>>>    available.
>>>>  * Run your application with MPI_THREAD_SINGLE.
>>>>  * Set the MCA parameter btl_base_verbose to 100 (or mtl_base_verbose,
>>>>    if using MTL-based communications) to see exactly which
>>>>    communication plugins were considered and/or discarded.
>>>> ------------------------------------------------------------
>>>> --------------
>>>> [openpower:88867] 1 more process has sent help message
>>>> help-mca-bml-r2.txt / unreachable proc
>>>> [openpower:88867] Set MCA parameter "orte_base_help_aggregate" to 0 to
>>>> see all help / error messages
>>>> [openpower:88867] 1 more process has sent help message
>>>> help-mpi-runtime.txt / mpi_init:startup:pml-add-procs-fail
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> 2017-05-19 9:22 GMT+02:00 Gabriele Fatigati <g.fatig...@cineca.it
>>>> <mailto:g.fatig...@cineca.it>>:
>>>>
>>>>     Hi GIlles,
>>>>
>>>>     using your command with one MPI procs I get:
>>>>
>>>>     findActiveDevices Error
>>>>     We found no active IB device ports
>>>>     Hello world from rank 0  out of 1 processors
>>>>
>>>>     So it seems to work apart the error message.
>>>>
>>>>
>>>>     2017-05-19 9:10 GMT+02:00 Gilles Gouaillardet <gil...@rist.or.jp
>>>>     <mailto:gil...@rist.or.jp>>:
>>>>
>>>>         Gabriele,
>>>>
>>>>
>>>>         so it seems pml/pami assumes there is an infiniband card
>>>>         available (!)
>>>>
>>>>         i guess IBM folks will comment on that shortly.
>>>>
>>>>
>>>>         meanwhile, you do not need pami since you are running on a
>>>>         single node
>>>>
>>>>         mpirun --mca pml ^pami ...
>>>>
>>>>         should do the trick
>>>>
>>>>         (if it does not work, can run and post the logs)
>>>>
>>>>         mpirun --mca pml ^pami --mca pml_base_verbose 100 ...
>>>>
>>>>
>>>>         Cheers,
>>>>
>>>>
>>>>         Gilles
>>>>
>>>>
>>>>         On 5/19/2017 4:01 PM, Gabriele Fatigati wrote:
>>>>
>>>>             Hi John,
>>>>             Infiniband is not used, there is a single node on this
>>>>             machine.
>>>>
>>>>             2017-05-19 8:50 GMT+02:00 John Hearns via users
>>>>             <users@lists.open-mpi.org
>>>>             <mailto:users@lists.open-mpi.org>
>>>>             <mailto:users@lists.open-mpi.org
>>>>             <mailto:users@lists.open-mpi.org>>>:
>>>>
>>>>                 Gabriele,   pleae run  'ibv_devinfo'
>>>>                 It looks to me like you may have the physical
>>>>             interface cards in
>>>>                 these systems, but you do not have the correct drivers
>>>> or
>>>>                 libraries loaded.
>>>>
>>>>                 I have had similar messages when using Infiniband on
>>>>             x86 systems -
>>>>                 which did not have libibverbs installed.
>>>>
>>>>
>>>>                 On 19 May 2017 at 08:41, Gabriele Fatigati
>>>>             <g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>
>>>>                 <mailto:g.fatig...@cineca.it
>>>>             <mailto:g.fatig...@cineca.it>>> wrote:
>>>>
>>>>                     Hi Gilles, using your command:
>>>>
>>>>                     [openpower:88536] mca: base: components_register:
>>>>             registering
>>>>                     framework pml components
>>>>                     [openpower:88536] mca: base: components_register:
>>>>             found loaded
>>>>                     component pami
>>>>                     [openpower:88536] mca: base: components_register:
>>>>             component
>>>>                     pami register function successful
>>>>                     [openpower:88536] mca: base: components_open:
>>>>             opening pml
>>>>                     components
>>>>                     [openpower:88536] mca: base: components_open:
>>>>             found loaded
>>>>                     component pami
>>>>                     [openpower:88536] mca: base: components_open:
>>>>             component pami
>>>>                     open function successful
>>>>                     [openpower:88536] select: initializing pml
>>>>             component pami
>>>>                     findActiveDevices Error
>>>>                     We found no active IB device ports
>>>>                     [openpower:88536] select: init returned failure
>>>>             for component pami
>>>>                     [openpower:88536] PML pami cannot be selected
>>>>                                -----------------------------
>>>> ---------------------------------------------
>>>>                     No components were able to be opened in the pml
>>>>             framework.
>>>>
>>>>                     This typically means that either no components of
>>>>             this type were
>>>>                     installed, or none of the installed componnets can
>>>>             be loaded.
>>>>                     Sometimes this means that shared libraries
>>>>             required by these
>>>>                     components are unable to be found/loaded.
>>>>
>>>>                       Host:      openpower
>>>>                       Framework: pml
>>>>                                -----------------------------
>>>> ---------------------------------------------
>>>>
>>>>
>>>>                     2017-05-19 7:03 GMT+02:00 Gilles Gouaillardet
>>>>                     <gil...@rist.or.jp <mailto:gil...@rist.or.jp>
>>>>             <mailto:gil...@rist.or.jp <mailto:gil...@rist.or.jp>>>:
>>>>
>>>>
>>>>                         Gabriele,
>>>>
>>>>
>>>>                         pml/pami is here, at least according to
>>>> ompi_info
>>>>
>>>>
>>>>                         can you update your mpirun command like this
>>>>
>>>>                         mpirun --mca pml_base_verbose 100 ..
>>>>
>>>>
>>>>                         and post the output ?
>>>>
>>>>
>>>>                         Cheers,
>>>>
>>>>                         Gilles
>>>>
>>>>                         On 5/18/2017 10:41 PM, Gabriele Fatigati wrote:
>>>>
>>>>                             Hi Gilles, attached the requested info
>>>>
>>>>                             2017-05-18 15:04 GMT+02:00 Gilles
>>>> Gouaillardet
>>>>                             <gilles.gouaillar...@gmail.com
>>>>             <mailto:gilles.gouaillar...@gmail.com>
>>>>                             <mailto:gilles.gouaillar...@gmail.com
>>>>             <mailto:gilles.gouaillar...@gmail.com>>
>>>>                             <mailto:gilles.gouaillar...@gmail.com
>>>>             <mailto:gilles.gouaillar...@gmail.com>
>>>>                             <mailto:gilles.gouaillar...@gmail.com
>>>>             <mailto:gilles.gouaillar...@gmail.com>>>>:
>>>>
>>>>                                 Gabriele,
>>>>
>>>>                                 can you
>>>>                                 ompi_info --all | grep pml
>>>>
>>>>                                 also, make sure there is nothing in your
>>>>                             environment pointing to
>>>>                                 an other Open MPI install
>>>>                                 for example
>>>>                                 ldd a.out
>>>>                                 should only point to IBM libraries
>>>>
>>>>                                 Cheers,
>>>>
>>>>                                 Gilles
>>>>
>>>>
>>>>                                 On Thursday, May 18, 2017, Gabriele
>>>>             Fatigati
>>>>                             <g.fatig...@cineca.it
>>>>             <mailto:g.fatig...@cineca.it> <mailto:g.fatig...@cineca.it
>>>>             <mailto:g.fatig...@cineca.it>>
>>>>                                 <mailto:g.fatig...@cineca.it
>>>>             <mailto:g.fatig...@cineca.it>
>>>>
>>>>                             <mailto:g.fatig...@cineca.it
>>>>             <mailto:g.fatig...@cineca.it>>>> wrote:
>>>>
>>>>                                     Dear OpenMPI users and developers,
>>>>             I'm using
>>>>                             IBM Spectrum MPI
>>>>                                     10.1.0 based on OpenMPI, so I hope
>>>>             there are
>>>>                             some MPI expert
>>>>                                     can help me to solve the problem.
>>>>
>>>>                                     When I run a simple Hello World
>>>>             MPI program, I
>>>>                             get the follow
>>>>                                     error message:
>>>>
>>>>
>>>>                                     A requested component was not
>>>>             found, or was
>>>>                             unable to be
>>>>                                     opened.  This
>>>>                                     means that this component is
>>>>             either not
>>>>                             installed or is unable
>>>>                                     to be
>>>>                                     used on your system (e.g.,
>>>>             sometimes this
>>>>                             means that shared
>>>>                                     libraries
>>>>                                     that the component requires are
>>>>             unable to be
>>>>                             found/loaded).         Note that
>>>>                                     Open MPI stopped checking at the
>>>> first
>>>>                             component that it did
>>>>                                     not find.
>>>>
>>>>                                     Host:      openpower
>>>>                                     Framework: pml
>>>>                                     Component: pami
>>>>             ------------------------------
>>>> --------------------------------------------
>>>>             ------------------------------
>>>> --------------------------------------------
>>>>                                     It looks like MPI_INIT failed for
>>>>             some reason;
>>>>                             your parallel
>>>>                                     process is
>>>>                                     likely to abort. There are many
>>>>             reasons that a
>>>>                             parallel
>>>>                                     process can
>>>>                                     fail during MPI_INIT; some of
>>>>             which are due to
>>>>                             configuration
>>>>                                     or environment
>>>>                                     problems.  This failure appears to
>>>>             be an
>>>>                             internal failure;
>>>>                                     here's some
>>>>                                     additional information (which may
>>>>             only be
>>>>                             relevant to an Open MPI
>>>>                                     developer):
>>>>
>>>>                                     mca_pml_base_open() failed
>>>>                                       --> Returned "Not found" (-13)
>>>>             instead of
>>>>                             "Success" (0)
>>>>             ------------------------------
>>>> --------------------------------------------
>>>>                                     *** An error occurred in MPI_Init
>>>>                                     *** on a NULL communicator
>>>>                                     *** MPI_ERRORS_ARE_FATAL
>>>>             (processes in this
>>>>                             communicator will
>>>>                                     now abort,
>>>>                                     ***    and potentially your MPI job)
>>>>
>>>>                                     My sysadmin used official IBM
>>>> Spectrum
>>>>                             packages to install
>>>>                                     MPI, so It's quite strange that
>>>>             there are some
>>>>                             components
>>>>                                     missing (pami). Any help? Thanks
>>>>
>>>>
>>>>                                     --         Ing. Gabriele Fatigati
>>>>
>>>>                                     HPC specialist
>>>>
>>>>                                     SuperComputing Applications and
>>>>             Innovation
>>>>                             Department
>>>>
>>>>                                     Via Magnanelli 6/3, Casalecchio di
>>>>             Reno (BO) Italy
>>>>
>>>>             www.cineca.it <http://www.cineca.it> <http://www.cineca.it>
>>>>                             <http://www.cineca.it>             Tel: +39
>>>>             051 6171722 <tel:051%206171722 <051%206171722>> <
>>>> tel:051%206171722 <051%206171722>>
>>>>
>>>>             <tel:051%20617%201722 <051%20617%201722>>
>>>>
>>>>                                     g.fatigati [AT] cineca.it
>>>>             <http://cineca.it> <http://cineca.it>
>>>>                             <http://cineca.it>
>>>>
>>>>
>>>>             _______________________________________________
>>>>                                 users mailing list
>>>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>>>                             <mailto:users@lists.open-mpi.org
>>>>             <mailto:users@lists.open-mpi.org>>
>>>>                             <mailto:users@lists.open-mpi.org
>>>>             <mailto:users@lists.open-mpi.org>
>>>>                             <mailto:users@lists.open-mpi.org
>>>>             <mailto:users@lists.open-mpi.org>>>
>>>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >
>>>>                                        <https://rfd.newmexicoconsort
>>>> ium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >>
>>>>                                                           <
>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >
>>>>                                        <https://rfd.newmexicoconsort
>>>> ium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >>>
>>>>
>>>>
>>>>
>>>>
>>>>                             --                 Ing. Gabriele Fatigati
>>>>
>>>>                             HPC specialist
>>>>
>>>>                             SuperComputing Applications and Innovation
>>>>             Department
>>>>
>>>>                             Via Magnanelli 6/3, Casalecchio di Reno
>>>>             (BO) Italy
>>>>
>>>>             www.cineca.it <http://www.cineca.it> <http://www.cineca.it>
>>>>                             <http://www.cineca.it> Tel: +39 051
>>>>             6171722 <tel:%2B39%20051%206171722 <%2B39%20051%206171722>>
>>>>             <tel:%2B39%20051%206171722 <%2B39%20051%206171722>>
>>>>
>>>>                             g.fatigati [AT] cineca.it
>>>>             <http://cineca.it> <http://cineca.it>
>>>>                             <http://cineca.it>
>>>>
>>>>
>>>>                                        _____________________________
>>>> __________________
>>>>                             users mailing list
>>>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>>>             <mailto:users@lists.open-mpi.org
>>>>             <mailto:users@lists.open-mpi.org>>
>>>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >
>>>>                                        <https://rfd.newmexicoconsort
>>>> ium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>                     --         Ing. Gabriele Fatigati
>>>>
>>>>                     HPC specialist
>>>>
>>>>                     SuperComputing Applications and Innovation
>>>> Department
>>>>
>>>>                     Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>>>
>>>>             www.cineca.it <http://www.cineca.it>
>>>>             <http://www.cineca.it>                  Tel:
>>>>             +39 051 6171722 <tel:%2B39%20051%206171722
>>>> <%2B39%20051%206171722>>
>>>>             <tel:+39%20051%20617%201722 <+39%20051%20617%201722>>
>>>>
>>>>
>>>>                     g.fatigati [AT] cineca.it <http://cineca.it>
>>>>             <http://cineca.it>
>>>>
>>>>                     _______________________________________________
>>>>                     users mailing list
>>>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>>>             <mailto:users@lists.open-mpi.org
>>>>             <mailto:users@lists.open-mpi.org>>
>>>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >
>>>>                                <https://rfd.newmexicoconsort
>>>> ium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >>
>>>>
>>>>
>>>>
>>>>                 _______________________________________________
>>>>                 users mailing list
>>>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>>>             <mailto:users@lists.open-mpi.org
>>>>             <mailto:users@lists.open-mpi.org>>
>>>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >
>>>>                            <https://rfd.newmexicoconsort
>>>> ium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >>
>>>>
>>>>
>>>>
>>>>
>>>>             --             Ing. Gabriele Fatigati
>>>>
>>>>             HPC specialist
>>>>
>>>>             SuperComputing Applications and Innovation Department
>>>>
>>>>             Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>>>
>>>>             www.cineca.it <http://www.cineca.it>
>>>>             <http://www.cineca.it> Tel: +39 051 6171722
>>>>             <tel:%2B39%20051%206171722 <%2B39%20051%206171722>>
>>>>
>>>>             g.fatigati [AT] cineca.it <http://cineca.it>
>>>>             <http://cineca.it>
>>>>
>>>>
>>>>             _______________________________________________
>>>>             users mailing list
>>>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> >
>>>>
>>>>
>>>>         _______________________________________________
>>>>         users mailing list
>>>>         users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>>>         https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>>         <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>>>>
>>>>
>>>>
>>>>
>>>>     --     Ing. Gabriele Fatigati
>>>>
>>>>     HPC specialist
>>>>
>>>>     SuperComputing Applications and Innovation Department
>>>>
>>>>     Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>>>
>>>>     www.cineca.it <http://www.cineca.it>       Tel: +39 051 6171722
>>>>     <tel:051%20617%201722 <051%20617%201722>>
>>>>
>>>>     g.fatigati [AT] cineca.it <http://cineca.it>
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Ing. Gabriele Fatigati
>>>>
>>>> HPC specialist
>>>>
>>>> SuperComputing Applications and Innovation Department
>>>>
>>>> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>>>
>>>> www.cineca.it <http://www.cineca.it> Tel:   +39 051 6171722
>>>>
>>>> g.fatigati [AT] cineca.it <http://cineca.it>
>>>>
>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> users@lists.open-mpi.org
>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>>
>>>
>>> _______________________________________________
>>> users mailing list
>>> users@lists.open-mpi.org
>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>
>>
>>
>>
>> --
>> Ing. Gabriele Fatigati
>>
>> HPC specialist
>>
>> SuperComputing Applications and Innovation Department
>>
>> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>
>> www.cineca.it                    Tel:   +39 051 6171722
>> <+39%20051%20617%201722>
>>
>> g.fatigati [AT] cineca.it
>> <output_mpirun>_______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>
>>
>>
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>
>
>
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>



-- 
Ing. Gabriele Fatigati

HPC specialist

SuperComputing Applications and Innovation Department

Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy

www.cineca.it                    Tel:   +39 051 6171722

g.fatigati [AT] cineca.it
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to