Jeff Squyres wrote:
> Thanks for the feedback!  Sorry the lack of a clear error message caused
> you pain.  :-(
> 
> FWIW, we do have this feature in the upcoming v1.3 release.  For example:

That's very good to know. Thanks for the prompt and informative reply.


> 
> [6:46] svbu-mpi:~/mpi % mpirun -np 2 --mca btl openibbbb,self hello
> --------------------------------------------------------------------------
> A requested component was not found, or was unable to be opened.  This
> means that this component is either not installed or is unable to be
> used on your system (e.g., sometimes this means that shared libraries
> that the component requires are unable to be found/loaded).  Note that
> Open MPI stopped checking at the first component that it did not find.
> 
> Host:      svbu-mpi.cisco.com
> Framework: btl
> Component: openibbbb
> --------------------------------------------------------------------------
> 
> Hopefully, this will help such situations in the future.
> 
> 
> 
> On Aug 4, 2008, at 9:23 AM, Bernhard Hermann wrote:
> 
>> Hi,
>>
>> We would like to request a warning/error message to be printed whenever
>> a component (like "openib") is specified, but not actually loaded.
>>
>> Our case: Running a job via "tcp" over IB went OK, trying to use
>> "openib" failed. It turned out that while compiling the "openib" module
>> wasn't made (missing path), so OpenMPI croaked when trying to access it.
>>
>> This was the error that was of little help to us in tracing the problem:
>> --------------------------------------------------------------------------
>>
>> Process 0.1.7 is unable to reach 0.1.0 for MPI communication.
>> If you specified the use of a BTL component, you may have
>> forgotten a component (such as "self") in the list of
>> usable components.
>> --------------------------------------------------------------------------
>>
>> --------------------------------------------------------------------------
>>
>> It looks like MPI_INIT failed for some reason; your parallel process is
>> likely to abort.  There are many reasons that a parallel process can
>> fail during MPI_INIT; some of which are due to configuration or
>> environment
>> problems.  This failure appears to be an internal failure; here's some
>> additional information (which may only be relevant to an Open MPI
>> developer):
>>
>>  PML add procs failed
>>  --> Returned "Unreachable" (-12) instead of "Success" (0)
>> --------------------------------------------------------------------------
>>
>> *** An error occurred in MPI_Init
>> *** before MPI was initialized
>> *** MPI_ERRORS_ARE_FATAL (goodbye)
>>
>>
>>
>>
>> Unfortunately we can't provide a patch and would be glad if someone
>> could implement and include this.
>>
>> Thanks for your attention,
>> best regards,
>> Bernhard HERMANN
>>
>> init.at informationstechnologie GmbH.
>>
>> Mariahilfer Str. 61/1/10
>> 1060 Wien/Vienna
>> Austria - EU
>>
>> Tel: +43 1 522 53 77 - 35
>>
>> http://www.init.at
>> herm...@init.at
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 

-- 
Bernhard HERMANN

init.at informationstechnologie GmbH.

Mariahilfer Str. 61/1/10
1060 Wien/Vienna
Austria - EU

Tel: +43 1 522 53 77 - 35

http://www.init.at
herm...@init.at

Reply via email to