I have filed a ticket for this:

    https://svn.open-mpi.org/trac/ompi/ticket/972


On Apr 3, 2007, at 5:18 PM, Xie, Hugh wrote:


I think that workaround you purposed would resolve this problem.

-----Original Message-----
From: users-boun...@open-mpi.org [mailto:users-bounces@open- mpi.org] On
Behalf Of Jeff Squyres
Sent: Tuesday, April 03, 2007 5:05 PM
To: Open MPI Users
Subject: Re: [OMPI users] Mpirun failed for machines not in the same
subnet.

Do your different subnets violate the assumptions listed here?

     http://www.open-mpi.org/faq/?category=tcp#tcp-routability

We have not implemented any workarounds to say "subnet X is routable to
subnet Y" because no one had asked for them.  Such workarounds are
possible, of course, but I don't know what kind of timeframe we would be able to implement them in. Contributions would always be accepted! :-)

Probably the easiest workaround would be a top-level MCA parameter that
effectively tells OMPI to assume that *all* TCP addresses are routable
to each other.  That might not be too difficult to implement.


On Apr 3, 2007, at 4:11 PM, Xie, Hugh wrote:


Hi,

I got the follow error message while running: 'mpirun -v -np 2
-machinefile hosts.txt testc.x'

Process 0.1.1 is unable to reach 0.1.0 for MPI communication.
If you specified the use of a BTL component, you may have forgotten a
component (such as "self") in the list of usable components.
--------------------------------------------------------------------- -
--
--
--------------------------------------------------------------------- -
--
--
Process 0.1.0 is unable to reach 0.1.1 for MPI communication.
If you specified the use of a BTL component, you may have forgotten a
component (such as "self") in the list of usable components.
--------------------------------------------------------------------- -
--
--
--------------------------------------------------------------------- -
--
--
It looks like MPI_INIT failed for some reason; your parallel process
is likely to abort.  There are many reasons that a parallel process
can fail during MPI_INIT; some of which are due to configuration or
environment problems. This failure appears to be an internal failure;

here's some additional information (which may only be relevant to an
Open MPI
developer):

  PML add procs failed
  --> Returned "Unreachable" (-12) instead of "Success" (0)
--------------------------------------------------------------------- -
--
--


The same commands works if the content in hosts.txt is in same subnet.
Once I switch to hosts in different subnet, it stop working. I am
using ompi 1.2.
Please help.

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
- - - - - - - - - -

This message is intended only for the personal and confidential use of

the designated recipient(s) named above.  If you are not the intended
recipient of this message you are hereby notified that any review,
dissemination, distribution or copying of this message is strictly
prohibited.  This communication is for information purposes only and
should not be regarded as an offer to sell or as a solicitation of an
offer to buy any financial product, an official confirmation of any
transaction, or as an official statement of Lehman Brothers.  Email
transmission cannot be guaranteed to be secure or error-free.
Therefore, we do not represent that this information is complete or
accurate and it should not be relied upon as such. All information is

subject to change without notice.

--------
IRS Circular 230 Disclosure:
Please be advised that any discussion of U.S. tax matters contained
within this communication (including any attachments) is not intended
or written to be used and cannot be used for the purpose of (i)
avoiding U.S. tax related penalties or (ii) promoting, marketing or
recommending to another party any transaction or matter addressed
herein.



_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


--
Jeff Squyres
Cisco Systems

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

This message is intended only for the personal and confidential use of the designated recipient(s) named above. If you are not the intended recipient of this message you are hereby notified that any review, dissemination, distribution or copying of this message is strictly prohibited. This communication is for information purposes only and should not be regarded as an offer to sell or as a solicitation of an offer to buy any financial product, an official confirmation of any transaction, or as an official statement of Lehman Brothers. Email transmission cannot be guaranteed to be secure or error-free. Therefore, we do not represent that this information is complete or accurate and it should not be relied upon as such. All information is subject to change without notice.

--------
IRS Circular 230 Disclosure:
Please be advised that any discussion of U.S. tax matters contained within this communication (including any attachments) is not intended or written to be used and cannot be used for the purpose of (i) avoiding U.S. tax related penalties or (ii) promoting, marketing or recommending to another party any transaction or matter addressed herein.



_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


--
Jeff Squyres
Cisco Systems

Reply via email to