Re: [OMPI users] MPI_Comm_accept()/connect() errors

2009-10-08 Thread Blesson Varghese
The PATH variable contains
/home/hx019035/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/
bin:/usr/games:/usr/local/maui/bin/:

/home/hx019035/bin contains the local installation of OMPI 1.3.3



The LD_LIBRARY_PATH variable contains /home/hx019035/lib:



These variables are being set in the .profile file on the hpcc00 node. 



Would there be a change anywhere else?





From: Ralph Castain [mailto:rhc.open...@gmail.com] On Behalf Of Ralph
Castain
Sent: 07 October 2009 13:32
To: Blesson Varghese
Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors



Yes, it does. But the error message indicates a 1.2 version is running on
hpcc00.



On Oct 7, 2009, at 5:46 AM, Blesson Varghese wrote:



Just a quick question. Would mpirun -version give me the version of the
mpirun being executed? I am getting the result of that as 1.3.3.



From: Ralph Castain [mailto:rhc.open...@gmail.com] On Behalf Of Ralph
Castain
Sent: 07 October 2009 11:58
To: Blesson Varghese
Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors



Hate to tell you this, but your output clearly indicates you are NOT running
1.3.3 - that is an output from a 1.2.x version of OMPI.



Check you path and ld_library_path - you're still picking up the 1.2.5
version somewhere.





On Oct 7, 2009, at 4:05 AM, Blesson Varghese wrote:





Hi,



Please refer to the emails below.



I have made an upgrade to Open MPI 1.3.3 as suggested. The necessary
environment variables have all been set. Attaching the output of ompi_info
-all. However, the errors continue to persist.



[hpcc00:31864] [0,0,0] ORTE_ERROR_LOG: Not found in file dss/dss_unpack.c at
line 209

[hpcc00:31864] [0,0,0] ORTE_ERROR_LOG: Not found in file
communicator/comm_dyn.c at line 186

[hpcc00:31864] *** An error occurred in MPI_Comm_connect

[hpcc00:31864] *** on communicator MPI_COMM_WORLD

[hpcc00:31864] *** MPI_ERR_INTERN: internal error

[hpcc00:31864] *** MPI_ERRORS_ARE_FATAL (goodbye)





The server program is as follows:



#include 

#include 

#include 



int main( int argc, char **argv )

{

  MPI_Comm client;

  MPI_Status status;

  char port_name[MPI_MAX_PORT_NAME];

  int buf;

  int size, again;

  MPI_Info portInfo;



  MPI_Init( ,  );



  MPI_Comm_size(MPI_COMM_WORLD, );



  MPI_Open_port(MPI_INFO_NULL, port_name);



  printf("server available at %s\n",port_name);



  MPI_Comm_accept(port_name, MPI_INFO_NULL, 0, MPI_COMM_WORLD, 
);

  MPI_Recv(, 1, MPI_INT, MPI_ANY_SOURCE, MPI_ANY_TAG, client,
 );

  MPI_Comm_disconnect(  );

}



The client program is as follows:



#include 

#include 

#include 

#include 



int main( int argc, char **argv )

{

MPI_Comm server;

int buf = 8;

char port_name[MPI_MAX_PORT_NAME];

MPI_Info portInfo;



MPI_Init( ,  );



strcpy(port_name, "0.0.0:2000"); //The port name is hardcoded since
0.0.0:2000 is generated by the server program

MPI_Comm_connect(port_name, MPI_INFO_NULL, 0, MPI_COMM_WORLD,  );



MPI_Send(, 1, MPI_INT, 0, 1, server );



MPI_Comm_disconnect(  );

MPI_Finalize();

return 0;

}



Would you please advise?



Regards,

Blesson.





-Original Message-
From: Blesson Varghese [mailto:hx019...@reading.ac.uk] 
Sent: 03 October 2009 12:20
To: 'Jeff Squyres'
Subject: RE: [OMPI users] MPI_Comm_accept()/connect() errors



Thank you. I shall try the upgrade very soon.



-Original Message-

From: Jeff Squyres [mailto:jsquy...@cisco.com]

Sent: 03 October 2009 12:18

To: Blesson Varghese

Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors



On Oct 3, 2009, at 7:14 AM, Blesson Varghese wrote:



> Thanks for your reply Jeff. Since, it is a teaching cluster of the

> University,  I am quite unsure if I would be able to upgrade it very 

> soon.

> 

> Do you reckon that the error is due to the Open MPI version?

> 



You can always install your own version of Open MPI under your $HOME 

or somesuch -- there is no requirement that Open MPI is installed by 

root in a central location.



That being said, you might want to check with your administrator to 

ensure that this is ok with local policies -- see if they did any 

special setup for Open MPI, etc.



But yes, we made a bunch of COMM_SPAWN improvements since the 1.2 

series.



--

Jeff Squyres

jsquy...@cisco.com





From: Blesson Varghese [mailto:hx019...@reading.ac.uk] 
Sent: 01 October 2009 12:01
To: 'Open MPI Users'; 'Ralph Castain'
Subject: RE: [OMPI users] MPI_Comm_accept()/connect() errors



The following is the information regarding the error. I am running Open MPI
1.2.5 on Ubuntu 4.2.4, kernel version 2.6.24



I ran the server program as mpirun -np 1 server. This program gave me the
output port as 0.1.0:2000. I used this port name value as the command line
argument for the client program: mpirun -np 1 client 0.1.1:2000.



- The output of the "ompi_info --all&q

Re: [OMPI users] MPI_Comm_accept()/connect() errors

2009-10-03 Thread Jeff Squyres

On Oct 1, 2009, at 7:00 AM, Blesson Varghese wrote:

The following is the information regarding the error. I am running  
Open MPI 1.2.5 on Ubuntu 4.2.4, kernel version 2.6.24


Is there any chance that you can upgrade to the Open MPI v1.3 series?

--
Jeff Squyres
jsquy...@cisco.com



Re: [OMPI users] MPI_Comm_accept()/connect() errors

2009-10-01 Thread Blesson Varghese
The following is the information regarding the error. I am running Open MPI
1.2.5 on Ubuntu 4.2.4, kernel version 2.6.24



I ran the server program as mpirun -np 1 server. This program gave me the
output port as 0.1.0:2000. I used this port name value as the command line
argument for the client program: mpirun -np 1 client 0.1.1:2000.



- The output of the "ompi_info --all" is attached with the email

- PATH Variable:
/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr
/local/maui/bin/:

- LD_LIBRARY_PATH variable was empty

- The following is the output of ifconfig on hpcc00 from where the error has
been generated:

eth0  Link encap:Ethernet  HWaddr 00:12:3f:4c:2d:78

  inet addr:134.225.200.100  Bcast:134.225.200.255
Mask:255.255.255.0

  inet6 addr: fe80::212:3fff:fe4c:2d78/64 Scope:Link

  UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

  RX packets:15912728 errors:0 dropped:0 overruns:0 frame:0

  TX packets:15312376 errors:0 dropped:0 overruns:0 carrier:0

  collisions:0 txqueuelen:1000

  RX bytes:2951880321 (2.7 GB)  TX bytes:2788249498 (2.5 GB)

  Interrupt:16



loLink encap:Local Loopback

  inet addr:127.0.0.1  Mask:255.0.0.0

  inet6 addr: ::1/128 Scope:Host

  UP LOOPBACK RUNNING  MTU:16436  Metric:1

  RX packets:3507489 errors:0 dropped:0 overruns:0 frame:0

  TX packets:3507489 errors:0 dropped:0 overruns:0 carrier:0

  collisions:0 txqueuelen:0

  RX bytes:1794266658 (1.6 GB)  TX bytes:1794266658 (1.6 GB)



Regards,

Blesson.



From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] On
Behalf Of Ralph Castain
Sent: 29 September 2009 23:59
To: Open MPI Users
Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors



I will ask the obvious - what version of Open MPI are you running? In what
environment? What was your command line?



:-)



On Sep 29, 2009, at 3:50 PM, Blesson Varghese wrote:



Hi,



I have been trying to execute the server.c and client.c program provided in
http://www.mpi-forum.org/docs/mpi21-report/node213.htm#Node213, using
accept() and connect() function in MPI. However, the following errors are
generated.



[hpcc00:16522] *** An error occurred in MPI_Comm_connect

[hpcc00:16522] *** on communicator MPI_COMM_WORLD

[hpcc00:16522] *** MPI_ERR_INTERN: internal error

[hpcc00:16522] *** MPI_ERRORS_ARE_FATAL (goodbye)



Could anybody please help me?



Many thanks,
Blesson.

___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



Open MPI: 1.2.5
   Open MPI SVN revision: r16989
Open RTE: 1.2.5
   Open RTE SVN revision: r16989
OPAL: 1.2.5
   OPAL SVN revision: r16989
   MCA backtrace: execinfo (MCA v1.0, API v1.0, Component v1.2.5)
  MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.2.5)
   MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2.5)
   MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.2.5)
   MCA timer: linux (MCA v1.0, API v1.0, Component v1.2.5)
 MCA installdirs: env (MCA v1.0, API v1.0, Component v1.2.5)
 MCA installdirs: config (MCA v1.0, API v1.0, Component v1.2.5)
   MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
   MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
MCA coll: basic (MCA v1.0, API v1.0, Component v1.2.5)
MCA coll: self (MCA v1.0, API v1.0, Component v1.2.5)
MCA coll: sm (MCA v1.0, API v1.0, Component v1.2.5)
MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2.5)
  MCA io: romio (MCA v1.0, API v1.0, Component v1.2.5)
   MCA mpool: rdma (MCA v1.0, API v1.0, Component v1.2.5)
   MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2.5)
 MCA pml: cm (MCA v1.0, API v1.0, Component v1.2.5)
 MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2.5)
 MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2.5)
  MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2.5)
 MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2.5)
 MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2.5)
 MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0)
MCA topo: unity (MCA v1.0, API v1.0, Component v1.2.5)
 MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2.5)
  MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2.5)
  MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2.5)
  MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2.5)
 MCA gpr: null (MCA v1.0, API v1.0, Component v1.2.5)
 MCA gpr: proxy (MCA v1.0, API v1.0, Compon

Re: [OMPI users] MPI_Comm_accept()/connect() errors

2009-09-30 Thread Blesson Varghese
Thanks for your reply. 



I am running MPI 2.0 on Ubuntu 4.2.4, kernel version 2.6.24.



I ran the server program as mpirun -np 1 server. This program gave me the
output port as 0.1.0:2000. I used this port name value as the command line
argument for the client program: mpirun -np 1 client 0.1.1:2000



Regards,

Blesson. 



From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] On
Behalf Of Ralph Castain
Sent: 29 September 2009 23:59
To: Open MPI Users
Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors



I will ask the obvious - what version of Open MPI are you running? In what
environment? What was your command line?



:-)



On Sep 29, 2009, at 3:50 PM, Blesson Varghese wrote:





Hi,



I have been trying to execute the server.c and client.c program provided in
http://www.mpi-forum.org/docs/mpi21-report/node213.htm#Node213, using
accept() and connect() function in MPI. However, the following errors are
generated.



[hpcc00:16522] *** An error occurred in MPI_Comm_connect

[hpcc00:16522] *** on communicator MPI_COMM_WORLD

[hpcc00:16522] *** MPI_ERR_INTERN: internal error

[hpcc00:16522] *** MPI_ERRORS_ARE_FATAL (goodbye)



Could anybody please help me?



Many thanks,
Blesson.

___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users





Re: [OMPI users] MPI_Comm_accept()/connect() errors

2009-09-29 Thread Ralph Castain
I will ask the obvious - what version of Open MPI are you running? In  
what environment? What was your command line?


:-)

On Sep 29, 2009, at 3:50 PM, Blesson Varghese wrote:


Hi,

I have been trying to execute the server.c and client.c program  
provided in http://www.mpi-forum.org/docs/mpi21-report/node213.htm#Node213 
, using accept() and connect() function in MPI. However, the  
following errors are generated.


[hpcc00:16522] *** An error occurred in MPI_Comm_connect
[hpcc00:16522] *** on communicator MPI_COMM_WORLD
[hpcc00:16522] *** MPI_ERR_INTERN: internal error
[hpcc00:16522] *** MPI_ERRORS_ARE_FATAL (goodbye)

Could anybody please help me?

Many thanks,
Blesson.
___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users




[OMPI users] MPI_Comm_accept()/connect() errors

2009-09-29 Thread Blesson Varghese
Hi,



I have been trying to execute the server.c and client.c program provided in
http://www.mpi-forum.org/docs/mpi21-report/node213.htm#Node213, using
accept() and connect() function in MPI. However, the following errors are
generated.



[hpcc00:16522] *** An error occurred in MPI_Comm_connect

[hpcc00:16522] *** on communicator MPI_COMM_WORLD

[hpcc00:16522] *** MPI_ERR_INTERN: internal error

[hpcc00:16522] *** MPI_ERRORS_ARE_FATAL (goodbye)



Could anybody please help me?



Many thanks,
Blesson.