Can someone reply to my mail please? -----Original Message----- From: devel-boun...@open-mpi.org [mailto:devel-boun...@open-mpi.org] On Behalf Of Sridhar Chirravuri Sent: Wednesday, August 17, 2005 4:18 PM To: Open MPI Developers Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI
Hi, I checked out the latest code drop r6911 today morning and ran Pallas with in the same node (2 procs). It ran fine. I didn't see any hangs this time whereas I could see the following statements in the pallas output and I feel they are just warnings, which can be ignored. Am I correct? Request for 0 bytes (coll_basic_reduce_scatter.c, 80) Request for 0 bytes (coll_basic_reduce.c, 194) Request for 0 bytes (coll_basic_reduce_scatter.c, 80) Request for 0 bytes (coll_basic_reduce.c, 194) Request for 0 bytes (coll_basic_reduce_scatter.c, 80) Request for 0 bytes (coll_basic_reduce.c, 194) Here is the output of ompi_info command [root@micrompi-1 SRC_PMB]# ompi_info Open MPI: 1.0a1r6911 Open MPI SVN revision: r6911 Open RTE: 1.0a1r6911 Open RTE SVN revision: r6911 OPAL: 1.0a1r6911 OPAL SVN revision: r6911 Prefix: /openmpi Configured architecture: x86_64-redhat-linux-gnu Configured by: root Configured on: Wed Aug 17 23:59:09 IST 2005 Configure host: micrompi-1 Built by: root Built on: Thu Aug 18 00:09:33 IST 2005 Built host: micrompi-1 C bindings: yes C++ bindings: yes Fortran77 bindings: yes (all) Fortran90 bindings: no C compiler: gcc C compiler absolute: /usr/bin/gcc C++ compiler: g++ C++ compiler absolute: /usr/bin/g++ Fortran77 compiler: g77 Fortran77 compiler abs: /usr/bin/g77 Fortran90 compiler: none Fortran90 compiler abs: none C profiling: yes C++ profiling: yes Fortran77 profiling: yes Fortran90 profiling: no C++ exceptions: no Thread support: posix (mpi: no, progress: no) Internal debug support: yes MPI parameter check: runtime Memory profiling support: yes Memory debugging support: yes Memory hook support: yes libltdl support: 1 MCA memory: malloc_hooks (MCA v1.0, API v1.0, Component v1.0) MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.0) MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0) MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0) MCA coll: basic (MCA v1.0, API v1.0, Component v1.0) MCA coll: self (MCA v1.0, API v1.0, Component v1.0) MCA io: romio (MCA v1.0, API v1.0, Component v1.0) MCA mpool: mvapi (MCA v1.0, API v1.0, Component v1.0) MCA mpool: sm (MCA v1.0, API v1.0, Component v1.0) MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.0) MCA ptl: self (MCA v1.0, API v1.0, Component v1.0) MCA ptl: sm (MCA v1.0, API v1.0, Component v1.0) MCA ptl: tcp (MCA v1.0, API v1.0, Component v1.0) MCA btl: mvapi (MCA v1.0, API v1.0, Component v1.0) MCA btl: self (MCA v1.0, API v1.0, Component v1.0) MCA btl: sm (MCA v1.0, API v1.0, Component v1.0) MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0) MCA topo: unity (MCA v1.0, API v1.0, Component v1.0) MCA gpr: null (MCA v1.0, API v1.0, Component v1.0) MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.0) MCA gpr: replica (MCA v1.0, API v1.0, Component v1.0) MCA iof: proxy (MCA v1.0, API v1.0, Component v1.0) MCA iof: svc (MCA v1.0, API v1.0, Component v1.0) MCA ns: proxy (MCA v1.0, API v1.0, Component v1.0) MCA ns: replica (MCA v1.0, API v1.0, Component v1.0) MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0) MCA ras: host (MCA v1.0, API v1.0, Component v1.0) MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.0) MCA rds: resfile (MCA v1.0, API v1.0, Component v1.0) MCA rmaps: round_robin (MCA v1.0, API v1.0, Component v1.0) MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.0) MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.0) MCA rml: oob (MCA v1.0, API v1.0, Component v1.0) MCA pls: fork (MCA v1.0, API v1.0, Component v1.0) MCA pls: proxy (MCA v1.0, API v1.0, Component v1.0) MCA pls: rsh (MCA v1.0, API v1.0, Component v1.0) MCA sds: env (MCA v1.0, API v1.0, Component v1.0) MCA sds: pipe (MCA v1.0, API v1.0, Component v1.0) MCA sds: seed (MCA v1.0, API v1.0, Component v1.0) MCA sds: singleton (MCA v1.0, API v1.0, Component v1.0) Here is the output of sample MPI program which sends a char and recvs a char. [root@micrompi-1 ~]# mpirun -np 2 ./a.out Could not join a running, existing universe Establishing a new one named: default-universe-12913 [0,0,0] mca_oob_tcp_init: calling orte_gpr.subscribe [0,0,0] mca_oob_tcp_init: calling orte_gpr.put(orte-job-0) [0,0,0] mca_oob_tcp_accept: 10.52.0.98:52866 [0,0,0]-[0,0,1] accepted: 10.52.0.98 - 10.52.0.98 nodelay 0 sndbuf 50724 rcvbuf 87552 flags 00000802 [0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2 [0,1,0] mca_oob_tcp_init: calling orte_gpr.subscribe [0,1,0] mca_oob_tcp_init: calling orte_gpr.put(orte-job-1) [0,0,0] mca_oob_tcp_accept: 10.52.0.98:53634 [0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_peer_start_connect: connecting port 33487 to: 10.52.0.98:33483 [0,0,0]-[0,1,0] accepted: 10.52.0.98 - 10.52.0.98 nodelay 0 sndbuf 50724 rcvbuf 87552 flags 00000802 [0,1,0]-[0,0,0] connected: 10.52.0.98 - 10.52.0.98 nodelay 0 sndbuf 50724 rcvbuf 87552 flags 00000802 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,1,0]-[18446744073709551615,18446744073709551615,18446744073709551615] mca_oob_tcp_recv: tag 10 [0,0,0] mca_oob_tcp_accept: 10.52.0.97:13953 [0,0,0]-[0,0,2] accepted: 10.52.0.98 - 10.52.0.97 nodelay 0 sndbuf 16384 rcvbuf 87380 flags 00000802 [0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2 [0,1,1] mca_oob_tcp_init: calling orte_gpr.subscribe [0,1,1] mca_oob_tcp_init: calling orte_gpr.put(orte-job-1) [0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2 [0,0,0] mca_oob_tcp_accept: 10.52.0.97:14721 [0,1,1]-[0,0,0] mca_oob_tcp_peer_start_connect: connecting port 33079 to: 10.52.0.98:33483 [0,0,0]-[0,1,1] accepted: 10.52.0.98 - 10.52.0.97 nodelay 0 sndbuf 16384 rcvbuf 87380 flags 00000802 [0,1,1]-[0,0,0] connected: 10.52.0.97 - 10.52.0.98 nodelay 0 sndbuf 16384 rcvbuf 87380 flags 00000802 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 3 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 3 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 10 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 10 [0,1,0] mca_oob_tcp_registry_callback [0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,1,0] mca_oob_tcp_registry_callback: received peer [0,1,0] [0,1,0] mca_oob_tcp_registry_callback: received peer [0,1,1] [0,1,1]-[18446744073709551615,18446744073709551615,18446744073709551615] mca_oob_tcp_recv: tag 10 [0,1,1] mca_oob_tcp_registry_callback [0,1,1] mca_oob_tcp_registry_callback: received peer [0,1,0] [0,1,1] mca_oob_tcp_registry_callback: received peer [0,1,1] [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 10 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 10 [0,1,1]-[18446744073709551615,18446744073709551615,18446744073709551615] mca_oob_tcp_recv: tag 10 [0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,1,0]-[18446744073709551615,18446744073709551615,18446744073709551615] mca_oob_tcp_recv: tag 10 [0,1,0]-[0,1,1] mca_oob_tcp_peer_start_connect: connecting port 33487 to: 10.52.0.97:33079 [0,1,1] mca_oob_tcp_accept: 10.52.0.98:53890 [0,1,0]-[0,1,1] connected: 10.52.0.98 - 10.52.0.97 nodelay 0 sndbuf 16384 rcvbuf 87380 flags 00000802 [0,1,1]-[0,1,0] accepted: 10.52.0.97 - 10.52.0.98 nodelay 0 sndbuf 16384 rcvbuf 87380 flags 00000802 Task 1: Received 1 char(s) from task 0 with tag 1 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2 Task 0: Received 1 char(s) from task 1 with tag 1 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 10 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 10 [0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,1,0]-[18446744073709551615,18446744073709551615,18446744073709551615] mca_oob_tcp_recv: tag 10 [0,1,1]-[18446744073709551615,18446744073709551615,18446744073709551615] mca_oob_tcp_recv: tag 10 [0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2 [0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,1,0] mca_oob_tcp_send: tag 10 [0,0,0]-[0,1,1] mca_oob_tcp_send: tag 10 [0,1,0]-[18446744073709551615,18446744073709551615,18446744073709551615] mca_oob_tcp_recv: tag 10 [0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2 [0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2 [0,1,1]-[18446744073709551615,18446744073709551615,18446744073709551615] mca_oob_tcp_recv: tag 10 [0,0,0]-[0,1,0] mca_oob_tcp_msg_recv: peer closed connection [0,0,0]-[0,1,0] mca_oob_tcp_peer_close(0x526810) sd 6 state 4 [0,1,1]-[0,1,0] mca_oob_tcp_msg_recv: peer closed connection [0,1,1]-[0,1,0] mca_oob_tcp_peer_close(0x521210) sd 13 state 4 [0,0,0]-[0,1,1] mca_oob_tcp_msg_recv: peer closed connection [0,0,0]-[0,1,1] mca_oob_tcp_peer_close(0x526e30) sd 8 state 4 [0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,2] mca_oob_tcp_msg_recv: peer closed connection [0,0,0]-[0,0,2] mca_oob_tcp_peer_close(0x526b20) sd 7 state 4 [0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2 [0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2 My configure command looks like ./configure --prefix=/openmpi --with-btl-mvapi=/usr/local/topspin/ --enable-mca-no-build=btl-openib,pml-teg,pml-uniq Since I am working with mvapi component, I disabled openib. But I could see that data is going over TCP/GigE and not on Infiniband. I have run pallas, it simply hangs again :-( Note: I added pml=ob1 in the conf file /openmpi/etc/openmpi-mca-params.conf Any latest options being added to the configure command? Please let me know. Thanks -Sridhar -----Original Message----- From: devel-boun...@open-mpi.org [mailto:devel-boun...@open-mpi.org] On Behalf Of George Bosilca Sent: Wednesday, August 10, 2005 11:56 PM To: Open MPI Developers Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI I just commit the collective fix. Now I'm able to run the ex-pallas test on 2, 4 and 8 nodes without any errors. However, I'm unable to test the IB BTL as I don't have access to any machine having such hardware. george. PS: if one still notice the malloc with 0 bytes length, please sent a bug report. On Aug 10, 2005, at 6:28 AM, Sridhar Chirravuri wrote: > > Hi, > > I got the latest code drop of 6791 today morning. > > I have removed .ompi_ignore and .ompi_unignore files from > ompi/mca/mpool/mvapi directory. If I don't remove and build, the MPI > program fails with signal 11. After removing those hidden files from > that directory and building, signal 11 error disappeared. > > I have configured with the options given by Galen. > > ./configure --prefix=/openmpi --with-btl-mvapi=/usr/local/topspin/ > --enable-mca-no-build=btl-openib,pml-teg,pml-uniq > > After make all install, I have run pallas but I got the same error > messages (please see down below for error messages). I have run 3-4 > times, sometimes I didn't get any output but pallas just hungs. I have > run pingpong only. I have run pallas (all functions including reduce), > but got the following messages in intra-node case. > > Request for 0 bytes (coll_basic_reduce_scatter.c, 79) > Request for 0 bytes (coll_basic_reduce.c, 193) > Request for 0 bytes (coll_basic_reduce_scatter.c, 79) > Request for 0 bytes (coll_basic_reduce.c, 193) > > Since these types of messages seen by George, upcoming patch might > resolve this issue. > > Also, I have run mpi-ping.c program given by Galen with the latest > code > drop and it just hung. Here is the output > > [root@micrompi-1 ~]# mpirun -np 2 ./a.out -r 10 0 100000 1000 > Could not join a running, existing universe > Establishing a new one named: default-universe-12461 > mpi-ping: ping-pong > nprocs=2, reps=10, min bytes=0, max bytes=100000 inc bytes=1000 > 0 pings 1 > > > ... I just did ctrl+c here after 10 mins ... > > 2 processes killed (possibly by Open MPI) > > I have no clue whether the George patch will fix this problem or not. > > Before running mpi-ping program, I have export > OMPI_MCA_btl_base_debug=2 > in my shell. > > Thanks > -Sridhar > > -----Original Message----- > From: devel-boun...@open-mpi.org [mailto:devel-bounces@open- > mpi.org] On > Behalf Of Galen Shipman > Sent: Tuesday, August 09, 2005 11:10 PM > To: Open MPI Developers > Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI > > Hi > On Aug 9, 2005, at 8:15 AM, Sridhar Chirravuri wrote: > > > >> The same kind of output while running Pallas "pingpong" test. >> >> -Sridhar >> >> -----Original Message----- >> From: devel-boun...@open-mpi.org [mailto:devel-boun...@open-mpi.org] >> >> > On > > >> Behalf Of Sridhar Chirravuri >> Sent: Tuesday, August 09, 2005 7:44 PM >> To: Open MPI Developers >> Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI >> >> >> I have run sendrecv function in Pallas but it failed to run it. Here >> >> > is > > >> the output >> >> [root@micrompi-2 SRC_PMB]# mpirun -np 2 PMB-MPI1 sendrecv >> Could not join a running, existing universe >> Establishing a new one named: default-universe-5097 >> [0,1,1][btl_mvapi.c:130:mca_btl_mvapi_del_procs] Stub >> [0,1,1][btl_mvapi.c:130:mca_btl_mvapi_del_procs] Stub >> >> >> [0,1,0][btl_mvapi.c:130:mca_btl_mvapi_del_procs] Stub >> >> [0,1,0][btl_mvapi.c:130:mca_btl_mvapi_del_procs] Stub >> >> [0,1,0][btl_mvapi_endpoint.c:542:mca_btl_mvapi_endpoint_send] >> Connection >> to endpoint closed ... connecting ... >> [0,1,0][btl_mvapi_endpoint.c: >> 318:mca_btl_mvapi_endpoint_start_connect] >> Initialized High Priority QP num = 263177, Low Priority QP num = >> 263178, >> LID = 785 >> >> [0,1,0][btl_mvapi_endpoint.c:190: >> mca_btl_mvapi_endpoint_send_connect_req >> ] Sending High Priority QP num = 263177, Low Priority QP num = >> 263178, >> LID = 785[0,1,0][btl_mvapi_endpoint.c: >> 542:mca_btl_mvapi_endpoint_send] >> Connection to endpoint closed ... connecting ... >> [0,1,0][btl_mvapi_endpoint.c: >> 318:mca_btl_mvapi_endpoint_start_connect] >> Initialized High Priority QP num = 263179, Low Priority QP num = >> 263180, >> LID = 786 >> >> [0,1,0][btl_mvapi_endpoint.c:190: >> mca_btl_mvapi_endpoint_send_connect_req >> ] Sending High Priority QP num = 263179, Low Priority QP num = >> 263180, >> LID = 786#--------------------------------------------------- >> # PALLAS MPI Benchmark Suite V2.2, MPI-1 part >> #--------------------------------------------------- >> # Date : Tue Aug 9 07:11:25 2005 >> # Machine : x86_64# System : Linux >> # Release : 2.6.9-5.ELsmp >> # Version : #1 SMP Wed Jan 5 19:29:47 EST 2005 >> >> # >> # Minimum message length in bytes: 0 >> # Maximum message length in bytes: 4194304 >> # >> # MPI_Datatype : MPI_BYTE >> # MPI_Datatype for reductions : MPI_FLOAT >> # MPI_Op : MPI_SUM >> # >> # >> >> # List of Benchmarks to run: >> >> # Sendrecv >> [0,1,1][btl_mvapi_endpoint.c:368: >> mca_btl_mvapi_endpoint_reply_start_conn >> ect] Initialized High Priority QP num = 263177, Low Priority QP num = >> 263178, LID = 777 >> >> [0,1,1][btl_mvapi_endpoint.c:266: >> mca_btl_mvapi_endpoint_set_remote_info] >> Received High Priority QP num = 263177, Low Priority QP num 263178, >> LID >> = 785 >> >> [0,1,1][btl_mvapi_endpoint.c: >> 756:mca_btl_mvapi_endpoint_qp_init_query] >> Modified to init..Qp >> 7080096[0,1,1][btl_mvapi_endpoint.c:791: >> mca_btl_mvapi_endpoint_qp_init_q >> uery] Modified to RTR..Qp >> 7080096[0,1,1][btl_mvapi_endpoint.c:814: >> mca_btl_mvapi_endpoint_qp_init_q >> uery] Modified to RTS..Qp 7080096 >> >> [0,1,1][btl_mvapi_endpoint.c: >> 756:mca_btl_mvapi_endpoint_qp_init_query] >> Modified to init..Qp 7240736 >> [0,1,1][btl_mvapi_endpoint.c: >> 791:mca_btl_mvapi_endpoint_qp_init_query] >> Modified to RTR..Qp >> 7240736[0,1,1][btl_mvapi_endpoint.c:814: >> mca_btl_mvapi_endpoint_qp_init_q >> uery] Modified to RTS..Qp 7240736 >> [0,1,1][btl_mvapi_endpoint.c:190: >> mca_btl_mvapi_endpoint_send_connect_req >> ] Sending High Priority QP num = 263177, Low Priority QP num = >> 263178, >> LID = 777 >> [0,1,0][btl_mvapi_endpoint.c:266: >> mca_btl_mvapi_endpoint_set_remote_info] >> Received High Priority QP num = 263177, Low Priority QP num 263178, >> LID >> = 777 >> [0,1,0][btl_mvapi_endpoint.c: >> 756:mca_btl_mvapi_endpoint_qp_init_query] >> Modified to init..Qp 7081440 >> [0,1,0][btl_mvapi_endpoint.c: >> 791:mca_btl_mvapi_endpoint_qp_init_query] >> Modified to RTR..Qp 7081440 >> [0,1,0][btl_mvapi_endpoint.c: >> 814:mca_btl_mvapi_endpoint_qp_init_query] >> Modified to RTS..Qp 7081440 >> [0,1,0][btl_mvapi_endpoint.c: >> 756:mca_btl_mvapi_endpoint_qp_init_query] >> Modified to init..Qp 7241888 >> [0,1,0][btl_mvapi_endpoint.c: >> 791:mca_btl_mvapi_endpoint_qp_init_query] >> Modified to RTR..Qp >> 7241888[0,1,0][btl_mvapi_endpoint.c:814: >> mca_btl_mvapi_endpoint_qp_init_q >> uery] Modified to RTS..Qp 7241888 >> [0,1,1][btl_mvapi_component.c:523:mca_btl_mvapi_component_progress] >> >> > Got > > >> a recv completion >> >> >> Thanks >> -Sridhar >> >> >> >> >> -----Original Message----- >> From: devel-boun...@open-mpi.org [mailto:devel-boun...@open-mpi.org] >> >> > On > > >> Behalf Of Brian Barrett >> Sent: Tuesday, August 09, 2005 7:35 PM >> To: Open MPI Developers >> Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI >> >> On Aug 9, 2005, at 8:48 AM, Sridhar Chirravuri wrote: >> >> >> >>> Does r6774 has lot of changes that are related to 3rd generation >>> point-to-point? I am trying to run some benchmark tests (ex: >>> pallas) with Open MPI stack and just want to compare the >>> performance figures with MVAPICH 095 and MVAPICH 092. >>> >>> In order to use 3rd generation p2p communication, I have added the >>> following line in the /openmpi/etc/openmpi-mca-params.conf >>> >>> pml=ob1 >>> >>> I also exported (as double check) OMPI_MCA_pml=ob1. >>> >>> Then, I have tried running on the same machine. My machine has got >>> 2 processors. >>> >>> Mpirun -np 2 ./PMB-MPI1 >>> >>> I still see the following lines >>> >>> Request for 0 bytes (coll_basic_reduce_scatter.c, 79) >>> Request for 0 bytes (coll_basic_reduce.c, 193) >>> Request for 0 bytes (coll_basic_reduce_scatter.c, 79) >>> Request for 0 bytes (coll_basic_reduce.c, 193) >>> >>> >> >> These errors are coming from the collective routines, not the PML/BTL >> layers. It looks like the reduction codes are trying to call malloc >> (0), which doesn't work so well. We'll take a look as soon as we >> can. In the mean time, can you just not run the tests that call the >> reduction collectives? >> >> Brian >> >> >> -- >> Brian Barrett >> Open MPI developer >> http://www.open-mpi.org/ >> >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/devel >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/devel >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/devel >> >> > > _______________________________________________ > devel mailing list > de...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/devel > > _______________________________________________ > devel mailing list > de...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/devel > > _______________________________________________ devel mailing list de...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/devel _______________________________________________ devel mailing list de...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/devel