Re: [OMPI users] no allocated resources for the application........(mpirun)
Hi Gus Correa, Thank you for your response. I'll check what your saying and see if it'll work. many thanks sincerly Mouhamad Gus Correaa écrit : Hi Mouhamad Did you check if your Open MPI setup is working with the simple programs in the "examples" subdirectory (connectivity_c.c, hello_c.c, ring_c.c)? This will tell you if the problem is with Open MPI or with arpege (whatever program arpege is). You can compile the examples with mpicc and run them with mpirun with the same hostfile that you used for arpege. Also, make sure your PATH and LD_LIBRARY_PATH point to the bin and lib subdirectories of your Open MPI installation. You can set this in your .cshrc/.bashrc file. See: http://www.open-mpi.org/faq/?category=running#adding-ompi-to-path The Open MPI FAQ are worth reading, and may help you with this and other problems: http://www.open-mpi.org/faq/ Just a suggestion. Gus Correa Mouhamad Al-Sayed-Ali wrote: Hello all, I have been trying to run the executable "arpege" with mpirun -np 2 --host file arpege where file contains the name of the machines. But, I get the following error: - -- There are no allocated resources for the application arpege that match the requested mapping: Verify that you have mapped the allocated resources properly using the --host or --hostfile specification. -- -- A daemon (pid unknown) died unexpectedly on signal 1 while attempting to launch so we are aborting. There may be more information reported by the environment (see above). This may be because the daemon was unable to find all the needed shared libraries on the remote node. You may set your LD_LIBRARY_PATH to have the location of the shared libraries on the remote nodes and this will automatically be forwarded to the remote nodes. -- -- mpirun noticed that the job aborted, but has no info as to the process that caused that situation. -- mpirun: clean termination accomplished - Can anyone help me, please ? Sincerly Mouhamad Al sayed ali ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] no allocated resources for the application........(mpirun)
Hi Mouhamad Did you check if your Open MPI setup is working with the simple programs in the "examples" subdirectory (connectivity_c.c, hello_c.c, ring_c.c)? This will tell you if the problem is with Open MPI or with arpege (whatever program arpege is). You can compile the examples with mpicc and run them with mpirun with the same hostfile that you used for arpege. Also, make sure your PATH and LD_LIBRARY_PATH point to the bin and lib subdirectories of your Open MPI installation. You can set this in your .cshrc/.bashrc file. See: http://www.open-mpi.org/faq/?category=running#adding-ompi-to-path The Open MPI FAQ are worth reading, and may help you with this and other problems: http://www.open-mpi.org/faq/ Just a suggestion. Gus Correa Mouhamad Al-Sayed-Ali wrote: Hello all, I have been trying to run the executable "arpege" with mpirun -np 2 --host file arpege where file contains the name of the machines. But, I get the following error: - -- There are no allocated resources for the application arpege that match the requested mapping: Verify that you have mapped the allocated resources properly using the --host or --hostfile specification. -- -- A daemon (pid unknown) died unexpectedly on signal 1 while attempting to launch so we are aborting. There may be more information reported by the environment (see above). This may be because the daemon was unable to find all the needed shared libraries on the remote node. You may set your LD_LIBRARY_PATH to have the location of the shared libraries on the remote nodes and this will automatically be forwarded to the remote nodes. -- -- mpirun noticed that the job aborted, but has no info as to the process that caused that situation. -- mpirun: clean termination accomplished - Can anyone help me, please ? Sincerly Mouhamad Al sayed ali ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users
[OMPI users] no allocated resources for the application........(mpirun)
Hello all, I have been trying to run the executable "arpege" with mpirun -np 2 --host file arpege where file contains the name of the machines. But, I get the following error: - -- There are no allocated resources for the application arpege that match the requested mapping: Verify that you have mapped the allocated resources properly using the --host or --hostfile specification. -- -- A daemon (pid unknown) died unexpectedly on signal 1 while attempting to launch so we are aborting. There may be more information reported by the environment (see above). This may be because the daemon was unable to find all the needed shared libraries on the remote node. You may set your LD_LIBRARY_PATH to have the location of the shared libraries on the remote nodes and this will automatically be forwarded to the remote nodes. -- -- mpirun noticed that the job aborted, but has no info as to the process that caused that situation. -- mpirun: clean termination accomplished - Can anyone help me, please ? Sincerly Mouhamad Al sayed ali
Re: [OMPI users] problems with parallel MPI-IO...
On Tuesday, July 19, 2011 16:13:38 Jonathan Dursi wrote: > On 19 Jul 9:02AM, Alexander Beck-Ratzka wrote: > >integer ierr, i, myrank, BUFSIZE, thefile, intsize > >parameter (BUFSIZE=100) > >integer buf(BUFSIZE) > > > >do i=0,BUFSIZE > > > > buf(i) = myrank*BUFSIZE + i > > print*, 'i =', i, 'myrank =', myrank, 'buf(i)=',buf(i) > > > >end do > > [...] > > > When I am reading the data in again and print them out, I always have: > > > > buf(0)=0 > > If you compile your code with -check bounds and run, you'll get an error > pointing out that buf(0) is an illegal access; in Fortran arrays start at > 1. > Thanks a lot, that solved the problem! Best wishes Alexander
Re: [OMPI users] problems with parallel MPI-IO...
On 19 Jul 9:02AM, Alexander Beck-Ratzka wrote: integer ierr, i, myrank, BUFSIZE, thefile, intsize parameter (BUFSIZE=100) integer buf(BUFSIZE) do i=0,BUFSIZE buf(i) = myrank*BUFSIZE + i print*, 'i =', i, 'myrank =', myrank, 'buf(i)=',buf(i) end do [...] When I am reading the data in again and print them out, I always have: buf(0)=0 If you compile your code with -check bounds and run, you'll get an error pointing out that buf(0) is an illegal access; in Fortran arrays start at 1. - Jonathan -- Jonathan Dursi | SciNet, Compute/Calcul Canada
[OMPI users] problems with parallel MPI-IO...
Hi Folks, I am using the following f90 example program for writing a file in parallel with MIP: [snip] program parallel_io use mpi implicit none integer ierr, i, myrank, BUFSIZE, thefile, intsize parameter (BUFSIZE=100) integer buf(BUFSIZE) ! integer (kind=MPI_OFFSET_KIND) disp integer*8 disp call mpi_init(ierr) call mpi_comm_rank(mpi_comm_world, myrank,ierr) do i=0,BUFSIZE buf(i) = myrank*BUFSIZE + i print*, 'i =', i, 'myrank =', myrank, 'buf(i)=',buf(i) end do call mpi_file_open(mpi_comm_world, 'testfile', MPI_MODE_WRONLY + MPI_MODE_CREATE, mpi_info_null, thefile, ierr) call mpi_type_size(MPI_INTEGER, intsize, ierr) disp = myrank * BUFSIZE * intsize ! call mpi_file_set_view(thefile, disp, MPI_INTEGER, MPI_INTEGER, 'native', mpi_info_null, ierr) call mpi_file_write_at(thefile, disp, buf, BUFSIZE, MPI_INTEGER, mpi_status_ignore, ierr) call mpi_file_close(thefile, ierr) call mpi_finalize(ierr) end program parallel_io [snip] And the follwoing program shall read all the data in again, and print them out [snip] program parallel_read_io use mpi implicit none integer ierr, i, myrank, BUFSIZE, thefile, intsize parameter (BUFSIZE=100) integer buf(BUFSIZE) ! integer (kind=MPI_OFFSET_KIND) disp integer*8 disp call mpi_init(ierr) call mpi_comm_rank(mpi_comm_world, myrank,ierr) ! do i=0,BUFSIZE ! buf(i) = myrank*BUFSIZE + i ! end do call mpi_file_open(mpi_comm_world, 'testfile', MPI_MODE_RDONLY, mpi_info_null, thefile, ierr) call mpi_type_size(MPI_INTEGER, intsize, ierr) disp = myrank * BUFSIZE * intsize ! call mpi_file_set_view(thefile, disp, MPI_INTEGER, MPI_INTEGER, 'native', mpi_info_null, ierr) ! call mpi_file_read(thefile, buf, BUFSIZE, MPI_INTEGER, mpi_status_ignore, ierr) call mpi_file_read_at(thefile, disp, buf, BUFSIZE, MPI_INTEGER, mpi_status_ignore, ierr) call mpi_file_close(thefile, ierr) ! print the data read in... if (myrank.eq.1) then do i = 0,BUFSIZE print*, 'i =', i, 'myrank =', myrank, 'buf(i)=', buf(i) end do endif call mpi_finalize(ierr) [snip] I have maid several tests, also with do loops only from 0 to (BUFSIZE-1), with MPI_FILE_SET_VIEW and MPI_FILE_READ, etc... When I am reading the data in again and print them out, I always have: buf(0)=0 for every rank, so I assume that something with the offset is wrong. I am using openmpi with an Intel f90 compiler. What am I doing wrong? Best wishes Alexander
Re: [OMPI users] InfiniBand, different OpenFabrics transport types
Yevgeny, Sorry for the delay in replying -- I'd been out for a few days. - Original Message - > From: Yevgeny Kliteynik> Sent: Thursday, July 14, 2011 12:51 AM > Subject: Re: [OMPI users] InfiniBand, different OpenFabrics transport types > While I'm trying to find an old HCA somewhere, could you please > post here the output of "ibv_devinfo -v" on mthca? :~$ ibv_devinfo -v hca_id: mthca0 transport: InfiniBand (0) fw_ver: 4.8.917 node_guid: 0005:ad00:000b:60c0 sys_image_guid: 0005:ad00:0100:d050 vendor_id: 0x05ad vendor_part_id: 25208 hw_ver: 0xA0 board_id: MT_00A001 phys_port_cnt: 2 max_mr_size: 0x page_size_cap: 0xf000 max_qp: 64512 max_qp_wr: 65535 device_cap_flags: 0x1c76 max_sge: 59 max_sge_rd: 0 max_cq: 65408 max_cqe: 131071 max_mr: 131056 max_pd: 32768 max_qp_rd_atom: 4 max_ee_rd_atom: 0 max_res_rd_atom: 258048 max_qp_init_rd_atom: 128 max_ee_init_rd_atom: 0 atomic_cap: ATOMIC_HCA (1) max_ee: 0 max_rdd: 0 max_mw: 0 max_raw_ipv6_qp: 0 max_raw_ethy_qp: 0 max_mcast_grp: 8192 max_mcast_qp_attach: 56 max_total_mcast_qp_attach: 458752 max_ah: 0 max_fmr: 0 max_srq: 960 max_srq_wr: 65535 max_srq_sge: 31 max_pkeys: 64 local_ca_ack_delay: 15 port: 1 state: PORT_ACTIVE (4) max_mtu: 2048 (4) active_mtu: 2048 (4) sm_lid: 2 port_lid: 49 port_lmc: 0x00 link_layer: IB max_msg_sz: 0x8000 port_cap_flags: 0x02510a68 max_vl_num: 8 (4) bad_pkey_cntr: 0x0 qkey_viol_cntr: 0x0 sm_sl: 0 pkey_tbl_len: 64 gid_tbl_len: 32 subnet_timeout: 8 init_type_reply: 0 active_width: 4X (2) active_speed: 2.5 Gbps (1) phys_state: LINK_UP (5) GID[ 0]: fe80::::0005:ad00:000b:60c1 port: 2 state: PORT_DOWN (1) max_mtu: 2048 (4) active_mtu: 512 (2) sm_lid: 0 port_lid: 0 port_lmc: 0x00 link_layer: IB max_msg_sz: 0x8000 port_cap_flags: 0x02510a68 max_vl_num: 8 (4) bad_pkey_cntr: 0x0 qkey_viol_cntr: 0x0 sm_sl: 0 pkey_tbl_len: 64 gid_tbl_len: 32 subnet_timeout: 0 init_type_reply: 0 active_width: 4X (2) active_speed: 2.5 Gbps (1) phys_state: POLLING (2) GID[ 0]: fe80::::0005:ad00:000b:60c2