Re: [OMPI users] no allocated resources for the application........(mpirun)

2011-07-19 Thread Mouhamad Al-Sayed-Ali

Hi Gus Correa,

 Thank you for your response.

 I'll check what your saying and see if it'll work.

many thanks

sincerly

Mouhamad
Gus Correa  a écrit :


Hi Mouhamad

Did you check if your Open MPI setup is working with the simple
programs in the "examples" subdirectory (connectivity_c.c,  
hello_c.c, ring_c.c)?

This will tell you if the problem is with Open MPI or with arpege
(whatever program arpege is).
You can compile the examples with mpicc and run them with mpirun
with the same hostfile that you used for arpege.

Also, make sure your PATH and LD_LIBRARY_PATH point to the bin and
lib subdirectories of your Open MPI installation.
You can set this in your .cshrc/.bashrc file.
See:
http://www.open-mpi.org/faq/?category=running#adding-ompi-to-path

The Open MPI FAQ are worth reading, and may help you with this and
other problems:
http://www.open-mpi.org/faq/

Just a suggestion.
Gus Correa

Mouhamad Al-Sayed-Ali wrote:

Hello all,

I have been trying to run the executable "arpege" with

mpirun -np 2 --host file arpege


where file contains the name of the machines. But, I get the  
following error:


-
--
There are no allocated resources for the application
 arpege
that match the requested mapping:


Verify that you have mapped the allocated resources properly using the
--host or --hostfile specification.
--
--
A daemon (pid unknown) died unexpectedly on signal 1  while attempting to
launch so we are aborting.

There may be more information reported by the environment (see above).

This may be because the daemon was unable to find all the needed shared
libraries on the remote node. You may set your LD_LIBRARY_PATH to have the
location of the shared libraries on the remote nodes and this will
automatically be forwarded to the remote nodes.
--
--
mpirun noticed that the job aborted, but has no info as to the process
that caused that situation.
--
mpirun: clean termination accomplished
-


Can anyone help me, please ?


Sincerly


Mouhamad Al sayed ali
___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users







Re: [OMPI users] no allocated resources for the application........(mpirun)

2011-07-19 Thread Gus Correa

Hi Mouhamad

Did you check if your Open MPI setup is working with the simple
programs in the "examples" subdirectory (connectivity_c.c, hello_c.c, 
ring_c.c)?

This will tell you if the problem is with Open MPI or with arpege
(whatever program arpege is).
You can compile the examples with mpicc and run them with mpirun
with the same hostfile that you used for arpege.

Also, make sure your PATH and LD_LIBRARY_PATH point to the bin and
lib subdirectories of your Open MPI installation.
You can set this in your .cshrc/.bashrc file.
See:
http://www.open-mpi.org/faq/?category=running#adding-ompi-to-path

The Open MPI FAQ are worth reading, and may help you with this and
other problems:
http://www.open-mpi.org/faq/

Just a suggestion.
Gus Correa

Mouhamad Al-Sayed-Ali wrote:

Hello all,

 I have been trying to run the executable "arpege" with

mpirun -np 2 --host file arpege


where file contains the name of the machines. But, I get the following 
error:


-
--
There are no allocated resources for the application
  arpege
that match the requested mapping:


Verify that you have mapped the allocated resources properly using the
--host or --hostfile specification.
--
--
A daemon (pid unknown) died unexpectedly on signal 1  while attempting to
launch so we are aborting.

There may be more information reported by the environment (see above).

This may be because the daemon was unable to find all the needed shared
libraries on the remote node. You may set your LD_LIBRARY_PATH to have the
location of the shared libraries on the remote nodes and this will
automatically be forwarded to the remote nodes.
--
--
mpirun noticed that the job aborted, but has no info as to the process
that caused that situation.
--
mpirun: clean termination accomplished
-


Can anyone help me, please ?


Sincerly


Mouhamad Al sayed ali
___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users




[OMPI users] no allocated resources for the application........(mpirun)

2011-07-19 Thread Mouhamad Al-Sayed-Ali

Hello all,

 I have been trying to run the executable "arpege" with

mpirun -np 2 --host file arpege


where file contains the name of the machines. But, I get the following error:

-
--
There are no allocated resources for the application
  arpege
that match the requested mapping:


Verify that you have mapped the allocated resources properly using the
--host or --hostfile specification.
--
--
A daemon (pid unknown) died unexpectedly on signal 1  while attempting to
launch so we are aborting.

There may be more information reported by the environment (see above).

This may be because the daemon was unable to find all the needed shared
libraries on the remote node. You may set your LD_LIBRARY_PATH to have the
location of the shared libraries on the remote nodes and this will
automatically be forwarded to the remote nodes.
--
--
mpirun noticed that the job aborted, but has no info as to the process
that caused that situation.
--
mpirun: clean termination accomplished
-


Can anyone help me, please ?


Sincerly


Mouhamad Al sayed ali


Re: [OMPI users] problems with parallel MPI-IO...

2011-07-19 Thread Alexander Beck-Ratzka
On Tuesday, July 19, 2011 16:13:38 Jonathan Dursi wrote:
> On 19 Jul 9:02AM, Alexander Beck-Ratzka wrote:
> >integer ierr, i, myrank, BUFSIZE, thefile, intsize
> >parameter (BUFSIZE=100)
> >integer buf(BUFSIZE)
> >
> >do i=0,BUFSIZE
> >
> >   buf(i) = myrank*BUFSIZE + i
> >   print*, 'i =', i, 'myrank =', myrank, 'buf(i)=',buf(i)
> >
> >end do
> 
> [...]
> 
> > When I am reading the data in again and print them out, I always have:
> > 
> > buf(0)=0
> 
> If you compile your code with -check bounds and run, you'll get an error
> pointing out that buf(0) is an illegal access; in Fortran arrays start at
> 1.
> 

Thanks a lot, that solved the problem!

Best wishes

Alexander


Re: [OMPI users] problems with parallel MPI-IO...

2011-07-19 Thread Jonathan Dursi

On 19 Jul 9:02AM, Alexander Beck-Ratzka wrote:

   integer ierr, i, myrank, BUFSIZE, thefile, intsize
   parameter (BUFSIZE=100)
   integer buf(BUFSIZE)

   do i=0,BUFSIZE
  buf(i) = myrank*BUFSIZE + i
  print*, 'i =', i, 'myrank =', myrank, 'buf(i)=',buf(i)
   end do

[...]

When I am reading the data in again and print them out, I always have:

buf(0)=0


If you compile your code with -check bounds and run, you'll get an error 
pointing out that buf(0) is an illegal access; in Fortran arrays start at 1.


- Jonathan
--
Jonathan Dursi | SciNet, Compute/Calcul Canada


[OMPI users] problems with parallel MPI-IO...

2011-07-19 Thread Alexander Beck-Ratzka
Hi Folks,

I am using the following f90 example program for writing a file in parallel 
with MIP:

[snip]
  program parallel_io

  use mpi

  implicit none

  integer ierr, i, myrank, BUFSIZE, thefile, intsize
  parameter (BUFSIZE=100)
  integer buf(BUFSIZE)
  !  integer (kind=MPI_OFFSET_KIND) disp
  integer*8 disp

  call mpi_init(ierr)
  call mpi_comm_rank(mpi_comm_world, myrank,ierr)

  do i=0,BUFSIZE
 buf(i) = myrank*BUFSIZE + i
 print*, 'i =', i, 'myrank =', myrank, 'buf(i)=',buf(i)
  end do

  call mpi_file_open(mpi_comm_world, 'testfile', MPI_MODE_WRONLY + 
MPI_MODE_CREATE, mpi_info_null, thefile, ierr)
  call mpi_type_size(MPI_INTEGER, intsize, ierr)

  disp = myrank * BUFSIZE * intsize

  !  call mpi_file_set_view(thefile, disp, MPI_INTEGER, MPI_INTEGER, 'native', 
mpi_info_null, ierr)
  call mpi_file_write_at(thefile, disp, buf, BUFSIZE, MPI_INTEGER, 
mpi_status_ignore, ierr)

  call mpi_file_close(thefile, ierr)
  call mpi_finalize(ierr)


end program parallel_io
[snip]


And the follwoing program shall read all the data in again, and print them out

[snip]
  program parallel_read_io

  use mpi

  implicit none

  integer ierr, i, myrank, BUFSIZE, thefile, intsize
  parameter (BUFSIZE=100)
  integer buf(BUFSIZE)
  !  integer (kind=MPI_OFFSET_KIND) disp
  integer*8 disp

  call mpi_init(ierr)
  call mpi_comm_rank(mpi_comm_world, myrank,ierr)

  !  do i=0,BUFSIZE
  ! buf(i) = myrank*BUFSIZE + i
  !  end do

  call mpi_file_open(mpi_comm_world, 'testfile', MPI_MODE_RDONLY, 
mpi_info_null, 
thefile, ierr)
  call mpi_type_size(MPI_INTEGER, intsize, ierr)

  disp = myrank * BUFSIZE * intsize

  !  call mpi_file_set_view(thefile, disp, MPI_INTEGER, MPI_INTEGER, 'native', 
mpi_info_null, ierr)
  !  call mpi_file_read(thefile, buf, BUFSIZE, MPI_INTEGER, mpi_status_ignore, 
ierr)
  call mpi_file_read_at(thefile, disp, buf, BUFSIZE, MPI_INTEGER, 
mpi_status_ignore, ierr)

  call mpi_file_close(thefile, ierr)

  !  print the data read in...

  if (myrank.eq.1) then
 do i = 0,BUFSIZE
print*, 'i =', i, 'myrank =', myrank, 'buf(i)=', buf(i)
 end do
  endif

  call mpi_finalize(ierr)
[snip]

I have maid several tests, also with do loops only from 0 to (BUFSIZE-1), with 
MPI_FILE_SET_VIEW and MPI_FILE_READ, etc...

When I am reading the data in again and print them out, I always have:

buf(0)=0

for every rank, so I assume that something with the offset is wrong. I am using 
openmpi with an Intel f90 compiler.

What am I doing wrong?

Best wishes

Alexander


Re: [OMPI users] InfiniBand, different OpenFabrics transport types

2011-07-19 Thread Bill Johnstone
Yevgeny,

Sorry for the delay in replying -- I'd been out for a few days.



- Original Message -
> From: Yevgeny Kliteynik 
> Sent: Thursday, July 14, 2011 12:51 AM
> Subject: Re: [OMPI users] InfiniBand, different OpenFabrics transport types
 
> While I'm trying to find an old HCA somewhere, could you please
> post here the output of "ibv_devinfo -v" on mthca?

:~$ ibv_devinfo -v
hca_id:    mthca0
    transport:            InfiniBand (0)
    fw_ver:                4.8.917
    node_guid:            0005:ad00:000b:60c0
    sys_image_guid:            0005:ad00:0100:d050
    vendor_id:            0x05ad
    vendor_part_id:            25208
    hw_ver:                0xA0
    board_id:            MT_00A001
    phys_port_cnt:            2
    max_mr_size:            0x
    page_size_cap:            0xf000
    max_qp:                64512
    max_qp_wr:            65535
    device_cap_flags:        0x1c76
    max_sge:            59
    max_sge_rd:            0
    max_cq:                65408
    max_cqe:            131071
    max_mr:                131056
    max_pd:                32768
    max_qp_rd_atom:            4
    max_ee_rd_atom:            0
    max_res_rd_atom:        258048
    max_qp_init_rd_atom:        128
    max_ee_init_rd_atom:        0
    atomic_cap:            ATOMIC_HCA (1)
    max_ee:                0
    max_rdd:            0
    max_mw:                0
    max_raw_ipv6_qp:        0
    max_raw_ethy_qp:        0
    max_mcast_grp:            8192
    max_mcast_qp_attach:        56
    max_total_mcast_qp_attach:    458752
    max_ah:                0
    max_fmr:            0
    max_srq:            960
    max_srq_wr:            65535
    max_srq_sge:            31
    max_pkeys:            64
    local_ca_ack_delay:        15
        port:    1
            state:            PORT_ACTIVE (4)
            max_mtu:        2048 (4)
            active_mtu:        2048 (4)
            sm_lid:            2
            port_lid:        49
            port_lmc:        0x00
            link_layer:        IB
            max_msg_sz:        0x8000
            port_cap_flags:        0x02510a68
            max_vl_num:        8 (4)
            bad_pkey_cntr:        0x0
            qkey_viol_cntr:        0x0
            sm_sl:            0
            pkey_tbl_len:        64
            gid_tbl_len:        32
            subnet_timeout:        8
            init_type_reply:    0
            active_width:        4X (2)
            active_speed:        2.5 Gbps (1)
            phys_state:        LINK_UP (5)
            GID[  0]:        fe80::::0005:ad00:000b:60c1

        port:    2
            state:            PORT_DOWN (1)
            max_mtu:        2048 (4)
            active_mtu:        512 (2)
            sm_lid:            0
            port_lid:        0
            port_lmc:        0x00
            link_layer:        IB
            max_msg_sz:        0x8000
            port_cap_flags:        0x02510a68
            max_vl_num:        8 (4)
            bad_pkey_cntr:        0x0
            qkey_viol_cntr:        0x0
            sm_sl:            0
            pkey_tbl_len:        64
            gid_tbl_len:        32
            subnet_timeout:        0
            init_type_reply:    0
            active_width:        4X (2)
            active_speed:        2.5 Gbps (1)
            phys_state:        POLLING (2)
            GID[  0]:        fe80::::0005:ad00:000b:60c2