I replied a couple days ago (with OMPI users in CC) but got an error
last night:

Action: failed
Status: 5.0.0 (permanent failure)
Diagnostic-Code: smtp; 5.4.7 - Delivery expired (message too old) 'timeout' 
(delivery attempts: 0)


I resent the mail this morning, it looks like it wasn't delivered yet
either?
I'll try again with another SMTP.

Brice



Le 10/09/2012 20:37, Douglas Eadline a écrit :
> I got no response for this question. Is Open-MX
> no longer supported in Open MPI? Or is there someplace else
> I should submit this information? I also attached my ompi_info
> and omx_info output
>
> --
> Doug
>
>
>> I built open-mpi 1.6.1 using the open-mx libraries.
>> This worked previously and now I get the following
>> error. Here is my system:
>>
>> kernel:  2.6.32-279.5.1.el6.x86_64
>> open-mx: 1.5.2
>>
>> BTW, open-mx worked previously with open-mpi and the current
>> version works with mpich2
>>
>>
>> $  mpiexec -np 8 -machinefile machines cpi
>> Process 0 on limulus
>> FatalError: Failed to lookup peer by addr, driver replied Bad file
>> descriptor
>> cpi: ../omx_misc.c:89: omx__ioctl_errno_to_return_checked: Assertion `0'
>> failed.
>> [limulus:04448] *** Process received signal ***
>> [limulus:04448] Signal: Aborted (6)
>> [limulus:04448] Signal code:  (-6)
>> [limulus:04448] [ 0] /lib64/libpthread.so.0() [0x3324e0f500]
>> [limulus:04448] [ 1] /lib64/libc.so.6(gsignal+0x35) [0x33246328a5]
>> [limulus:04448] [ 2] /lib64/libc.so.6(abort+0x175) [0x3324634085]
>> [limulus:04448] [ 3] /lib64/libc.so.6() [0x332462ba1e]
>> [limulus:04448] [ 4] /lib64/libc.so.6(__assert_perror_fail+0)
>> [0x332462bae0]
>> [limulus:04448] [ 5]
>> /usr/open-mx/lib/libopen-mx.so.0(omx__ioctl_errno_to_return_checked+0x197)
>> [0x7fb587418b37]
>> [limulus:04448] [ 6]
>> /usr/open-mx/lib/libopen-mx.so.0(omx__peer_addr_to_index+0x55)
>> [0x7fb58741a5d5]
>> [limulus:04448] [ 7] /usr/open-mx/lib/libopen-mx.so.0(+0xdc7a)
>> [0x7fb587419c7a]
>> [limulus:04448] [ 8] /usr/open-mx/lib/libopen-mx.so.0(omx_connect+0x8c)
>> [0x7fb58741a27c]
>> [limulus:04448] [ 9] /usr/open-mx/lib/libopen-mx.so.0(mx_connect+0x15)
>> [0x7fb587425865]
>> [limulus:04448] [10]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_btl_mx_proc_connect+0x5e)
>> [0x7fb5876fe40e]
>> [limulus:04448] [11]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_btl_mx_send+0x2d4)
>> [0x7fb5876fbd94]
>> [limulus:04448] [12]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_pml_ob1_send_request_start_prepare+0xcb)
>> [0x7fb58777d6fb]
>> [limulus:04448] [13]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_pml_ob1_isend+0x4cb)
>> [0x7fb58777509b]
>> [limulus:04448] [14]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(ompi_coll_tuned_bcast_intra_generic+0x37b)
>> [0x7fb58770b55b]
>> [limulus:04448] [15]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(ompi_coll_tuned_bcast_intra_binomial+0xd8)
>> [0x7fb58770b8b8]
>> [limulus:04448] [16]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(ompi_coll_tuned_bcast_intra_dec_fixed+0xcc)
>> [0x7fb587702d8c]
>> [limulus:04448] [17]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_coll_sync_bcast+0x78)
>> [0x7fb587712e88]
>> [limulus:04448] [18]
>> /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(MPI_Bcast+0x130) [0x7fb5876ce1b0]
>> [limulus:04448] [19] cpi(main+0x10b) [0x400cc4]
>> [limulus:04448] [20] /lib64/libc.so.6(__libc_start_main+0xfd)
>> [0x332461ecdd]
>> [limulus:04448] [21] cpi() [0x400ac9]
>> [limulus:04448] *** End of error message ***
>> Process 2 on limulus
>> Process 4 on limulus
>> Process 6 on limulus
>> Process 1 on n0
>> Process 7 on n0
>> Process 3 on n0
>> Process 5 on n0
>> --------------------------------------------------------------------------
>> mpiexec noticed that process rank 0 with PID 4448 on node limulus exited
>> on signal 6 (Aborted).
>> --------------------------------------------------------------------------
>>
> [root@limulus ~]# ompi_info
>                  Package: Open MPI deadline@limulus Distribution
>                 Open MPI: 1.6.1
>    Open MPI SVN revision: r27106
>    Open MPI release date: Aug 22, 2012
>                 Open RTE: 1.6.1
>    Open RTE SVN revision: r27106
>    Open RTE release date: Aug 22, 2012
>                     OPAL: 1.6.1
>        OPAL SVN revision: r27106
>        OPAL release date: Aug 22, 2012
>                  MPI API: 2.1
>             Ident string: 1.6.1
>                   Prefix: /opt/mpi/openmpi-gnu4-1.6.1
>  Configured architecture: x86_64-unknown-linux-gnu
>           Configure host: limulus
>            Configured by: deadline
>            Configured on: Wed Sep  5 15:03:09 EDT 2012
>           Configure host: limulus
>                 Built by: deadline
>                 Built on: Wed Sep  5 15:21:19 EDT 2012
>               Built host: limulus
>               C bindings: yes
>             C++ bindings: yes
>       Fortran77 bindings: yes (all)
>       Fortran90 bindings: yes
>  Fortran90 bindings size: small
>               C compiler: gcc
>      C compiler absolute: /usr/bin/gcc
>   C compiler family name: GNU
>       C compiler version: 4.4.6
>             C++ compiler: g++
>    C++ compiler absolute: /usr/bin/g++
>       Fortran77 compiler: gfortran
>   Fortran77 compiler abs: /usr/bin/gfortran
>       Fortran90 compiler: /usr/bin/gfortran
>   Fortran90 compiler abs:
>              C profiling: yes
>            C++ profiling: yes
>      Fortran77 profiling: yes
>      Fortran90 profiling: yes
>           C++ exceptions: no
>           Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no)
>            Sparse Groups: no
>   Internal debug support: no
>   MPI interface warnings: no
>      MPI parameter check: runtime
> Memory profiling support: no
> Memory debugging support: no
>          libltdl support: yes
>    Heterogeneous support: no
>  mpirun default --prefix: no
>          MPI I/O support: yes
>        MPI_WTIME support: gettimeofday
>      Symbol vis. support: yes
>    Host topology support: yes
>           MPI extensions: affinity example
>    FT Checkpoint support: no (checkpoint thread: no)
>      VampirTrace support: yes
>   MPI_MAX_PROCESSOR_NAME: 256
>     MPI_MAX_ERROR_STRING: 256
>      MPI_MAX_OBJECT_NAME: 64
>         MPI_MAX_INFO_KEY: 36
>         MPI_MAX_INFO_VAL: 256
>        MPI_MAX_PORT_NAME: 1024
>   MPI_MAX_DATAREP_STRING: 128
>            MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA memory: linux (MCA v2.0, API v2.0, Component v1.6.1)
>            MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA carto: file (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6.1)
>            MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6.1)
>            MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA timer: linux (MCA v2.0, API v2.0, Component v1.6.1)
>          MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6.1)
>          MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6.1)
>              MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6.1)
>            MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6.1)
>            MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA coll: basic (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA coll: inter (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA coll: self (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA coll: sm (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA coll: sync (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6.1)
>                   MCA io: romio (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA pml: cm (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA pml: csum (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA pml: v (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA btl: self (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA btl: mx (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA btl: sm (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA mtl: mx (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA topo: unity (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA iof: orted (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA iof: tool (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6.1)
>                 MCA odls: default (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ras: cm (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ras: gridengine (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA rml: oob (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA routed: cm (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA routed: direct (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA routed: linear (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA routed: radix (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA routed: slave (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6.1)
>                MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6.1)
>               MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ess: env (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ess: slave (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6.1)
>                  MCA ess: tool (MCA v2.0, API v2.0, Component v1.6.1)
>              MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6.1)
>              MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6.1)
>              MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6.1)
>             MCA notifier: command (MCA v2.0, API v1.0, Component v1.6.1)
>             MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6.1)
>
>
> [root@limulus ~]# /usr/open-mx/bin/omx_info
> Open-MX version 1.5.2
>  build: deadline@limulus:/raid1/home/deadline/rpms-sl6/BUILD/open-mx-1.5.2
> Mon Sep 10 08:44:16 EDT 2012
>
> Found 1 boards (32 max) supporting 32 endpoints each:
>  limulus:0 (board #0 name eth0 addr e0:69:95:35:d7:71)
>    managed by driver 'e1000e'
>
> Peer table is ready, mapper is 00:00:00:00:00:00
> ================================================
>   0) e0:69:95:35:d7:71 limulus:0
>   1) 00:1c:c0:9b:66:d0 n1:0
>   2) 00:1a:4d:4a:bf:85 n0:0
>   3) 00:1a:4d:4a:bf:83 n2:0
>
>
>
>
>
>
> --
> Doug
>

Reply via email to