The real problem is not the -g it is the -O0 option which will be 
automatically added by -g. If you compile with "-g -ON" for 0 < N everythings 
works as expected.

Thanks,
Sven

On Friday 11 August 2006 11:54, Bettina Krammer wrote:
> Hi,
> 
> when I use the attached hello.f with Open MPI 1.1.0 and underlying Intel 
> 9.0 or 9.1 compiler on our Xeon cluster, it is deadlocking when compiled 
> with -g option but works without -g:
> 
> ===================
> output with -g:
> 
> $mpirun -np 2 ./hello-g
> 
> My rank is            0 !
> waiting for message from            1
> My rank is            1 !
> Greetings from process            1 !
> Sending message from            1 !
> Message recieved: HelloFromMexxxxxxxxx!
> waiting for message from            1
> 
>       [...deadlock...]
> ===================
> 
>  output without -g:
> 
> $mpirun -np 2 ./hello-no-g
> 
> My rank is            0 !
>  waiting for message from            1
>  My rank is            1 !
>  Greetings from process            1 !
>  Sending message from            1 !
>  Message recieved: HelloFromMexxxxxxxxx!
>  All done...           0
>  All done...           1
> ===================
> 
> Thanks, Bettina Krammer
> 
> (The example is taken from the distribution of DDT, to be found in 
> ddt/examples. The problem is reproducible with the simplified 
> hello-simple.f. The deadlock occurs in the DO source... MPI_Recv(...) 
> .... loop)
> ===================
> The config.log is not available to me.
> 
> hpc43203 cacau1 219$ompi_info
>                 Open MPI: 1.1
>    Open MPI SVN revision: r10477
>                 Open RTE: 1.1
>    Open RTE SVN revision: r10477
>                     OPAL: 1.1
>        OPAL SVN revision: r10477
>                   Prefix: /opt/OpenMPI/1.1.0/
>  Configured architecture: x86_64-unknown-linux-gnu
>            Configured by: hpcraink
>            Configured on: Mon Jul 31 12:55:30 CEST 2006
>           Configure host: cacau1
>                 Built by: hpcraink
>                 Built on: Mon Jul 31 13:16:04 CEST 2006
>               Built host: cacau1
>               C bindings: yes
>             C++ bindings: yes
>       Fortran77 bindings: yes (all)
>       Fortran90 bindings: yes
>  Fortran90 bindings size: small
>               C compiler: icc
>      C compiler absolute: /opt/intel/compiler/9.1/cce/bin/icc
>             C++ compiler: icpc
>    C++ compiler absolute: /opt/intel/compiler/9.1/cce/bin/icpc
>       Fortran77 compiler: ifc
>   Fortran77 compiler abs: /opt/intel/compiler/9.1/fce/bin/ifc
>       Fortran90 compiler: ifc
>   Fortran90 compiler abs: /opt/intel/compiler/9.1/fce/bin/ifc
>              C profiling: yes
>            C++ profiling: yes
>      Fortran77 profiling: yes
>      Fortran90 profiling: yes
>           C++ exceptions: no
>           Thread support: posix (mpi: no, progress: no)
>   Internal debug support: no
>      MPI parameter check: runtime
> Memory profiling support: no
> Memory debugging support: no
>          libltdl support: yes
>               MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.1)
>            MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.1)
>            MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.1)
>            MCA maffinity: libnuma (MCA v1.0, API v1.0, Component v1.1)
>                MCA timer: linux (MCA v1.0, API v1.0, Component v1.1)
>            MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
>            MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
>                 MCA coll: basic (MCA v1.0, API v1.0, Component v1.1)
>                 MCA coll: hierarch (MCA v1.0, API v1.0, Component v1.1)
>                 MCA coll: self (MCA v1.0, API v1.0, Component v1.1)
>                 MCA coll: sm (MCA v1.0, API v1.0, Component v1.1)
>                 MCA coll: tuned (MCA v1.0, API v1.0, Component v1.1)
>                   MCA io: romio (MCA v1.0, API v1.0, Component v1.1)
>                MCA mpool: sm (MCA v1.0, API v1.0, Component v1.1)
>                MCA mpool: mvapi (MCA v1.0, API v1.0, Component v1.1)
>                  MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.1)
>                  MCA bml: r2 (MCA v1.0, API v1.0, Component v1.1)
>               MCA rcache: rb (MCA v1.0, API v1.0, Component v1.1)
>                  MCA btl: self (MCA v1.0, API v1.0, Component v1.1)
>                  MCA btl: sm (MCA v1.0, API v1.0, Component v1.1)
>                  MCA btl: mvapi (MCA v1.0, API v1.0, Component v1.1)
>                  MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
>                 MCA topo: unity (MCA v1.0, API v1.0, Component v1.1)
>                  MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.0)
>                  MCA gpr: null (MCA v1.0, API v1.0, Component v1.1)
>                  MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.1)
>                  MCA gpr: replica (MCA v1.0, API v1.0, Component v1.1)
>                  MCA iof: proxy (MCA v1.0, API v1.0, Component v1.1)
>                  MCA iof: svc (MCA v1.0, API v1.0, Component v1.1)
>                   MCA ns: proxy (MCA v1.0, API v1.0, Component v1.1)
>                   MCA ns: replica (MCA v1.0, API v1.0, Component v1.1)
>                  MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
>                  MCA ras: dash_host (MCA v1.0, API v1.0, Component v1.1)
>                  MCA ras: hostfile (MCA v1.0, API v1.0, Component v1.1)
>                  MCA ras: localhost (MCA v1.0, API v1.0, Component v1.1)
>                  MCA ras: slurm (MCA v1.0, API v1.0, Component v1.1)
>                  MCA ras: tm (MCA v1.0, API v1.0, Component v1.1)
>                  MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.1)
>                  MCA rds: resfile (MCA v1.0, API v1.0, Component v1.1)
>                MCA rmaps: round_robin (MCA v1.0, API v1.0, Component v1.1)
>                 MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.1)
>                 MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.1)
>                  MCA rml: oob (MCA v1.0, API v1.0, Component v1.1)
>                  MCA pls: fork (MCA v1.0, API v1.0, Component v1.1)
>                  MCA pls: rsh (MCA v1.0, API v1.0, Component v1.1)
>                  MCA pls: slurm (MCA v1.0, API v1.0, Component v1.1)
>                  MCA pls: tm (MCA v1.0, API v1.0, Component v1.1)
>                  MCA sds: env (MCA v1.0, API v1.0, Component v1.1)
>                  MCA sds: seed (MCA v1.0, API v1.0, Component v1.1)
>                  MCA sds: singleton (MCA v1.0, API v1.0, Component v1.1)
>                  MCA sds: pipe (MCA v1.0, API v1.0, Component v1.1)
>                  MCA sds: slurm (MCA v1.0, API v1.0, Component v1.1)
> 
> 
> -- 
> ---------------------------------------------
> * NEW PHONE AND FAX-NUMBERS *
> ---------------------------------------------
> Dipl.-Math. Bettina Krammer
> High Performance Computing Center (HLRS)
> University of Stuttgart
> Nobelstrasse 19
> D-70569 Stuttgart
> 
> Phone: ++49 (0)711-685-65890
> Fax: ++49 (0)711-685-65832
> email: kram...@hlrs.de
> URL: http://www.hlrs.de
> ---------------------------------------------
> 

Reply via email to