The real problem is not the -g it is the -O0 option which will be automatically added by -g. If you compile with "-g -ON" for 0 < N everythings works as expected.
Thanks, Sven On Friday 11 August 2006 11:54, Bettina Krammer wrote: > Hi, > > when I use the attached hello.f with Open MPI 1.1.0 and underlying Intel > 9.0 or 9.1 compiler on our Xeon cluster, it is deadlocking when compiled > with -g option but works without -g: > > =================== > output with -g: > > $mpirun -np 2 ./hello-g > > My rank is 0 ! > waiting for message from 1 > My rank is 1 ! > Greetings from process 1 ! > Sending message from 1 ! > Message recieved: HelloFromMexxxxxxxxx! > waiting for message from 1 > > [...deadlock...] > =================== > > output without -g: > > $mpirun -np 2 ./hello-no-g > > My rank is 0 ! > waiting for message from 1 > My rank is 1 ! > Greetings from process 1 ! > Sending message from 1 ! > Message recieved: HelloFromMexxxxxxxxx! > All done... 0 > All done... 1 > =================== > > Thanks, Bettina Krammer > > (The example is taken from the distribution of DDT, to be found in > ddt/examples. The problem is reproducible with the simplified > hello-simple.f. The deadlock occurs in the DO source... MPI_Recv(...) > .... loop) > =================== > The config.log is not available to me. > > hpc43203 cacau1 219$ompi_info > Open MPI: 1.1 > Open MPI SVN revision: r10477 > Open RTE: 1.1 > Open RTE SVN revision: r10477 > OPAL: 1.1 > OPAL SVN revision: r10477 > Prefix: /opt/OpenMPI/1.1.0/ > Configured architecture: x86_64-unknown-linux-gnu > Configured by: hpcraink > Configured on: Mon Jul 31 12:55:30 CEST 2006 > Configure host: cacau1 > Built by: hpcraink > Built on: Mon Jul 31 13:16:04 CEST 2006 > Built host: cacau1 > C bindings: yes > C++ bindings: yes > Fortran77 bindings: yes (all) > Fortran90 bindings: yes > Fortran90 bindings size: small > C compiler: icc > C compiler absolute: /opt/intel/compiler/9.1/cce/bin/icc > C++ compiler: icpc > C++ compiler absolute: /opt/intel/compiler/9.1/cce/bin/icpc > Fortran77 compiler: ifc > Fortran77 compiler abs: /opt/intel/compiler/9.1/fce/bin/ifc > Fortran90 compiler: ifc > Fortran90 compiler abs: /opt/intel/compiler/9.1/fce/bin/ifc > C profiling: yes > C++ profiling: yes > Fortran77 profiling: yes > Fortran90 profiling: yes > C++ exceptions: no > Thread support: posix (mpi: no, progress: no) > Internal debug support: no > MPI parameter check: runtime > Memory profiling support: no > Memory debugging support: no > libltdl support: yes > MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.1) > MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.1) > MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.1) > MCA maffinity: libnuma (MCA v1.0, API v1.0, Component v1.1) > MCA timer: linux (MCA v1.0, API v1.0, Component v1.1) > MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0) > MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0) > MCA coll: basic (MCA v1.0, API v1.0, Component v1.1) > MCA coll: hierarch (MCA v1.0, API v1.0, Component v1.1) > MCA coll: self (MCA v1.0, API v1.0, Component v1.1) > MCA coll: sm (MCA v1.0, API v1.0, Component v1.1) > MCA coll: tuned (MCA v1.0, API v1.0, Component v1.1) > MCA io: romio (MCA v1.0, API v1.0, Component v1.1) > MCA mpool: sm (MCA v1.0, API v1.0, Component v1.1) > MCA mpool: mvapi (MCA v1.0, API v1.0, Component v1.1) > MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.1) > MCA bml: r2 (MCA v1.0, API v1.0, Component v1.1) > MCA rcache: rb (MCA v1.0, API v1.0, Component v1.1) > MCA btl: self (MCA v1.0, API v1.0, Component v1.1) > MCA btl: sm (MCA v1.0, API v1.0, Component v1.1) > MCA btl: mvapi (MCA v1.0, API v1.0, Component v1.1) > MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0) > MCA topo: unity (MCA v1.0, API v1.0, Component v1.1) > MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.0) > MCA gpr: null (MCA v1.0, API v1.0, Component v1.1) > MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.1) > MCA gpr: replica (MCA v1.0, API v1.0, Component v1.1) > MCA iof: proxy (MCA v1.0, API v1.0, Component v1.1) > MCA iof: svc (MCA v1.0, API v1.0, Component v1.1) > MCA ns: proxy (MCA v1.0, API v1.0, Component v1.1) > MCA ns: replica (MCA v1.0, API v1.0, Component v1.1) > MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0) > MCA ras: dash_host (MCA v1.0, API v1.0, Component v1.1) > MCA ras: hostfile (MCA v1.0, API v1.0, Component v1.1) > MCA ras: localhost (MCA v1.0, API v1.0, Component v1.1) > MCA ras: slurm (MCA v1.0, API v1.0, Component v1.1) > MCA ras: tm (MCA v1.0, API v1.0, Component v1.1) > MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.1) > MCA rds: resfile (MCA v1.0, API v1.0, Component v1.1) > MCA rmaps: round_robin (MCA v1.0, API v1.0, Component v1.1) > MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.1) > MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.1) > MCA rml: oob (MCA v1.0, API v1.0, Component v1.1) > MCA pls: fork (MCA v1.0, API v1.0, Component v1.1) > MCA pls: rsh (MCA v1.0, API v1.0, Component v1.1) > MCA pls: slurm (MCA v1.0, API v1.0, Component v1.1) > MCA pls: tm (MCA v1.0, API v1.0, Component v1.1) > MCA sds: env (MCA v1.0, API v1.0, Component v1.1) > MCA sds: seed (MCA v1.0, API v1.0, Component v1.1) > MCA sds: singleton (MCA v1.0, API v1.0, Component v1.1) > MCA sds: pipe (MCA v1.0, API v1.0, Component v1.1) > MCA sds: slurm (MCA v1.0, API v1.0, Component v1.1) > > > -- > --------------------------------------------- > * NEW PHONE AND FAX-NUMBERS * > --------------------------------------------- > Dipl.-Math. Bettina Krammer > High Performance Computing Center (HLRS) > University of Stuttgart > Nobelstrasse 19 > D-70569 Stuttgart > > Phone: ++49 (0)711-685-65890 > Fax: ++49 (0)711-685-65832 > email: kram...@hlrs.de > URL: http://www.hlrs.de > --------------------------------------------- >