I'm not quite sure I understand -- does the application hang in an MPI call?
Or is there some compiler error that is causing it to execute a DO loop
incorrectly?


On 8/11/06 6:25 AM, "Sven Stork" <st...@hlrs.de> wrote:

> The real problem is not the -g it is the -O0 option which will be
> automatically added by -g. If you compile with "-g -ON" for 0 < N everythings
> works as expected.
> 
> Thanks,
> Sven
> 
> On Friday 11 August 2006 11:54, Bettina Krammer wrote:
>> Hi,
>> 
>> when I use the attached hello.f with Open MPI 1.1.0 and underlying Intel
>> 9.0 or 9.1 compiler on our Xeon cluster, it is deadlocking when compiled
>> with -g option but works without -g:
>> 
>> ===================
>> output with -g:
>> 
>> $mpirun -np 2 ./hello-g
>> 
>> My rank is            0 !
>> waiting for message from            1
>> My rank is            1 !
>> Greetings from process            1 !
>> Sending message from            1 !
>> Message recieved: HelloFromMexxxxxxxxx!
>> waiting for message from            1
>> 
>>       [...deadlock...]
>> ===================
>> 
>>  output without -g:
>> 
>> $mpirun -np 2 ./hello-no-g
>> 
>> My rank is            0 !
>>  waiting for message from            1
>>  My rank is            1 !
>>  Greetings from process            1 !
>>  Sending message from            1 !
>>  Message recieved: HelloFromMexxxxxxxxx!
>>  All done...           0
>>  All done...           1
>> ===================
>> 
>> Thanks, Bettina Krammer
>> 
>> (The example is taken from the distribution of DDT, to be found in
>> ddt/examples. The problem is reproducible with the simplified
>> hello-simple.f. The deadlock occurs in the DO source... MPI_Recv(...)
>> .... loop)
>> ===================
>> The config.log is not available to me.
>> 
>> hpc43203 cacau1 219$ompi_info
>>                 Open MPI: 1.1
>>    Open MPI SVN revision: r10477
>>                 Open RTE: 1.1
>>    Open RTE SVN revision: r10477
>>                     OPAL: 1.1
>>        OPAL SVN revision: r10477
>>                   Prefix: /opt/OpenMPI/1.1.0/
>>  Configured architecture: x86_64-unknown-linux-gnu
>>            Configured by: hpcraink
>>            Configured on: Mon Jul 31 12:55:30 CEST 2006
>>           Configure host: cacau1
>>                 Built by: hpcraink
>>                 Built on: Mon Jul 31 13:16:04 CEST 2006
>>               Built host: cacau1
>>               C bindings: yes
>>             C++ bindings: yes
>>       Fortran77 bindings: yes (all)
>>       Fortran90 bindings: yes
>>  Fortran90 bindings size: small
>>               C compiler: icc
>>      C compiler absolute: /opt/intel/compiler/9.1/cce/bin/icc
>>             C++ compiler: icpc
>>    C++ compiler absolute: /opt/intel/compiler/9.1/cce/bin/icpc
>>       Fortran77 compiler: ifc
>>   Fortran77 compiler abs: /opt/intel/compiler/9.1/fce/bin/ifc
>>       Fortran90 compiler: ifc
>>   Fortran90 compiler abs: /opt/intel/compiler/9.1/fce/bin/ifc
>>              C profiling: yes
>>            C++ profiling: yes
>>      Fortran77 profiling: yes
>>      Fortran90 profiling: yes
>>           C++ exceptions: no
>>           Thread support: posix (mpi: no, progress: no)
>>   Internal debug support: no
>>      MPI parameter check: runtime
>> Memory profiling support: no
>> Memory debugging support: no
>>          libltdl support: yes
>>               MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.1)
>>            MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.1)
>>            MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.1)
>>            MCA maffinity: libnuma (MCA v1.0, API v1.0, Component v1.1)
>>                MCA timer: linux (MCA v1.0, API v1.0, Component v1.1)
>>            MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
>>            MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
>>                 MCA coll: basic (MCA v1.0, API v1.0, Component v1.1)
>>                 MCA coll: hierarch (MCA v1.0, API v1.0, Component v1.1)
>>                 MCA coll: self (MCA v1.0, API v1.0, Component v1.1)
>>                 MCA coll: sm (MCA v1.0, API v1.0, Component v1.1)
>>                 MCA coll: tuned (MCA v1.0, API v1.0, Component v1.1)
>>                   MCA io: romio (MCA v1.0, API v1.0, Component v1.1)
>>                MCA mpool: sm (MCA v1.0, API v1.0, Component v1.1)
>>                MCA mpool: mvapi (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA bml: r2 (MCA v1.0, API v1.0, Component v1.1)
>>               MCA rcache: rb (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA btl: self (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA btl: sm (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA btl: mvapi (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
>>                 MCA topo: unity (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.0)
>>                  MCA gpr: null (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA gpr: replica (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA iof: proxy (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA iof: svc (MCA v1.0, API v1.0, Component v1.1)
>>                   MCA ns: proxy (MCA v1.0, API v1.0, Component v1.1)
>>                   MCA ns: replica (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
>>                  MCA ras: dash_host (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA ras: hostfile (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA ras: localhost (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA ras: slurm (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA ras: tm (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA rds: resfile (MCA v1.0, API v1.0, Component v1.1)
>>                MCA rmaps: round_robin (MCA v1.0, API v1.0, Component v1.1)
>>                 MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.1)
>>                 MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA rml: oob (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA pls: fork (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA pls: rsh (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA pls: slurm (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA pls: tm (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA sds: env (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA sds: seed (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA sds: singleton (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA sds: pipe (MCA v1.0, API v1.0, Component v1.1)
>>                  MCA sds: slurm (MCA v1.0, API v1.0, Component v1.1)
>> 
>> 
>> -- 
>> ---------------------------------------------
>> * NEW PHONE AND FAX-NUMBERS *
>> ---------------------------------------------
>> Dipl.-Math. Bettina Krammer
>> High Performance Computing Center (HLRS)
>> University of Stuttgart
>> Nobelstrasse 19
>> D-70569 Stuttgart
>> 
>> Phone: ++49 (0)711-685-65890
>> Fax: ++49 (0)711-685-65832
>> email: kram...@hlrs.de
>> URL: http://www.hlrs.de
>> ---------------------------------------------
>> 
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel


-- 
Jeff Squyres
Server Virtualization Business Unit
Cisco Systems

Reply via email to