reassign 625832 libopenmpi1.3
thanks

Since hangs in non-python programs it has to do with a  generic
install/configuration of openmpi, thus reassigning to openmpi.


On Sun, 08 May 2011, Juha Jäykkä wrote:

> Let's see...

> > python -c 'import mpi4py.MPI'

> No use, hangs similarly.

> > 2. where does it hang? (according to strace or may be gdb)?

> Strace and gdb trace attached. Here are the relevant parts of ps and netstat 
> outpot, too.

> orted --hnp --set-sid --report-uri 7 --singleton-died-pipe 8

> tcp        0      0 *:60561                 *:*                     LISTEN    
>   
> 6238/orted      
> tcp6       0      0 [::]:34267              [::]:*                  LISTEN    
>   
> 6238/orted      

> > 3. does any other (non-Python) application  built against openmpi ran
> >    with opterun hangs? (i.e. is it configured properly, etc)

> Hangs in exaxtly the same place:

> #0  0x00007ffff690e668 in __poll (fds=0x629ac0, nfds=4, timeout=<value 
> optimized out>)
>     at ../sysdeps/unix/sysv/linux/poll.c:83
> #1  0x00007ffff76a3921 in ?? () from /usr/lib/libopen-pal.so.0
> #2  0x00007ffff76a28cf in ?? () from /usr/lib/libopen-pal.so.0
> #3  0x00007ffff76970b1 in opal_progress () from /usr/lib/libopen-pal.so.0
> #4  0x00007ffff5a2c7e5 in ?? () from 
> /usr/lib/openmpi/lib/openmpi/mca_rml_oob.so
> #5  0x00007ffff5a2ccb0 in ?? () from 
> /usr/lib/openmpi/lib/openmpi/mca_rml_oob.so
> #6  0x00007ffff7915986 in orte_routed_base_register_sync () from 
> /usr/lib/libopen-rte.so.0
> #7  0x00007ffff60384ce in ?? () from 
> /usr/lib/openmpi/lib/openmpi/mca_routed_binomial.so
> #8  0x00007ffff7902652 in orte_ess_base_app_setup () from /usr/lib/libopen-
> rte.so.0
> #9  0x00007ffff5c30fa1 in ?? () from 
> /usr/lib/openmpi/lib/openmpi/mca_ess_singleton.so
> #10 0x00007ffff78e9ba3 in orte_init () from /usr/lib/libopen-rte.so.0
> #11 0x00007ffff7b62397 in ?? () from /usr/lib/libmpi.so.0
> #12 0x00007ffff7b833e0 in PMPI_Init () from /usr/lib/libmpi.so.0
> #13 0x0000000000400916 in main (argc=1, argv=0x7fffffffddc8) at hello.c:5

> That result is independent of number of ranks and whether I start the program 
> with or without orterun. The hello.c is very short:

> #include <mpi.h>
> #include <stdio.h>
> int main(int argc, char **argv) {
>   int rc, id;  
>   rc=MPI_Init(&argc, &argv);
>   rc=MPI_Comm_rank(MPI_COMM_WORLD, &id);
>   printf("My id = %i\n", id);
>   MPI_Finalize();
> }

> And, as you see from the backtrace, it hangs in MPI_Init().

> Does this mean that the problem is not actually mpi4py, but openmpi (or some 
> such) instead?

> > altogether it might be simply a generic misconfiguration of mpi.  If we

> I do not believe this. I never configured openmpi AT ALL, every single 
> openmpi 
> setting is pristine, as they came out of the Debian package. I never had to 
> touch anything - which is precisely the reason I started using openmpi all 
> those years ago, when mpich needed tweaking to get it running at all.

> Cheers,
> Juha
-- 
=------------------------------------------------------------------=
Keep in touch                                     www.onerussian.com
Yaroslav Halchenko                 www.ohloh.net/accounts/yarikoptic



-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org

Reply via email to