reassign 625832 libopenmpi1.3 thanks Since hangs in non-python programs it has to do with a generic install/configuration of openmpi, thus reassigning to openmpi.
On Sun, 08 May 2011, Juha Jäykkä wrote: > Let's see... > > python -c 'import mpi4py.MPI' > No use, hangs similarly. > > 2. where does it hang? (according to strace or may be gdb)? > Strace and gdb trace attached. Here are the relevant parts of ps and netstat > outpot, too. > orted --hnp --set-sid --report-uri 7 --singleton-died-pipe 8 > tcp 0 0 *:60561 *:* LISTEN > > 6238/orted > tcp6 0 0 [::]:34267 [::]:* LISTEN > > 6238/orted > > 3. does any other (non-Python) application built against openmpi ran > > with opterun hangs? (i.e. is it configured properly, etc) > Hangs in exaxtly the same place: > #0 0x00007ffff690e668 in __poll (fds=0x629ac0, nfds=4, timeout=<value > optimized out>) > at ../sysdeps/unix/sysv/linux/poll.c:83 > #1 0x00007ffff76a3921 in ?? () from /usr/lib/libopen-pal.so.0 > #2 0x00007ffff76a28cf in ?? () from /usr/lib/libopen-pal.so.0 > #3 0x00007ffff76970b1 in opal_progress () from /usr/lib/libopen-pal.so.0 > #4 0x00007ffff5a2c7e5 in ?? () from > /usr/lib/openmpi/lib/openmpi/mca_rml_oob.so > #5 0x00007ffff5a2ccb0 in ?? () from > /usr/lib/openmpi/lib/openmpi/mca_rml_oob.so > #6 0x00007ffff7915986 in orte_routed_base_register_sync () from > /usr/lib/libopen-rte.so.0 > #7 0x00007ffff60384ce in ?? () from > /usr/lib/openmpi/lib/openmpi/mca_routed_binomial.so > #8 0x00007ffff7902652 in orte_ess_base_app_setup () from /usr/lib/libopen- > rte.so.0 > #9 0x00007ffff5c30fa1 in ?? () from > /usr/lib/openmpi/lib/openmpi/mca_ess_singleton.so > #10 0x00007ffff78e9ba3 in orte_init () from /usr/lib/libopen-rte.so.0 > #11 0x00007ffff7b62397 in ?? () from /usr/lib/libmpi.so.0 > #12 0x00007ffff7b833e0 in PMPI_Init () from /usr/lib/libmpi.so.0 > #13 0x0000000000400916 in main (argc=1, argv=0x7fffffffddc8) at hello.c:5 > That result is independent of number of ranks and whether I start the program > with or without orterun. The hello.c is very short: > #include <mpi.h> > #include <stdio.h> > int main(int argc, char **argv) { > int rc, id; > rc=MPI_Init(&argc, &argv); > rc=MPI_Comm_rank(MPI_COMM_WORLD, &id); > printf("My id = %i\n", id); > MPI_Finalize(); > } > And, as you see from the backtrace, it hangs in MPI_Init(). > Does this mean that the problem is not actually mpi4py, but openmpi (or some > such) instead? > > altogether it might be simply a generic misconfiguration of mpi. If we > I do not believe this. I never configured openmpi AT ALL, every single > openmpi > setting is pristine, as they came out of the Debian package. I never had to > touch anything - which is precisely the reason I started using openmpi all > those years ago, when mpich needed tweaking to get it running at all. > Cheers, > Juha -- =------------------------------------------------------------------= Keep in touch www.onerussian.com Yaroslav Halchenko www.ohloh.net/accounts/yarikoptic -- To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org