No, it's using 2 nodes.
  George.

On Tue, Oct 27, 2015 at 12:35 PM, Ralph Castain <r...@open-mpi.org> wrote:

> Is this on a single node?
>
> On Oct 27, 2015, at 9:25 AM, George Bosilca <bosi...@icl.utk.edu> wrote:
>
> I get intermittent deadlocks wit the latest trunk. The smallest reproducer
> is a shell for loop around a small (2 processes) short (20 seconds) MPI
> application. After few tens of iterations the MPI_Init will deadlock with
> the following backtrace:
>
> #0  0x00007fa94b5d9aed in nanosleep () from /lib64/libc.so.6
> #1  0x00007fa94b60ec94 in usleep () from /lib64/libc.so.6
> #2  0x00007fa94960ba08 in OPAL_PMIX_PMIX1XX_PMIx_Fence (procs=0x0,
> nprocs=0, info=0x7ffd7934fb90,
>     ninfo=1) at src/client/pmix_client_fence.c:100
> #3  0x00007fa9498376a2 in pmix1_fence (procs=0x0, collect_data=1) at
> pmix1_client.c:305
> #4  0x00007fa94bb39ba4 in ompi_mpi_init (argc=3, argv=0x7ffd793500a8,
> requested=3,
>     provided=0x7ffd7934ff94) at runtime/ompi_mpi_init.c:645
> #5  0x00007fa94bb77281 in PMPI_Init_thread (argc=0x7ffd7934ff8c,
> argv=0x7ffd7934ff80, required=3,
>     provided=0x7ffd7934ff94) at pinit_thread.c:69
> #6  0x000000000040150f in main (argc=3, argv=0x7ffd793500a8) at
> osu_mbw_mr.c:86
>
> On my machines this is reproducible at 100% after anywhere between 50 and
> 100 iterations.
>
>   Thanks,
>     George.
>
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2015/10/18280.php
>
>
>
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2015/10/18281.php
>

Reply via email to