Hi Artem, Actually some time ago there was a known issue with coll ml. I used to run my command lines with -mca coll ^ml to avoid these problems, so I don't know if it was fixed or not. It looks like you have the same problem.
Best regards, Elena On Fri, Oct 17, 2014 at 7:01 PM, Artem Polyakov <artpo...@gmail.com> wrote: > Gilles, > > I checked your patch and it doesn't solve the problem I observe. I think > the reason is somewhere else. > > 2014-10-17 19:13 GMT+07:00 Gilles Gouaillardet < > gilles.gouaillar...@gmail.com>: > >> Artem, >> >> There is a known issue #235 with modex and i made PR #238 with a >> tentative fix. >> >> Could you please give it a try and reports if it solves your problem ? >> >> Cheers >> >> Gilles >> >> >> Artem Polyakov <artpo...@gmail.com> wrote: >> Hello, I have troubles with latest trunk if I use PMI1. >> >> For example, if I use 2 nodes the application hangs. See backtraces from >> both nodes below. From them I can see that second (non launching) node >> hangs in bcol component selection. Here is the default setting of >> bcol_base_string parameter: >> bcol_base_string="basesmuma,basesmuma,iboffload,ptpcoll,ugni" >> according to ompi_info. I don't know if it is correct that basesmuma is >> duplicated or not. >> >> Experiments with this parameter showed that it directly influences the >> bug: >> export OMPI_MCA_bcol_base_string="" # [SEGFAULT] >> export OMPI_MCA_bcol_base_string="ptpcoll" # [OK] >> export OMPI_MCA_bcol_base_string="basesmuma,ptpcoll" # [OK] >> export OMPI_MCA_bcol_base_string="basesmuma,ptpcoll,iboffload" # [OK] >> export OMPI_MCA_bcol_base_string="basesmuma,ptpcoll,iboffload,ugni" # >> [OK] >> export >> OMPI_MCA_bcol_base_string="basesmuma,basesmuma,ptpcoll,iboffload,ugni" # >> [HANG] >> export OMPI_MCA_bcol_base_string="basesmuma,basesmuma,iboffload,ptpcoll" >> # [HANG] >> export OMPI_MCA_bcol_base_string="basesmuma,basesmuma,iboffload" # [OK] >> export OMPI_MCA_bcol_base_string="basesmuma,basesmuma,iboffload,ugni" # >> [OK] >> export OMPI_MCA_bcol_base_string="basesmuma,basesmuma,ptpcoll" # [HANG] >> export OMPI_MCA_bcol_base_string="ptpcoll,basesmuma" # [OK] >> export OMPI_MCA_bcol_base_string="ptpcoll,basesmuma,basesmuma" # [HANG] >> >> I can provide other information if nessesary. >> >> cn1: >> (gdb) bt >> 0 0x00007fdebd30ac6d in poll () from /lib/x86_64-linux-gnu/libc.so.6 >> 1 0x00007fdebcca64e0 in poll_dispatch (base=0x1d466b0, >> tv=0x7fff71aab880) at poll.c:165 >> 2 0x00007fdebcc9b041 in opal_libevent2021_event_base_loop >> (base=0x1d466b0, flags=2) at event.c:1631 >> 3 0x00007fdebcc35891 in opal_progress () at runtime/opal_progress.c:169 >> 4 0x00007fdeb32f78cb in opal_condition_wait (c=0x7fdebdb51bc0 >> <ompi_request_cond>, m=0x7fdebdb51cc0 <ompi_request_lock>) at >> ../../../../opal/threads/condition.h:78 >> 5 0x00007fdeb32f79b8 in ompi_request_wait_completion >> (req=0x7fff71aab920) at ../../../../ompi/request/request.h:381 >> 6 0x00007fdeb32f84b8 in mca_pml_ob1_recv (addr=0x7fff71aabd80, count=1, >> datatype=0x6026c0 <ompi_mpi_int>, src=1, tag=0, comm=0x6020a0 >> <ompi_mpi_comm_world>, >> status=0x7fff71aabd90) at pml_ob1_irecv.c:109 >> 7 0x00007fdebd88f54d in PMPI_Recv (buf=0x7fff71aabd80, count=1, >> type=0x6026c0 <ompi_mpi_int>, source=1, tag=0, comm=0x6020a0 >> <ompi_mpi_comm_world>, >> status=0x7fff71aabd90) at precv.c:78 >> 8 0x0000000000400c44 in main (argc=1, argv=0x7fff71aabe98) at >> hellompi.c:33 >> >> cn2: >> (gdb) bt >> 0 0x00007fa65aa78c6d in poll () from /lib/x86_64-linux-gnu/libc.so.6 >> 1 0x00007fa65a4144e0 in poll_dispatch (base=0x20e96b0, >> tv=0x7fff46f44a80) at poll.c:165 >> 2 0x00007fa65a409041 in opal_libevent2021_event_base_loop >> (base=0x20e96b0, flags=2) at event.c:1631 >> 3 0x00007fa65a3a3891 in opal_progress () at runtime/opal_progress.c:169 >> 4 0x00007fa65afbbc25 in opal_condition_wait (c=0x7fa65b2bfbc0 >> <ompi_request_cond>, m=0x7fa65b2bfcc0 <ompi_request_lock>) at >> ../opal/threads/condition.h:78 >> 5 0x00007fa65afbc1b5 in ompi_request_default_wait_all (count=2, >> requests=0x7fff46f44c70, statuses=0x0) at request/req_wait.c:287 >> 6 0x00007fa65afc7906 in comm_allgather_pml (src_buf=0x7fff46f44da0, >> dest_buf=0x233dac0, count=288, dtype=0x7fa65b29fee0 <ompi_mpi_char>, >> my_rank_in_group=1, >> n_peers=2, ranks_in_comm=0x210a760, comm=0x6020a0 >> <ompi_mpi_comm_world>) at patterns/comm/allgather.c:250 >> 7 0x00007fa64f14ba08 in bcol_basesmuma_smcm_allgather_connection >> (sm_bcol_module=0x7fa64e64d010, module=0x232c800, >> peer_list=0x7fa64f3513e8 <mca_bcol_basesmuma_component+456>, >> back_files=0x7fa64eae2690, comm=0x6020a0 <ompi_mpi_comm_world>, input=..., >> base_fname=0x7fa64f14ca8c "sm_ctl_mem_", map_all=false) at >> bcol_basesmuma_smcm.c:205 >> 8 0x00007fa64f146525 in base_bcol_basesmuma_setup_ctl >> (sm_bcol_module=0x7fa64e64d010, cs=0x7fa64f351220 >> <mca_bcol_basesmuma_component>) at bcol_basesmuma_setup.c:344 >> 9 0x00007fa64f146cbb in base_bcol_basesmuma_setup_library_buffers >> (sm_bcol_module=0x7fa64e64d010, cs=0x7fa64f351220 >> <mca_bcol_basesmuma_component>) >> at bcol_basesmuma_setup.c:550 >> 10 0x00007fa64f1418d0 in mca_bcol_basesmuma_comm_query (module=0x232c800, >> num_modules=0x232e570) at bcol_basesmuma_module.c:532 >> 11 0x00007fa64fd9e5f2 in mca_coll_ml_tree_hierarchy_discovery >> (ml_module=0x232fbe0, topo=0x232fd98, n_hierarchies=3, >> exclude_sbgp_name=0x0, include_sbgp_name=0x0) >> at coll_ml_module.c:1964 >> 12 0x00007fa64fd9f3a3 in mca_coll_ml_fulltree_hierarchy_discovery >> (ml_module=0x232fbe0, n_hierarchies=3) at coll_ml_module.c:2211 >> 13 0x00007fa64fd9cbe4 in ml_discover_hierarchy (ml_module=0x232fbe0) at >> coll_ml_module.c:1518 >> 14 0x00007fa64fda164f in mca_coll_ml_comm_query (comm=0x6020a0 >> <ompi_mpi_comm_world>, priority=0x7fff46f45358) at coll_ml_module.c:2970 >> 15 0x00007fa65b02f6aa in query_2_0_0 (component=0x7fa64fffe4e0 >> <mca_coll_ml_component>, comm=0x6020a0 <ompi_mpi_comm_world>, >> priority=0x7fff46f45358, >> module=0x7fff46f45390) at base/coll_base_comm_select.c:374 >> 16 0x00007fa65b02f66e in query (component=0x7fa64fffe4e0 >> <mca_coll_ml_component>, comm=0x6020a0 <ompi_mpi_comm_world>, >> priority=0x7fff46f45358, module=0x7fff46f45390) >> at base/coll_base_comm_select.c:357 >> 17 0x00007fa65b02f581 in check_one_component (comm=0x6020a0 >> <ompi_mpi_comm_world>, component=0x7fa64fffe4e0 <mca_coll_ml_component>, >> module=0x7fff46f45390) >> at base/coll_base_comm_select.c:319 >> 18 0x00007fa65b02f3c7 in check_components (components=0x7fa65b2a9530 >> <ompi_coll_base_framework+80>, comm=0x6020a0 <ompi_mpi_comm_world>) >> at base/coll_base_comm_select.c:283 >> 19 0x00007fa65b027d45 in mca_coll_base_comm_select (comm=0x6020a0 >> <ompi_mpi_comm_world>) at base/coll_base_comm_select.c:119 >> 20 0x00007fa65afbdb2c in ompi_mpi_init (argc=1, argv=0x7fff46f45a78, >> requested=0, provided=0x7fff46f4590c) at runtime/ompi_mpi_init.c:858 >> 21 0x00007fa65aff20ef in PMPI_Init (argc=0x7fff46f4594c, >> argv=0x7fff46f45940) at pinit.c:84 >> 22 0x0000000000400b66 in main (argc=1, argv=0x7fff46f45a78) at >> hellompi.c:11 >> >> >> >> -- >> С Уважением, Поляков Артем Юрьевич >> Best regards, Artem Y. Polyakov >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >> Link to this post: >> http://www.open-mpi.org/community/lists/devel/2014/10/16055.php >> > > > > -- > С Уважением, Поляков Артем Юрьевич > Best regards, Artem Y. Polyakov > > _______________________________________________ > devel mailing list > de...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > Link to this post: > http://www.open-mpi.org/community/lists/devel/2014/10/16067.php >