Hartmut, The code related to this error was derived from the sequential, recursive, dataflow code from the last week. It sounds like a reasonable generalization to some of these issues relates to variables going out of scope (that's a great answer to the question!). I'll be modifying the test and will send over the modified test to if there's still an issue.
Chris On Sun, May 21, 2017 at 6:49 AM, Hartmut Kaiser <[email protected]> wrote: > Chris, > > Sorry for the late reply. > > > I've an application that is running a dataflow of asynchronous functions > > over a linear sequence of doubles stored in a partitioned_vector. > > > > The application processes a fair number of elements before causing a > > segmentation fault. Upon inspection of the core file, the segmentation > > fault is happening when > > hpx::agas::server::primary_namespace::resolve_gid(hpx::naming::gid_type) > > () from libhpx.so.1 is called by the runtime. > > I've compiled the HPX runtime using the MPI ParcelPort. The application > is > > running on a single remote node (using slurm for scheduling). > > Any suggestions or recommendations of how to further debug the > application > > or any runtime flags to help further diagnose implementation errors would > > be appreciated. > > That is very difficult to diagnose from the distance. It looks like that > something went out of scope too early and an attempt to access it went > hiwire. Can you give us the code to have a closer look? > > Regards Hartmut > --------------- > http://boost-spirit.com > http://stellar.cct.lsu.edu > > > _______________________________________________ > hpx-users mailing list > [email protected] > https://mail.cct.lsu.edu/mailman/listinfo/hpx-users >
_______________________________________________ hpx-users mailing list [email protected] https://mail.cct.lsu.edu/mailman/listinfo/hpx-users
