Dear Roy,

I hope this does not come too late. But I finally got the debugger to work with 
some help of support on the Supercomputer that we are using. This is the 
complete stack:

To recap the problem:
We were reading in a mesh that was created outside of libmesh (rectilinear) and 
wanted to use that in the context of the MOOSE framework. Before calling the 
function prepare_for_use we wanted to make sure that the ghost elements for the 
neighbouring processes were found. This seemed to work fine when compiled and 
run in opt mode but caused the code to get stuck when done so in dbg mode (no 
crash, just a running but hung-up executable). So this is the complete stack of 
what is happening. It seems like MPI_Probe is blocking the code for some reason 
that it is not getting its message on one of the processes.
Have you ever come across something like that?

–Barna


> On 13 Jan 2017, at 17:04, Roy Stogner <royst...@ices.utexas.edu> wrote:
> 
> 
> On Fri, 13 Jan 2017, Barna Becsek wrote:
> 
>> Hmm, you are right but I cannot find a process that would correspond to the 
>> children of any of these.
> 
> Is this a cluster, by any chance?  Where slurm is running on the login
> node but the actual applications are running on separate compute
> nodes?
> 
> If I'm right, then I'd recommend testing on a simple workstation
> first, just to make debugging easier.  Even if you only have 2 or 4
> cores you can still "mpirun -np 8" if that's what's necessary to
> trigger the bug.
> ---
> Roy

------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, SlashDot.org! http://sdm.link/slashdot
_______________________________________________
Libmesh-users mailing list
Libmesh-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/libmesh-users

Reply via email to