Looks like this should be fixed in my PR #101 - could you please review it?
Thanks
Ralph
> On Nov 26, 2014, at 8:14 PM, Ralph Castain wrote:
>
> Aha - I see what happened. I have that param set to false in my default mca
> param file. If I set it to true on the cmd line,
Aha - I see what happened. I have that param set to false in my default mca
param file. If I set it to true on the cmd line, then I run without segfaulting.
Thanks!
Ralph
> On Nov 26, 2014, at 5:55 PM, Gilles Gouaillardet
> wrote:
>
> Ralph,
>
> let me
Ralph,
let me correct and enhance my previous statement :
- i cannot reproduce your crash in my environment (RHEL6 like vs your
RHEL7 like)
(i configured with --enable-debug --enable-picky)
- i can reproduce the crash with
mpirun --mca mpi_param_check false
- if you configured with
> On Nov 26, 2014, at 5:06 PM, Gilles Gouaillardet
> wrote:
>
> I will double check this(afk right now)
> Are you running on a rhel6 like distro with gcc ?
Yeah, I’m running CentOS7 and gcc 4.8.2
>
> Iirc, crash vs mpi error is ruled by --with-param-check or
I will double check this(afk right now)
Are you running on a rhel6 like distro with gcc ?
Iirc, crash vs mpi error is ruled by --with-param-check or something like
this...
Cheers,
Gilles
Ralph Castain さんのメール:
>I tried it with both the fortran and c versions - got the same
I tried it with both the fortran and c versions - got the same result.
This was indeed with a debug build. I wouldn’t expect a segfault even with an
optimized build, though - I would expect an MPI error, yes?
> On Nov 26, 2014, at 4:26 PM, Gilles Gouaillardet
>
I will have a look
Btw, i was running the fortran version, not the c one.
Did you confgure with --enable--debug ?
The program sends to a rank *not* in the communicator, so this behavior could
make some sense on an optimized build.
Cheers,
Gilles
Ralph Castain さんのメール:
>Ick -
Ick - I’m getting a segfault when trying to run that test:
MPITEST info (0): Starting MPI_Errhandler_fatal test
MPITEST info (0): This test will abort after printing the results message
MPITEST info (0): If it does not, then a f.a.i.l.u.r.e will be noted
[bend001:07714] *** Process received
Hmmm….yeah, I know we saw this and resolved it in the trunk, but it looks like
the fix indeed failed to come over to 1.8. I’ll take a gander (pretty sure I
remember how I fixed it) - thanks!
> On Nov 26, 2014, at 12:03 AM, Gilles Gouaillardet
> wrote:
>
>
Ralph,
i noted several hangs in mtt with the v1.8 branch.
a simple way to reproduce it is to use the MPI_Errhandler_fatal_f test
from the intel_tests suite,
invoke mpirun on one node and run the taks on an other node :
node0$ mpirun -np 3 -host node1 --mca btl tcp,self ./MPI_Errhandler_fatal_f
10 matches
Mail list logo