Can't really reproduce using PETSc 3.6.1:

Process 0: Number of global vertices: 10201
Process 0: Number of global cells: 20000
Expr.geometric_dimension() is deprecated, please use 
find_geometric_dimension(expr) instead.
Expr.geometric_dimension() is deprecated, please use 
find_geometric_dimension(expr) instead.
Expr.geometric_dimension() is deprecated, please use 
find_geometric_dimension(expr) instead.
10 None n =  100 100
10 None n =  100 100
10 None n =  100 100
Process 0: *** Warning: The underlying linear solver cannot be modified for 
this specified TAO solver. The options are all ignored.
Process 1: *** Warning: The underlying linear solver cannot be modified for 
this specified TAO solver. The options are all ignored.
Process 2: *** Warning: The underlying linear solver cannot be modified for 
this specified TAO solver. The options are all ignored.
1 2499.99999808
1 2499.99999808
1 2499.99999808
2 9.19912577662
2 9.19912577662
2 9.19912577662
3 38082.7834178
3 38082.7834178
3 38082.7834178
^Cmpirun: killing job...


Jan


On Mon, 12 Oct 2015 18:12:19 +0300
Giorgos Grekas <[email protected]> wrote:

> A simpler example where the same error happens with
> mpirun -np 4
> 
>  you can run the attached code, it is very shorter. I am sorry that i
> sent so many lines of code before.
> 
> 
> 
> On Mon, Oct 12, 2015 at 5:15 PM, Giorgos Grekas <[email protected]>
> wrote:
> 
> > I provide backtrace to the file bt.txt and my code. For my code you
> > need to run the file runMe.py.
> >
> >
> > On Mon, Oct 12, 2015 at 4:40 PM, Jan Blechta
> > <[email protected]> wrote:
> >
> >> PETSc error code 1 does not seem to indicate an expected problem,
> >> http://www.mcs.anl.gov/petsc/petsc-dev/include/petscerror.h.html.
> >> It seems as an error not handled by PETSc.
> >>
> >> You could provide us with your code or try investigating the
> >> problem with debugger
> >>
> >>   $ mpirun -n 3 xterm -e gdb -ex 'set breakpoint pending on' -ex
> >> 'break PetscError' -ex 'break dolfin::dolfin_error' -ex r -args
> >> python your_script.py
> >>   ...
> >>   Break point hit...
> >>   (gdb) bt
> >>
> >> and post a backtrace here.
> >>
> >> Jan
> >>
> >>
> >> On Mon, 12 Oct 2015 15:16:48 +0300
> >> Giorgos Grekas <[email protected]> wrote:
> >>
> >> > Hello,
> >> > i am using ncg from tao solver and i wanted to test my code
> >> > validity in a pc  with 4 processors
> >> > before its execution in a cluster. When i run my code with 2
> >> > processes (mpirun -np 2) everything
> >> > looks to work fine but when i use 3 or more processes i have the
> >> > following error:
> >> >
> >> >
> >> >  Error:   Unable to successfully call PETSc function
> >> > 'VecAssemblyBegin'. *** Reason:  PETSc error code is: 1.
> >> > *** Where:   This error was encountered inside
> >> >
> >> /home/ggrekas/.hashdist/tmp/dolfin-wphma2jn5fuw/dolfin/la/PETScVector.cpp.
> >> > *** Process: 3
> >> > ***
> >> > *** DOLFIN version: 1.7.0dev
> >> > *** Git changeset:  3fbd47ec249a3e4bd9d055f8a01b28287c5bcf6a
> >> > ***
> >> >
> >> -------------------------------------------------------------------------
> >> >
> >> >
> >> >
> >> ===================================================================================
> >> > =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
> >> > =   EXIT CODE: 134
> >> > =   CLEANING UP REMAINING PROCESSES
> >> > =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
> >> >
> >> ===================================================================================
> >> > YOUR APPLICATION TERMINATED WITH THE EXIT STRING: Aborted
> >> > (signal 6) This typically refers to a problem with your
> >> > application. Please see the FAQ page for debugging suggestions
> >> >
> >> >
> >> >
> >> >
> >> >
> >> > So, is it an issue that i must report to the tao team?
> >> >
> >> > Thank you in advance.
> >>
> >>
> >

_______________________________________________
fenics-support mailing list
[email protected]
http://fenicsproject.org/mailman/listinfo/fenics-support

Reply via email to