Hi Tri Dat,

was is your current status on this?

Can you provide a minimal example in terms of number of elements and process count where the problem occurs?

Kind regards
Bernd

On 11/28/2015 01:30 AM, Tri Dat NGO wrote:
Hi Birger and Alex,

Thank you very much for your shares and your advices.
I confirm that installing SuperLU solved my problem in the homogeneous case. But when trying to run my tests in heteogeneous media, I got the same convergence problem as Birger:

"Newton: Caught exception: "NumericalProblem [newtonSolveLinear:../dumux/nonlinear/newtoncontroller.hh:389]: Linear solver did not converge"

I don't get any convergence at all, even with very small time steps. Any hints are welcome.

@ Alex: Did you try run your test with fully implicit cc model? From my experience, in general, cc model works correctly in parallel (parallel run are faster than sequentiel run) but not box model.

Best regards,
Tri Dat

2015-11-26 17:26 GMT+01:00 Alexander Kissinger <[email protected] <mailto:[email protected]>>:

    Hi Birger and Tri Dat,

    are you not getting any convergence at all or do you get
    convergence at a smaller time step size?
    Do you get the error in the first Newton step of a time step or in
    between somewhere?

    I have similar problems for parallel computations with bad aspect
    ratios (300:10). Additionally I have horizontal layers with large
    differences in permeability. I am using AMG (also with SuperLU as
    coarse solver) and the fully implicit box model (1p2c).
    Running the same grid sequentially leads to much larger time step
    sizes until my time step is constrained by the linear solver.
    I also ran the my grid for a homogeneous case with simplified
    physics which the linear solver was able to handle.
    I assume that the problem in my case is a combination of
    heterogeneities and the bad aspect ratio.

    Unfortunately I was not able to resolve the problem. Some actions
    that slightly improved the behavior were:

    (1) In amgproperties.hh setting the Preconditioner from
    Dune::SeqSSOR to Dune::SeqILUn (This also improved convergence
    behavior of the linear solver for the sequential case)

    (2) If you use Dune ALUGrud. Try setting METIS_PartGraphRecursive
    in your alugrid.cfg file. Another point I observed is that the
    partitioning method played a huge role. Metis worked the best for me.

    (3) In amgbackend.hh increasing the coarsen target to a value
    larger than 2000 (for example 10000, 50000 or 100000). In the
    following line:
    Dune::Amg::Parameters params(15,2000,1.2,1.6,Dune::Amg::atOnceAccu);

    You can see the effect of the latter when you set the linear
    solver verbosity to 2 in your input file:

     [LinearSolver]
    Verbosity = 2

    AMG will print the different agglomeration levels for each Newton
    step. If you increase the coarsen target there should be fewer
    levels. On the highest level the coarse solver (SuperLU) is used.
    The convergence in my case improved with increasing number of
    unknowns in the highest level. The downside to this is that it
    takes much more time to solve the coarse linear system with
    SuperLU which for me meant that I was much faster in the
    sequential case than in the parallel one in the end. But maybe you
    have more luck.

    Best regards
    Alex







    On 11/26/2015 04:05 PM, Birger Hagemann wrote:

    Hi Tri Dat,

    I had recently the same problem with parallel computation on a
    realistic reservoir grid where also the horizontal dimensions are
    much larger than the vertical. The error message was the same.
    Bernd gave me the hint to install SuperLU. However, after
    installing SuperLU the error was only changed to:

    “Newton: Caught exception: "NumericalProblem

    >>[newtonSolveLinear:…/dumux/dumux/nonlinear/newtoncontroller.hh:380]:

    >>Linear solver did not converge"

    Maybe installing SuperLU will solve your problem. My problem is
    also still unsolved, thus, any hints are welcome. I am also using
    a fully implicit cell-centered model. The same simulation is
    working when started on only one processor and the same model is
    working in parallel on more simple grids.

    Kind regards

    Birger

    *Von:*Dumux [mailto:[email protected]] *Im
    Auftrag von *Tri Dat NGO
    *Gesendet:* Dienstag, 24. November 2015 20:07
    *An:* DuMuX User Forum
    *Betreff:* Re: [DuMuX] Convergence problem for 3D simulations (2p
    cell-centered model, ALUGrid)

    Sorry, I forgot to add my files.

    Kind regards,
    Tri Dat

    2015-11-24 20:03 GMT+01:00 Tri Dat NGO <[email protected]
    <mailto:[email protected]>>:

    Hi Dumuxers,

    I would like to run parallel simulations of CO2 injection into a
    3D reservoir of which the horizontal extents are much higher than
    the vertical one: Lx/Lz = Ly/Lz >> 1 (e.g. a reservoir of 1000m x
    1000m x 50 m). A 2p cell-centered model is used.

    The domain is described on a cartesian grid of 100x100x100 (1E+6)
    cells. The CO2 is injected into the domain at X=500m  and Y =
    500m by using the Peaceman's well model (bhp). The pressure at
    the boundaries perpendicular to the Y-axis is set to the initial
    hydrostatic pressure. All others boundaries are no-flux Neumann
    boundaries. At the first step, we consider, for simplicity, an
    homogeneous reservoir. You can find in attachment the *.hh files
    of my test.

    The simulation converges only if the reservoir is anisotropic
    (Kxx/Kzz = Kyy/Kzz >>1), meaning that the medium is lowly
    permeable in Z-direction. Nevertheless, when the domain is
    isotropic (Kxx = Kzz), a problem occurs: Newton solver did not
    converge. The error message is "MathError
    [mgc:/home/share/soft/dumux-2.6/include/dune/istl/paamg/amg.hh:825]:
    Coarse solver did not converge".


    Moreover, the same test works correctly on isotropic but thicker
    domains (e.g. 1000m x 1000m x 250 m, the grid is always of
    100x100x100).

    I don't understand why a same test works on a grid but not on
    another grid. Is this related to the numerical scheme
    (fully-implicit, cell-centered)?


    Any help will be greatly appreciated!

    Kind regards,

    Tri Dat



    _______________________________________________
    Dumux mailing list
    [email protected]
    <mailto:[email protected]>
    https://listserv.uni-stuttgart.de/mailman/listinfo/dumux


-- Alexander Kissinger
    Institut für Wasser- und Umweltsystemmodellierung
    Lehrstuhl für Hydromechanik und Hydrosystemmodellierung
    Pfaffenwaldring 61
    D-70569 Stuttgart

    Telefon:+49 (0) 711 685-64729 <tel:%2B49%20%280%29%20711%20685-64729>
    E-Mail:[email protected]
    <mailto:[email protected]>


    _______________________________________________
    Dumux mailing list
    [email protected]
    <mailto:[email protected]>
    https://listserv.uni-stuttgart.de/mailman/listinfo/dumux




_______________________________________________
Dumux mailing list
[email protected]
https://listserv.uni-stuttgart.de/mailman/listinfo/dumux


--
_______________________________________________________________

Bernd Flemisch                         phone: +49 711 685 69162
IWS, Universität Stuttgart             fax:   +49 711 685 60430
Pfaffenwaldring 61            email: [email protected]
D-70569 Stuttgart            url: www.hydrosys.uni-stuttgart.de
_______________________________________________________________

_______________________________________________
Dumux mailing list
[email protected]
https://listserv.uni-stuttgart.de/mailman/listinfo/dumux

Reply via email to