See src/snes/tutorials/ex70.c for the code that I think was used for that paper.

Alexander Lindsay <alexlindsay...@gmail.com> writes:

> Sorry for the spam. Looks like these authors have published multiple papers 
> on the subject 
>
> cover.jpg 
> Combining the Augmented Lagrangian Preconditioner with the Simple Schur 
> Complement Approximation | SIAM Journal on  
> Scientific Computingdoi.org  
> cover.jpg
>
>  On Jul 28, 2023, at 12:59 PM, Alexander Lindsay <alexlindsay...@gmail.com> 
> wrote:
>
>  Do you know of anyone who has applied the augmented Lagrange methodology to 
> a finite volume discretization?
>
>  On Jul 6, 2023, at 6:25 PM, Matthew Knepley <knep...@gmail.com> wrote:
>
>  On Thu, Jul 6, 2023 at 8:30 PM Alexander Lindsay <alexlindsay...@gmail.com> 
> wrote:
>
>  This is an interesting article that compares a multi-level ILU algorithm to 
> approximate commutator and augmented
>  Lagrange methods: https://doi.org/10.1002/fld.5039
>
>  That is for incompressible NS. The results are not better than 
> https://arxiv.org/abs/1810.03315, and that PC is considerably
>  simpler and already implemented in PETSc. There is an update in to this
>
>   
>  
> https://epubs.siam.org/doi/abs/10.1137/21M1430698?casa_token=Fp_XhuZStZ0AAAAA:YDhnkW9XvAom_b8KocWz-hBEI7FAt46aw3ICa0FvCrOVCtYr9bwvtqJ4aBOTkDSvANKh6YTQEw
>  
>
>  which removes the need for complicated elements.
>
>  You might need stuff like ILU for compressible flow, but I think 
> incompressible is solved.
>
>    Thanks,
>
>       Matt
>   
>  On Wed, Jun 28, 2023 at 11:37 AM Alexander Lindsay 
> <alexlindsay...@gmail.com> wrote:
>
>  I do believe that based off the results in https://doi.org/10.1137/040608817 
> we should be able to make LSC, with
>  proper scaling, compare very favorably with PCD
>
>  On Tue, Jun 27, 2023 at 10:41 AM Alexander Lindsay 
> <alexlindsay...@gmail.com> wrote:
>
>  I've opened https://gitlab.com/petsc/petsc/-/merge_requests/6642 which adds 
> a couple more scaling
>  applications of the inverse of the diagonal of A
>
>  On Mon, Jun 26, 2023 at 6:06 PM Alexander Lindsay <alexlindsay...@gmail.com> 
> wrote:
>
>  I guess that similar to the discussions about selfp, the approximation of 
> the velocity mass matrix by the
>  diagonal of the velocity sub-matrix will improve when running a transient as 
> opposed to a steady
>  calculation, especially if the time derivative is lumped.... Just thinking 
> while typing
>
>  On Mon, Jun 26, 2023 at 6:03 PM Alexander Lindsay <alexlindsay...@gmail.com> 
> wrote:
>
>  Returning to Sebastian's question about the correctness of the current LSC 
> implementation: in the
>  taxonomy paper that Jed linked to (which talks about SIMPLE, PCD, and LSC), 
> equation 21 shows four
>  applications of the inverse of the velocity mass matrix. In the PETSc 
> implementation there are at
>  most two applications of the reciprocal of the diagonal of A (an 
> approximation to the velocity mass
>  matrix without more plumbing, as already pointed out). It seems like for 
> code implementations in
>  which there are possible scaling differences between the velocity and 
> pressure equations, that this
>  difference in the number of inverse applications could be significant? I 
> know Jed said that these
>  scalings wouldn't really matter if you have a uniform grid, but I'm not 100% 
> convinced yet.
>
>  I might try fiddling around with adding two more reciprocal applications.
>
>  On Fri, Jun 23, 2023 at 1:09 PM Pierre Jolivet <pierre.joli...@lip6.fr> 
> wrote:
>
>  On 23 Jun 2023, at 10:06 PM, Pierre Jolivet <pierre.joli...@lip6.fr> wrote:
>
>  On 23 Jun 2023, at 9:39 PM, Alexander Lindsay <alexlindsay...@gmail.com> 
> wrote:
>
>  Ah, I see that if I use Pierre's new 'full' option for 
> -mat_schur_complement_ainv_type
>
>  That was not initially done by me
>
>  Oops, sorry for the noise, looks like it was done by me indeed in
>  9399e4fd88c6621aad8fe9558ce84df37bd6fada…
>
>  Thanks,
>  Pierre
>
>  (though I recently tweaked MatSchurComplementComputeExplicitOperator() a bit 
> to use
>  KSPMatSolve(), so that if you have a small Schur complement — which is not 
> really the case
>  for NS — this could be a viable option, it was previously painfully slow).
>
>  Thanks,
>  Pierre
>
>  that I get a single iteration for the Schur complement solve with LU. That's 
> a nice testing
>  option
>
>  On Fri, Jun 23, 2023 at 12:02 PM Alexander Lindsay <alexlindsay...@gmail.com>
>  wrote:
>
>  I guess it is because the inverse of the diagonal form of A00 becomes a poor
>  representation of the inverse of A00? I guess naively I would have thought 
> that the
>  blockdiag form of A00 is A00
>
>  On Fri, Jun 23, 2023 at 10:18 AM Alexander Lindsay <alexlindsay...@gmail.com>
>  wrote:
>
>  Hi Jed, I will come back with answers to all of your questions at some 
> point. I
>  mostly just deal with MOOSE users who come to me and tell me their solve is
>  converging slowly, asking me how to fix it. So I generally assume they have
>  built an appropriate mesh and problem size for the problem they want to solve
>  and added appropriate turbulence modeling (although my general assumption
>  is often violated).
>
>  > And to confirm, are you doing a nonlinearly implicit velocity-pressure 
> solve?
>
>  Yes, this is our default.
>
>  A general question: it seems that it is well known that the quality of selfp
>  degrades with increasing advection. Why is that?
>
>  On Wed, Jun 7, 2023 at 8:01 PM Jed Brown <j...@jedbrown.org> wrote:
>
>  Alexander Lindsay <alexlindsay...@gmail.com> writes:
>
>  > This has been a great discussion to follow. Regarding
>  >
>  >> when time stepping, you have enough mass matrix that cheaper
>  preconditioners are good enough
>  >
>  > I'm curious what some algebraic recommendations might be for high Re
>  in
>  > transients. 
>
>  What mesh aspect ratio and streamline CFL number? Assuming your model
>  is turbulent, can you say anything about momentum thickness Reynolds
>  number Re_θ? What is your wall normal spacing in plus units? (Wall
>  resolved or wall modeled?)
>
>  And to confirm, are you doing a nonlinearly implicit velocity-pressure
>  solve?
>
>  > I've found one-level DD to be ineffective when applied monolithically or
>  to the momentum block of a split, as it scales with the mesh size. 
>
>  I wouldn't put too much weight on "scaling with mesh size" per se. You
>  want an efficient solver for the coarsest mesh that delivers sufficient
>  accuracy in your flow regime. Constants matter.
>
>  Refining the mesh while holding time steps constant changes the advective
>  CFL number as well as cell Peclet/cell Reynolds numbers. A meaningful
>  scaling study is to increase Reynolds number (e.g., by growing the domain)
>  while keeping mesh size matched in terms of plus units in the viscous
>  sublayer and Kolmogorov length in the outer boundary layer. That turns
>  out to not be a very automatic study to do, but it's what matters and you
>  can spend a lot of time chasing ghosts with naive scaling studies.
>
>  -- 
>  What most experimenters take for granted before they begin their experiments 
> is infinitely more interesting than any
>  results to which their experiments lead.
>  -- Norbert Wiener
>
>  https://www.cse.buffalo.edu/~knepley/

Reply via email to