Barry,
Putting P_{A} either to the left or to the right of A means the same thing: we
avoid touching the subspace spanned by W. This is why the orthogonality
condition b-Ax_0 \perp W is needed. What the condition says is that the initial
residual should have no components on W, presumably the troublesome part of A.
Then in a Krylov method, all later residuals have no components on W. In other
words, the part of the solution on the subspace W is already fully computed
even in the 0-th step. Getting such an x_0 is not difficult; the difficult part
is to define/compute W.
When one adds another preconditioner M to the system, presumably W should be
the troublesome part of AM instead of A (I am always confused about the
notation M and M^{-1} but it does not affect my reasoning here). In the
aggressive way W can consist of eigenvectors of the pencil (A,M) corresponding
to the smallest eigenvalues in magnitude. On the other hand, if one already
found a good W for A and he/she is lazy and does not want to re-figuring out W,
then just use the old W. I guess the beauty of the deflation theory is that W
can be arbitrary but does not depend on A or M.
I sense that you want a preconditioner appearing in (4) and the formula for
x_0? I will add something there later.
Jie
----- Original Message -----
From: "Barry Smith" <[email protected]>
To: "For users of the development version of PETSc" <petsc-dev at mcs.anl.gov>
Sent: Saturday, March 2, 2013 10:33:05 PM
Subject: Re: [petsc-dev] Deflated Krylov solvers for PETSc
Jie,
Thanks. I think you know this but it is worth writing in the notes:
For symmetric positive definite problems you can write the preconditioned
operator (ignore the right hand side) as
right preconditioning
A[I ? W(W^{H}AW)^{?1}W^{H}A] x = A P_{A} x
and see exactly what the deflation is doing. It is projecting any vector x onto
the subspace spanned by W (in the A inner product) and removing it from the
vector before applying A. Hence A (and the Krylov method) never "sees" that
part of the space ( in other words that troublesome part of A). Similarly if
you write it as left preconditioning you have
P_{A}^{H} A x
Hence the Krylov space does not contain the null space of P_{A}^{H}. Again,
presumably removing some troublesome part of A.
In my book, top of page 17 is the statement:
"The procedure of restriction, (approximately) solving a small linear system,
and interpolation lies at the heart of all domain decomposition and
multigrid/multilevel algorithms. As indicated above and ? this procedure
calculates the projection of the unknown error onto a subspace and uses the
result as a correction term. An effective domain decomposition algorithm will
generate the appropriate subspaces that provide a good representation of the
solution".
This is why I am interested in understanding EXACTLY the relationship of
deflation and preconditioning and the subspaces used. Another way to phrase it
is does one (1) augment the preconditioner with another subspace to generate
the Krylov space (composing preconditioners in the DD/Multigrid sense) or (2)
does one remove those certain "modes" from ever getting into the Krylov
subspace? Or are the two different things actually more or less one and the
same?
I don't know the answer but if we knew the answer it would help us understand
how/where deflation might belong in PETSc (or maybe multiple places).
Barry
On Mar 2, 2013, at 9:43 PM, Jie Chen <jiechen at mcs.anl.gov> wrote:
> Mark, Barry, Jed,
>
> You guys raised quite a few questions, meanwhile I found a mistake in my
> note. I think I should rewrite the note to incorporate a more comprehensive
> view. Hopefully the new note will answer you questions (at least to a certain
> degree). I have something very important to finish before Wednesday, so my
> participation in the discussion may delay. See the new note in the old place:
>
> http://www.mcs.anl.gov/~jiechen/tmp/deflate.pdf
>
> Jie
>
>
>
>
> ----- Original Message -----
> From: "Mark F. Adams" <mark.adams at columbia.edu>
> To: "For users of the development version of PETSc" <petsc-dev at mcs.anl.gov>
> Sent: Saturday, March 2, 2013 12:54:34 PM
> Subject: Re: [petsc-dev] Deflated Krylov solvers for PETSc
>
>
> On Mar 2, 2013, at 12:45 PM, Barry Smith <bsmith at mcs.anl.gov> wrote:
>
>>
>> On Mar 2, 2013, at 11:16 AM, "Mark F. Adams" <mark.adams at columbia.edu>
>> wrote:
>>
>>>>
>>>> An alternative description for the method in the paper is "two-level
>>>> unsmoothed aggregation applied to the ASM-preconditioned operator".
>>>>
>>>
>>> Isn't this paper just doing two level unsmoothed aggregation with a V(0,k)
>>> cycle, where k (and the smoother) is K in the paper?
>>>
>>> It looks to me like you are moving the multigrid coarse grid correction
>>> (projection) from the PC into the operator, in which case this method is
>>> identical to two level MG with the V(0,k) cycle.
>>
>> Well, its not "identical" is it?
>
> My thinking is that with Q = P (RAP)^-1 R, and smoother S, then a two level
> V(0,1) PC (M) looks like: S (I - Q). (I think I'm missing an A in here,
> before the Q maybe) The preconditioned system being solved looks like:
>
> MAx = Mb
>
> with MG PC we have:
>
> (S (I-Q)) A x = ?
>
> Move the brackets around
>
> S ((I-Q)A) x = ..
>
> S is K in the paper and (I-Q)A is the new operator. The preconditioned
> system does not have a null space and thats all that matters.
>
> Anyway, I'm sure its not identical, I'm just not seeing it in this paper and
> don't have the patients work through it ? and maybe I'm having a bad
> arithmetic day but on page 935, the definition of P (first equation on page)
> and Pw (last equation on page) don't look consistent.
>
>
>> In the deflation approach it introduces a null space into the operator, in
>> the multigrid approach it does not. So identical only in analogy? So what I
>> am interested in EXACTLY how are they related, relationship between
>> eigenvalues or convergence rate.
>>
>> Barry
>>
>>>
>>> I'm sure I'm missing something. Jie's writeup has an orthogonality
>>> condition on the restriction operator, which I don't see in the Vuik paper.
>>
>>
>