Professional Documents
Culture Documents
(AM pk , pℓ ) = 0 for k, ℓ = 0, . . . , M − 1, k = ℓ.
292 13 Iterative Solution Methods
For an arbitrary given initial guess u0 ∈ RM we can write the unique solution
u ∈ RM of the linear system AM u = f as a linear combination of conjugate
vectors as
M−1
u = u0 − αℓ pℓ .
ℓ=0
Hence we have
M
−1
AM u = AM u0 − αℓ AM pℓ = f ,
ℓ=0
and from the AM –orthogonality of the basis vectors pℓ we can compute the
yet unknown coefficients from
(AM u0 − f , pℓ )
αℓ = for ℓ = 0, 1, . . . , M − 1.
(AM pℓ , pℓ )
uk+1 := uk − αk pk for k = 0, 1, . . . , M − 1,
If we denote by
rk := AM uk − f
the residual of the approximate solution uk we finally have
(rk , pk )
αk = . (13.1)
(AM pk , pk )
The above approach is based on the use of AM –orthogonal vectors {pℓ }M ℓ=0 .
1
Initialize for k = 0:
p0 := w0
Compute for k = 0, 1, . . . , M − 2:
k
(AM wk+1 , pℓ )
pk+1 := wk+1 − βkℓ pℓ , βkℓ =
(Apℓ , pℓ )
ℓ=0
(rk+1 , pℓ ) = 0 for ℓ = 0, 1, . . . , k.
(rk+1 , wℓ ) = 0 for ℓ = 0, 1, . . . , k.
294 13 Iterative Solution Methods
Hence we obtain
ℓ−1
(rk+1 , wℓ ) = (rk+1 , pℓ ) + βℓ−1,j (rk+1 , pj ),
j=0
w0 , w1 , . . . , wk , rk+1
are orthogonal to each other, and therefore linear independent. Since we need
only to know the search directions p0 , . . . , pk and therefore the initial vec-
tors w0 , . . . , wk to construct the approximate solution uk+1 and therefore the
residual rk+1 , we can define the new initial vector as
(rk+1 , rℓ ) = 0 for ℓ = 0, . . . , k; k = 0, . . . , M − 2.
(rk , rk )
αk = for k = 0, . . . , M − 1.
(AM pk , pk )
αℓ > 0 for ℓ = 0, . . . , k.
implying rℓ+1 = 0 and therefore uℓ+1 = u would be the exact solution of the
linear system AM u = f .
13.1 The Method of Conjugate Gradients 295
Now, by using wk+1 = rk+1 and by using the symmetry of the system matrix
AM = A⊤M we can compute the nominator of the coefficient βkℓ as
1 k+1 ℓ
(AM wk+1 , pℓ ) = (rk+1 , AM pℓ ) = (r , r − rℓ+1 )
αℓ
⎧
⎪
⎨ 0 for ℓ < k,
= (r k+1
, rk+1 )
⎪
⎩− for ℓ = k.
αk
The recursion of the Gram–Schmidt orthogonalization algorithm now reduces
to
1 (rk+1 , rk+1 )
pk+1 = rk+1 − βkk pk where βkk = − .
αk (AM pk , pk )
On the other hand we have
and therefore
(rk+1 , rk+1 )
pk+1 = rk+1 + βk pk where βk = .
(rk , rk )
.
1 κ2 (AFEM
h )+1 1
≈ ln . = ln qh .
2 κ2 (Ah ) − 1
FEM 2
13.1 The Method of Conjugate Gradients 297
Therefore, in the case of an uniform refinement step, i.e. halving the mesh
size h, the number of required iterations is doubled to reach the same relative
accuracy ε. As an example we choose ε = 10−10 . In Table 13.1 we give the
number of iterations of the conjugate gradient method to obtain the results
which were already presented in Table 11.2.
FEM BEM
L N κ2 (Ah ) Iter N κ2 (Vh ) Iter
2 64 12.66 13 16 24.14 8
3 256 51.55 38 32 47.86 18
4 1024 207.17 79 64 95.64 28
5 4096 829.69 157 128 191.01 39
6 16384 3319.76 309 256 381.32 52
7 65536 13280.04 607 512 760.73 69
8 262144 52592.92 1191 1024 1516.02 91
Theory: O(h−2 ) O(h−1 ) O(h−1 ) O(h−1/2 )
κ2 (ABEM
h ) = O(h−1 ) i.e. κ2 (ABEM BEM
h/2 ) ≈ 2 κ2 (Ah ).
satisfying
1/2 1/2 −1/2 1/2
CA = CA CA , CA := (CA )−1 .
Instead of the linear system AM u = f we now consider the equivalent system
298 13 Iterative Solution Methods
:= C −1/2 AM C −1/2
A A A
is again symmetric and positive definite. Hence we can apply the method
of conjugate gradients as described in Algorithm 13.2 to compute the trans-
1/2
= CA u. Inserting all the transformations we finally
formed solution vector u
obtain the preconditioned method of conjugate gradients, see Algorithm 13.3.
1/2
Note that for z = CA z we have
1/2 1/2
z 2A = (AC
2
A z, CA z) = (AM z, z) = zAM .
−1/2 k
Hence, for the approximate solution uk = CA u
we find the error estimate
2
q
uk − uAM ≤ u0 − uAM .
1 + q2k
13.2 A General Preconditioning Strategy 299
we
To bound the extremal eigenvalues of the transformed system matrix A
get from the Rayleigh quotient
z , z) z , z)
= min (A
λmin (A) ≤ max
(A
= λmax (A).
z∈RM (
z , z) z∈RM (
z , z)
cA A
1 (CA z, z) ≤ (AM z, z) ≤ c2 (CA z, z) for z ∈ RM (13.2)
−1/2 −1/2 −1 cA
2
κ2 (CA ACA ) = κ2 (CA AM ) ≤ .
cA
1
−1
If the spectral condition number κ2 (CA AM ) of the preconditioned system
matrix can be bounded independent of the dimension M , i.e. independent of
the mesh size h, then there is a fixed number kε of required iterations to reach
a certain given relative accuracy ε.
Av, v ≥ cA 2
1 vX , AvX ′ ≤ cA
2 vX for v ∈ X. (13.3)
Bf, f ≥ cB 2
1 f X ′ , Bf X ≤ cB
2 f X ′ .
300 13 Iterative Solution Methods
and define
BM [ℓ, k] = Bψk , ψℓ , MM [ℓ, k] = ϕk , ψℓ for k, ℓ = 1, . . . , M.
Note that the Galerkin matrix BM is symmetric and positive definite, and
therefore invertible. Therefore we can define an approximation of the precon-
ditioning matrix CA by
A := M ⊤ B −1 MM .
C (13.7)
M M
A is spec-
We need to prove that the approximated preconditioning matrix C
trally equivalent to CA , and therefore to AM .
Lemma 13.5. Let CA be the Galerkin matrix of B −1 as defined in (13.5),
A be the approximation as given in (13.7). Then there holds
and let C
A v, v) ≤ (CA v, v)
(C for v ∈ RM .
13.2 A General Preconditioning Strategy 301
Note that
(CA v, v) = B −1 vM , vM = w, vM = Bw, w. (13.8)
−1 ′
In the same way we define w = BM MM v ↔ wM ∈ XM as the unique solution
of the Galerkin variational problem
′
BwM , zM = vM , zM for zM ∈ XM ,
Again,
A v, v) = (B −1 MM v, MM v) = (w, MM v) = wM , vM = BwM , wM .
(C M
(13.9)
Moreover we have the Galerkin orthogonality
′
B(w − wM ), zM = 0 for zM ∈ XM .
0 ≤ cB 2
1 w − wM X ′ ≤ B(w − wM ), w − wM
and therefore
BwM , wM ≤ Bw, w.
By using (13.8) and (13.9) this finally gives the assertion. ⊓
⊔
Note that Lemma 13.5 holds for any arbitrary conforming trial spaces
′
XM ⊂ X and XM ⊂ X ′ . However, to prove the reverse estimate we need to
′
assume a certain stability condition of the trial space XM ⊂ X ′.
Then,
2
cB
1 A v, v)
cS (CA v, v) ≤ (C for all v ∈ RM .
cB
2
and hence
2 2
1 cB
2 1 cB
2
(CA v, v) ≤ wM 2X ′ ≤ BwM , wM .
cB
1 cS cS cB
1
The bilinear form of the inverse single layer potential can then be written as
[v, weq Γ ]2
V −1 v, vΓ = V −1 v, vΓ + .
1, weq Γ
vΓ ≤ 1 −1
cV1 cD
1 V
−1
v, vΓ ≤ Dv, V v, vΓ
4
for all v ∈ H 1/2 (Γ ).
By using Corollary 13.10 we now can define a preconditioner for the linear
system (12.15) of the Dirichlet boundary value problem, and for the sys-
tem (12.27) of the Neumann boundary value problem. The system matrix in
h := Dh + α a a⊤ where
(12.27) is D
for i, j = 1, . . . , M .
13.2 A General Preconditioning Strategy 305
Then,
≤ cQ wH −1/2 (Γ ) ,
h . In par-
defines a preconditioning matrix which is spectrally equivalent to D
ticular there hold the spectral equivalence inequalities
2
h v, v) ≤ 1 cV2
cV1 cD
1 (CD v, v) ≤ (D cQ V (CD v, v) for all v ∈ RM .
4 c1
In Table 13.2 the extremal eigenvalues and the resulting spectral condition
−1
numbers of the preconditioned system matrix CD Dh are listed for the L–
shaped domain as given in Fig. 10.1. For comparison we also give the corre-
sponding values in the case of a simple diagonal preconditioning which show
a linear dependency on the inverse mesh parameter h−1 .
By applying Corollary 13.10 we can use the Galerkin discretization of
the modified hypersingular boundary integral operator D as a preconditioner
for the discrete single layer potential Vh in (12.15). However, when using
piecewise constant basis functions to discretize the single layer potential, for
the Galerkin discretization of the hypersingular boundary integral operator
306 13 Iterative Solution Methods
- = diag Dh
CD CD −1
- = M̄h V̄h M̄h
−1 −1
L N λmin λmax κ(CD - Dh ) λmin λmax κ(CD - Dh )
0 28 9.05 –3 2.88 –2 3.18 1.02 –1 2.56 –1 2.50
1 112 4.07 –3 2.82 –2 6.94 9.24 –2 2.66 –1 2.88
2 448 1.98 –3 2.87 –2 14.47 8.96 –2 2.82 –1 3.14
3 1792 9.84 –3 2.90 –2 29.52 8.86 –2 2.89 –1 3.26
4 7168 4.91 –3 2.91 –2 59.35 8.80 –2 2.92 –1 3.31
5 28672 2.46 –4 2.92 –2 118.72 8.79 –2 2.92 –1 3.32
6 114688 1.23 –4 2.92 –2 237.66 8.78 –2 2.92 –1 3.33
Theory: O(h−1 ) O(1)
By construction we have
where Xh = Sh1 (Ω) ⊂ H 1 (Ω) is the trial space to be used for the Galerkin
−1 (Ω) which is induced by the
discretization of the operator A : H 1 (Ω) → H
bilinear form a(·, ·), i.e.
cB 2 B 2
1 f H −1 (Ω) ≤ Bf, f Ω ≤ c2 f H −1 (Ω) (13.13)
for all f ∈ H −1 (Ω) with some positive constants cB and cB . Such an op-
1 2
erator can be constructed when using an appropriately weighted multilevel
representation of L2 projection operators, see [28, 162].
For any trial space Vj ⊂ H 1 (Ω) let Qj : L2 (Ω) → Vj be the L2 projection
operator as defined in (9.23), i.e. Qj u ∈ Vj is the unique solution of the
variational problem
vj H 1 (Ω) ≤ cI h−1
j vj L2 (Ω) for all vj ∈ Vj . (13.15)
and therefore
Qk Qj v, vk L2 (Ω) = Qj v, vk L2 (Ω) = v, vk L2 (Ω) = Qk v, vk L2 (Ω)
cB 2 1 B 2
1 vH 1 (Ω) ≤ B v, vL2 (Ω) ≤ c2 vH 1 (Ω)
From the inverse inequalities of the trial spaces Vk and by using the error
estimates of the L2 projection operators Qk we further obtain from Corollary
13.14:
Lemma 13.15. For all v ∈ H 1 (Ω) there hold the spectral equivalence inequal-
ities
∞
∞
c1 (Qk −Qk−1 )v2H 1 (Ω) ≤ B 1 v, vL2 (Ω) ≤ c2 (Qk −Qk−1 )v2H 1 (Ω) .
k=0 k=0
Proof. By using Lemma 13.12, 3., the triangle inequality, the error estimate
(13.14), and assumption (13.12) we have
∞
B 1 v, vL2 (Ω) = h−2 2
k (Qk − Qk−1 )vL2 (Ω)
k=0
∞
= h−2 2
k (Qk − Qk−1 )(Qk − Qk−1 )vL2 (Ω)
k=0
∞ !
≤ 2 h−2
k (Qk − I)(Qk − Qk−1 )v2L2 (Ω)
k=0
"
+ (I − Qk−1 )(Qk − Qk−1 )v2L2 (Ω)
∞
! "
≤ 2c h−2
k h2k (Qk − Qk−1 )v2H 1 (Ω) + h2k−1 (Qk − Qk−1 )v2H 1 (Ω)
k=0
∞
≤ c2 (Qk − Qk−1 )v2H 1 (Ω)
k=0
and therefore the upper estimate. To prove the lower estimate we get from
the global inverse inequality (13.15) for (Qk − Qk−1 )v ∈ Vk−1 , and by using
assumption (13.12),
310 13 Iterative Solution Methods
∞
∞
(Qk − Qk−1 )v2H 1 (Ω) ≤ c2I h−2 2
k−1 (Qk − Qk−1 )vL2 (Ω)
k=0 k=0
∞
≤c h−2 2
k (Qk − Qk−1 )vL2 (Ω)
k=0
1
= c B v, vL2 (Ω) . ⊓
⊔
The statement of Theorem 13.13 now follows from Lemma 13.15 and from the
following spectral equivalence inequalities.
Lemma 13.16. For all v ∈ H 1 (Ω) there hold the spectral equivalence inequal-
ities
∞
c̄1 v2H 1 (Ω) ≤ (Qk − Qk−1 )v2H 1 (Ω) ≤ c̄2 v2H 1 (Ω) .
k=0
To prove Lemma 13.16 we first need a tool to estimate some matrix norms.
Hence we have
13.2 A General Preconditioning Strategy 311
# $# $
|vℓ |2 ≤ |A[ℓ, k]| 2α(k−ℓ) |A[ℓ, k]| 2α(ℓ−k) u2k
ℓ∈I ℓ∈I k∈I k∈I
# $ # $
α(k−ℓ) α(ℓ−k)
≤ sup |A[ℓ, k]| 2 |A[ℓ, k]| 2 u2k
ℓ∈I
k∈I ℓ∈I k∈I
# $ # $
α(k−ℓ) α(ℓ−k)
= sup |A[ℓ, k]| 2 |A[ℓ, k]| 2 u2k
ℓ∈I
k∈I k∈I ℓ∈I
# $ # $
≤ sup |A[ℓ, k]| 2α(k−ℓ) sup |A[ℓ, k]| 2α(ℓ−k) u2k
ℓ∈I k∈I
k∈I ℓ∈I k∈I
(Qi − Qi−1 )v, (Qj − Qj−1 )vH 1 (Ω) = (Qi − Qi−1 )v, Q1j (Qj − Qj−1 )vH 1 (Ω)
= Q1j (Qi − Qi−1 )v, (Qj − Qj−1 )vH 1 (Ω)
≤ Q1j (Qi − Qi−1 )vH 1 (Ω) (Qj − Qj−1 )vH 1 (Ω) .
Due to Vj = Sh1j (Ω) ⊂ H 1+σ (Ω) the H 1 projection as given in (9.29) is well
defined for u ∈ H 1−σ (Ω) and for σ ∈ (0, 12 ). Dependent on the regularity
312 13 Iterative Solution Methods
= cI h−s
j (Qi − Qi−1 )(Qi − Qi−1 )vH 1−s (Ω)
0
≤ cI h−s
j (Qi − I)(Qi − Qi−1 )vH 1−s (Ω)
1
+(I − Qi−1 )(Qi − Qi−1 )vH 1−s (Ω)
0 s 1
≤ c h−s
j hi + hsi−1 (Qi − Qi−1 )vH 1 (Ω)
For i < k we therefore have vi = (Q1i − Q1i−1 )v ∈ Vi−1 ⊂ Vk−1 , and thus
(Qk − Qk−1 )vi = 0. Hence, by interchanging the order of summation,
∞
∞
∞
(Qk − Qk−1 )v2H 1 (Ω) = (Qk − Qk−1 )vi , (Qk − Qk−1 )vj H 1 (Ω)
k=0 k=0 i,j=0
∞ min{i,j}
= (Qk − Qk−1 )vi , (Qk − Qk−1 )vj H 1 (Ω)
i,j=0 k=0
∞ min{i,j}
≤ (Qk − Qk−1 )vi H 1 (Ω) (Qk − Qk−1 )vj H 1 (Ω) .
i,j=0 k=0
13.2 A General Preconditioning Strategy 313
By using the global inverse inequality (13.15), the stability of the L2 projection
(see Remark 9.14), and applying some interpolation argument, we obtain from
assumption (13.12) for the already fixed parameter s ∈ (0, 1] the estimate
Moreover,
Hence we obtain
∞ ∞ min{i,j}
(Qk − Qk−1 )v2H 1 (Ω) ≤ c hk−2s hsi hsj vi H 1 (Ω) vj H 1 (Ω) .
k=0 i,j=0 k=0
Finally,
∞
∞
vi 2H 1 (Ω) = (Q1i − Q1i−1 )v, (Q1i − Q1i−1 )vH 1 (Ω)
i=0 i=0
∞
= (Q1i − Q1i−1 )(Q1i − Q1i−1 )v, vH 1 (Ω)
i=0
∞
= (Q1i − Q1i−1 )v, vH 1 (Ω) = v, vH 1 (Ω) = v2H 1 (Ω) ,
i=0
Remark 13.19. For s ∈ [0, 32 ) we may define the more general multilevel oper-
ator
∞
B s := hk−2s (Qk − Qk−1 )
k=0
Although the following considerations are only done for the special case s = 1,
these investigations can be extended to the more general case s ∈ [0, 23 ), too.
By using Theorem 13.13 the multilevel operator B 1 : H 1 (Ω) → H −1 (Ω)
1
is bounded and H (Ω)–elliptic. The inverse operator (B ) 1 −1
: H (Ω) →
−1
1 −1
H (Ω) is then bounded and H (Ω)–elliptic, in particular the spectral equiv-
alence inequalities (13.13) are valid. For the inverse operator (B 1 )−1 again a
multilevel representation can be given.
where s ∈ (− 12 , 21 ).
By using corollary 13.7 we can now establish the spectral equivalence of the
system matrix AhL with the discrete preconditioning matrix
A = M̄h B −1 M̄h
C L hL L
where
BhL [ℓ, k] = B −1 ϕL L
k , ϕℓ L2 (Ω) , M̄hL [ℓ, k] = ϕL L
k , ϕℓ L2 (Ω)
for all ϕL L 1
k , ϕℓ ∈ VL = ShL (Ω).
−1 r inside the algorithm of
It remains to describe the application v = CA
the preconditioned method of conjugate gradients, see Algorithm 13.3. There
we have to compute
−1 r = M −1 Bh M −1 r,
v := CA hL L hL
316 13 Iterative Solution Methods
or,
u := Mh−1
L
r, w := BhL u, v := Mh−1
L
w.
By using the isomorphism u ∈ RML ↔ uhL ∈ VL we obtain for the components
of w = BhL u
ML
ML
wℓ := BhL [ℓ, k]uk = B −1 ϕL L
k , ϕℓ L2 (Ω) uk = B
−1
uhL , ϕL
ℓ L2 (Ω) .
k=1 k=1
which is a finite sum due to Qk uhL = uhL for k ≥ L. For the components of
w = BhL u we then obtain
ML
−1
wℓ = B uhL , ϕL
ℓ L2 (Ω) = zhL , ϕL
ℓ L2 (Ω) = zk ϕL L
k , ϕℓ L2 (Ω) .
k=1
This is equivalent to
w = MhL z,
and therefore there is no need to invert the inverse mass matrix MhL when
computing the preconditioned residual,
v = Mh−1
L
w = Mh−1
L
MhL z = z .
where
Mk
ūhk = Qk uhL = ūkℓ ϕkℓ ∈ Vk
ℓ=1
we can define
L
z̄hL := h2k ūhk
k=0
Due to the inclusion Vk−1 ⊂ Vk we can write each basis function ϕk−1
ℓ ∈ Vk−1
as a linear combination of basis functions ϕkj ∈ Vk ,
Mk
ϕℓk−1 = k
rℓ,j ϕkj for all ℓ = 1, . . . , Mk−1 .
j=1
we obtain by induction
L
z̄ L = h2k Rk ūk .
k=0
ūhk , ϕkℓ L2 (Ω) = uhL , ϕkℓ L2 (Ω) for all ϕkℓ ∈ Vk .
318 13 Iterative Solution Methods
0 0 0 0 0
@
@ @
@ 1 @ 1
@ 2
@ @2 @
@ @ @
@ @ @0
@1
0
@ @
1
@ @ @
@ @
1 @1
@ 2
@
2@
@ @ @
0 0 0 0 0
Fig. 13.1. Basis functions ϕk−1
ℓ and coefficients k
rℓ,j (d = 2).
Mhk ūk = f k
where
Mhk [ℓ, j] = ϕkj , ϕkℓ L2 (Ω) , fℓk = uhL , ϕkℓ L2 (Ω) .
In particular for k = L we have
and therefore
ūL = Mh−1
L
r.
Due to
Mk
M
fℓk−1 = uhL , ϕℓk−1 L2 (Ω) = k
rℓ,j uhL , ϕkj L2 (Ω) = k
rℓ,j fjk
j=1 j=1
we get
f k−1 = Rk−1,k
⊤
f k = Rk−1
⊤
r.
By recursion we therefore have
ūk = Mh−1
k
Rk⊤ r,
Taking into account the spectral equivalence of the mass matrices with the
diagonal matrices hdk I, see Lemma 9.7, we then obtain for the application of
the multilevel preconditioner
L
v = hk2−d Rk Rk⊤ r. (13.19)
k=0
13.3 Solution Methods for Saddle Point Problems 319
CA = I CA = M̄hL Bh−1
L
M̄hL
−1 −1
L M λmin λmax κ(CA Ah ) λmin λmax κ(CA Ah )
1 13 2.88 –1 6.65 23.13 16.34 130.33 7.98
2 41 8.79 –2 7.54 85.71 16.69 160.04 9.59
3 145 2.42 –2 7.87 324.90 16.32 179.78 11.02
4 545 6.34 –3 7.96 1255.75 15.47 193.36 12.50
5 2113 1.62 –3 7.99 4925.47 15.48 202.94 13.11
6 8321 4.10 –4 8.00 19496.15 15.58 209.85 13.47
7 33025 ≈80000 15.76 214.87 13.63
8 131585 ≈320000 15.87 218.78 13.79
9 525313 ≈1280000 15.96 221.65 13.89
Theory: O(h−2 ) O(1)
In Table 13.3 we give the extremal eigenvalue and the resulting spec-
tral condition numbers of the preconditioned finite element stiffness matrix
−1
CA [a a⊤ +AhL ]. This preconditioner is also needed for an efficient solve of the
linear system (11.22), as it will be considered in the next section. The results
for the non–preconditioned system (CA = I) confirm the statement of Lemma
11.4, while the boundedness of the spectral condition of the preconditioned
systems coincides with the results of this section.
where the block A ∈ RM1 ×M1 is symmetric and positive definite, and
where D ∈ RM2 ×M2 is symmetric but positive semi–definite. Accordingly,
B ∈ RM1 ×M2 . Since the matrix A is assumed to be positive definite, we can
solve the first equation in (13.20) for u1 to obtain
320 13 Iterative Solution Methods
Inserting this into the second equation of (13.20) this results in the Schur
complement system
0 1
D + B ⊤ A−1 B u2 = f 2 − B ⊤ A−1 f 1 (13.21)
where
S = D + B ⊤ A−1 B ∈ RM2 ×M2 . (13.22)
is the Schur complement. From the symmetry properties of the block matrices
A, B and D we conclude the symmetry of S, while, at this point, we assume
the positive definiteness of S.
We assume that for the symmetric and positive definite matrices A and
S = D + B ⊤ A−1 B there are given some positive definite and symmetric
preconditioning matrices CA and CS satisfying the spectral equivalence in-
equalities
cA A
1 (CA x1 , x1 ) ≤ (Ax1 , x1 ) ≤ c2 (CA x1 , x1 ) (13.23)
for all x1 ∈ RM1 as well as
for all x2 ∈ RM2 . Hence, to solve the Schur complement system (13.21) we
can apply the CS preconditioned method of conjugate gradients (Algorithm
13.3). There, the matrix by vector multiplication sk = Spk for the Schur
complement (13.22) reads
Awk = Bpk .
This system can be solved either by a direct method, for example by the
Cholesky approach, or again by using a CA preconditioned method of con-
jugate gradients (Algorithm 13.3). Depending on the application under con-
sideration the Schur complement approach can be disadvantageous. Then, an
iterative solution strategy for the system (13.20) should be used. Possible iter-
ative solution methods for general non–symmetric linear systems of the form
(13.20) are the method of the generalized minimal residual (GMRES, [120]),
or the stabilized method of biorthogonal search directions (BiCGStab, [155]).
Here, following [26], we will describe a transformation of the block–skew
symmetric but positive definite system (13.20) leading to a symmetric and
positive definite system for which a preconditioned conjugate gradient ap-
proach can be used.
For the preconditioning matrix CA we need to assume that the spectral equiv-
alence inequalities (13.23) hold where
13.3 Solution Methods for Saddle Point Problems 321
cA
1 > 1 (13.25)
cM M
1 (CM x, x) ≤ (M x, x) ≤ c2 (CM x, x) for all x ∈ RM1 +M2
where
322 13 Iterative Solution Methods
8
1 A 1 A
cM
1
S
= c2 [1 + c1 ] − [c (1 + cS1 )]2 − cS1 cA
2 ,
2 4 2
8
1 1 A
cM
2 = cA [1 + cS2 ] + [c (1 + cS2 )]2 − cS2 cA
2 .
2 2 4 2
Proof. We need to estimate the extremal eigenvalues of the preconditioned
−1
system matrix CM M , in particular we have to consider the eigenvalue prob-
lem
& '& ' & '& '
−1 −1
ACA A − A (I − ACA )B x1 A − CA 0 x1
−1 −1 = λ .
B ⊤ (I − CA A) D + B ⊤ CA B x2 0 CS x2
Let λi be an eigenvalue with associated eigenvectors x1i and xi2 . From the first
equation,
−1 −1
(ACA A − A)xi1 + (I − ACA )Bxi2 = λi (A − CA )xi1 ,
For λi ∈ [1, cA A
2 ] nothing is to be shown. Hence we only consider λi ∈ [1, c2 ]
were λi CA − A is invertible. Thus,
Inserting this result into the second equation of the eigenvalue problem,
−1 −1
B ⊤ (I − CA A)xi1 + [D + B ⊤ CA B]xi2 = λi CS xi2 ,
this gives
−1 −1
−B ⊤ CA (CA − A)(λi CA − A)−1 Bxi2 + [D + B ⊤ CA B]xi2 = λi CS xi2 .
Due to
−1 −1
−CA (CA − A)(λi CA − A)−1 = −CA [λi CA − A + (1 − λi )CA ](λi CA − A)−1
−1
= (λi − 1)(λi CA − A)−1 − CA
this is equivalent to
λi − cA
2
(Ax1 , x1 ) ≤ ((λi CA − A)x1 , x1 )
cA
2
13.3 Solution Methods for Saddle Point Problems 323
cA
2
((λi CA − A)−1 x1 , x1 ) ≤ (A−1 x1 , x1 ) for all x1 ∈ Rn1 .
λi − cA
2
λi
(Sxi2 , xi2 ) ≤ λi (CS xi2 , xi2 )
cS2
= (Dxi2 , xi2 ) + (λi − 1)((λi CA − A)−1 Bxi2 , Bxi2 )
cA
2
≤ (Dxi2 , xi2 ) + (λi − 1) (A−1 Bxi2 , Bxi2 )
λi − cA
2
A λi − 1 i i
≤ c2 (Sx2 , x2 )
λi − cA 2
and therefore
λi λi − 1
S
≤ cA
2 ,
c2 λi − cA
2
i.e.
λ2i − cA S S A
2 [1 + c2 ]λi + c2 c2 ≤ 0 .
cA
2 − λi
((A − λi CA )x1 , x1 ) ≤ (Ax1 , x1 )
cA
2
and therefore
cA
2
((A − λi CA )−1 x1 , x1 ) ≥ (A−1 x1 , x1 )
cA
2 − λi
for all x1 ∈ RM1 . Again, by using the spectral equivalence inequalities (13.24)
we conclude
324 13 Iterative Solution Methods
λi
(Sxi2 , xi2 ) ≥ λi (CS xi2 , xi2 )
cS1
= (Dxi2 , xi2 ) + (1 − λi )((A − λi CA )−1 Bxi2 , Bxi2 )
cA
≥ (Dxi2 , xi2 ) + (1 − λi ) A 2 (A−1 Bxi2 , Bxi2 )
c2 − λi
cA
2
≥ (1 − λi ) (Sxi2 , xi2 )
cA
2 − λi
and therefore
1 − λi λi
cA
2 A
≤ S.
c2 − λi c1
This is equivalent to
λ2i − cA S S A
2 [c1 + 1]λi + c1 c2 ≤ 0,
i.e.
λ− ≤ λi ≤ λ+
where 8
1 A S 1 A
λ± = c2 [1 + c1 ] ± [c (1 + cS1 )]2 − cS1 cA
2.
2 4 2
Summarizing we have
8
1 A S 1 A M
1 > λi ≥ c2 [1 + c1 ] − [c (1 + cS1 )]2 − cS1 cA
2 = c1 .
2 4 2
This completes the proof. ⊓ ⊔
For the solution of the transformed linear system (13.26) Algorithm
13.3 of the preconditioned conjugate gradient approach can be applied. On
a first glance the multiplication with the inverse preconditioning matrix
−1 k+1
v k+1 = CM r , in particular the evaluation of v k+1
1 = (A−CA )−1 rk+1 seems
to be difficult. However, from the recursion of the residual,
rk+1 = rk − αk M pk , we find the representation
rk+1
1
−1
:= rk1 − αk (ACA − I)(Apk1 − Bpk2 ).
v k+1
1
−1
:= v k1 − αk CA (Apk1 − Bpk2 ).
When using the preconditioning matrix CD = M̄h V̄h−1 M̄h we then obtain the
estimate for the spectral condition number,
−1
κ2 (CD Dh ) ≤ c [1 + log |h|]2 .