Cours 4 Part 2
Cours 4 Part 2
Cours 4 Part 2
z(k) = Aq(k−1) ,
q(k) = z(k) /�z(k) �2 , (5.17)
ν (k) = (q(k) )H Aq(k) .
(k) Ak q(0)
q = , k ≥ 1. (5.18)
�Ak q(0) �2
This relation explains the role played by the powers of A in the method.
Because A is diagonalizable, its eigenvectors xi form a basis of Cn ; it is thus
possible to represent q(0) as
5.3 The Power Method 193
n
�
(0)
q = αi x i , αi ∈ C, i = 1, . . . , n. (5.19)
i=1
Since |λi /λ1 | < 1 for i = 2, . . . , n, as k increases the vector Ak q(0) (and
thus also q(k) , due to (5.18)), tends to assume an increasingly significant
component in the direction of the eigenvector x1 , while its components in the
other directions xj decrease. Using (5.18) and (5.20), we get
where µk is the sign of α1 λk1 and y(k) denotes a vector that vanishes as k → ∞.
As k → ∞, the vector q(k) thus aligns itself along the direction of eigen-
vector x1 , and the following error estimate holds at each step k.
where
�n � �k
(k) q(k) �Ak q(0) �2 αi λi
q̃ = = x1 + xi , k = 1, 2, . . . (5.22)
α1 λk1 i=2
α 1 λ 1
� n
�1/2
�
that is (5.21) with C = (αi /α1 )2 . �
i=2
194 5 Approximation of Eigenvalues and Eigenvectors
where cos(θ0 ) = |xT1 q(0) | �= 0. Inequality (5.23) outlines that the convergence
of the sequence ν (k) to λ1 is quadratic with respect to the ratio |λ2 /λ1 | (we
refer to Section 5.3.3 for numerical results).
We conclude the section by providing a stopping criterion for the iteration
(5.17). For this purpose, let us introduce the residual at step k
�r(k) �2
|λ1 − ν (k) | � , k ≥ 1, (5.25)
| cos(θλ )|
where θλ is the angle between the right and the left eigenvectors, x1 and y1 ,
associated with λ1 . Notice that, if A is an hermitian matrix, then cos(θλ ) = 1,
so that (5.25) yields an estimate which is analogue to (5.13).
In practice, in order to employ the estimate (5.25) it is necessary at each
step k to replace | cos(θλ )| with the module of the scalar product between two
approximations q(k) and w(k) of x1 and y1 , computed by the power method.
The following a posteriori estimate is thus obtained
�r(k) �2
|λ1 − ν (k) | � , k ≥ 1. (5.26)
|(w(k) )H q(k) |
where xi are the eigenvectors of M−1 µ (and thus also of A), while αi are as
in (5.19). As a consequence, recalling the definition of ξi and using (5.27), we
get
lim q̃(k) = xm , lim σ (k) = λm .
k→∞ k→∞
(k) ��r(k) �2
|λm − σ |� , k ≥ 1, (5.29)
� (k) )H q(k) |
|(w
The convergence analysis of Section 5.3.1 shows that the effectiveness of the
power method strongly depends on the dominant eigenvalues being well-
separated (that is, |λ2 |/|λ1 | � 1). Let us now analyze the behavior of
iteration (5.17) when two dominant eigenvalues of equal module exist (that
is, |λ2 | = |λ1 |). Three cases must be distinguished:
1. λ2 = λ1 : the two dominant eigenvalues are coincident. The method is still
convergent, since for k sufficiently large (5.20) yields
Matrix A has the following eigenvalues (to five significant figures): λ1 = 14.103,
λ2 = 10.385 and λ3 = 0.512, while the corresponding eigenvectors are the vector
columns of matrix V.
To approximate the pair (λ1 , x1 ), we have run the Program 26 with initial datum
z(0) = [1, 1, 1]T . After 71 iterations of the power method the absolute errors are
(71)
|λ1 − ν (71) | = 2.2341 · 10−10 and �x1 − x1 �∞ = 1.42 · 10−11 .
In a second run, we have used z(0) = x2 + x3 (notice that with this choice
α1 = 0). After 215 iterations the absolute errors are |λ1 − ν (215) | = 4.26 · 10−14 and
(215)
�x1 − x1 �∞ = 1.38 · 10−14 .
Figure 5.2 (left) shows the reliability of the a posteriori estimate (5.26). The se-
quences |λ1 − ν (k) | (solid line) and the corresponding a posteriori estimates (5.26)
5.4 The QR Iteration 199
100 100
10−2
10−2
10−4 (NS)
10−4
10−6
10−6
10−8
10−8
(S)
10−10
10−10 10−12
10−12 10−14
0 10 20 30 40 50 60 70 80 0 6 12 18
Fig. 5.2. Comparison between the a posteriori error estimate and the actual ab-
solute error for matrix A in (5.30) (left); convergence curves for the power method
applied to matrix A in (5.31) in its symmetric (S) and nonsymmetric (NS) forms
(right)
(dashed line) are plotted as a function of the number of iterations (in abscissae).
Notice the excellent agreement between the two curves.
Let us now consider the matrices
134 816
A = 3 1 2, T = 3 5 7 (5.31)
421 492
where A has the following spectrum: λ1 = 7.047, λ2 = −3.1879 and λ3 = −0.8868
(to five significant figures).
It is interesting to compare the behaviour of the power method when computing λ1
for the symmetric matrix A and for its similar matrix M = T−1 AT, where T is the
nonsingular (and nonorthogonal) matrix in (5.31).
Running Program 26 with z(0) = [1, 1, 1]T , the power method converges to the eigen-
value λ1 in 18 and 30 iterations, for matrices A and M, respectively. The sequence
of absolute errors |λ1 − ν (k) | is plotted in Figure 5.2 (right) where (S) and (NS) refer
to the computations on A and M, respectively. Notice the rapid error reduction in
the symmetric case, according to the quadratic convergence properties of the power
method (see Section 5.3.1).
We finally employ the inverse power method (5.28) to compute the eigenvalue of
smallest module
√ λ3 = 0.512 of matrix A in (5.30). Running Program 27 with q(0) =
[1, 1, 1] / 3, the method converges in 9 iterations, with absolute errors |λ3 −σ (9) | =
T
(9)
1.194 · 10−12 and �x3 − x3 �∞ = 4.59 · 10−13 . •
At each step k ≥ 1, the first phase of the iteration is the factorization of the
matrix T(k−1) into the product of an orthogonal matrix Q(k) with an upper
triangular matrix R(k) (see Section 5.6.3). The second phase is a simple matrix
product. Notice that
T(k) = R(k) Q(k) = (Q(k) )T (Q(k) R(k) )Q(k) = (Q(k) )T T(k−1) Q(k)
(5.33)
= (Q(0) Q(1) · · · Q(k) )T A(Q(0) Q(1) · · · Q(k) ), k ≥ 0,
where each block Rii is either a real number or a matrix of order 2 having
complex conjugate eigenvalues, and
� �
Q = lim Q(0) Q(1) · · · Q(k) (5.35)
k→∞
Q(k) being the orthogonal matrix generated by the k-th factorization step of
the QR iteration (5.32).
In the basic version of the QR method, one sets Q(0) = In in such a way
that T(0) = A. At each step k ≥ 1 the QR factorization of the matrix T(k−1)
can be carried out using the modified Gram-Schmidt procedure introduced in
Section 3.4.3, with a cost of the order of 2n3 flops (for a full matrix A). The
following convergence result holds (for the proof, see [GL89], Theorem 7.3.1,
or [Wil65], pp. 517-519).
Then
λ1 t12 . . . t1n
0 λ2 t23 . . .
lim T(k) =
.. .. . . .. .
(5.36)
k→+∞ .
. ..
0 0 . . . λn
As for the convergence rate, we have
�� �k �
(k) � λ i ��
|ti,i−1 | = O �� , i = 2, . . . , n, for k → +∞. (5.37)
λi−1 �
Example 5.4 We apply the QR method to the symmetric matrix A∈ R4×4 such
that aii = 4, for i = 1, . . . , 4, and aij = 4 + i − j for i < j ≤ 4, whose eigenvalues are
(to three significant figures) λ1 = 11.09, λ2 = 3.41, λ3 = 0.90 and λ4 = 0.59. After
20 iterations, we get
11.09 6.44 · 10−10 −3.62 · 10−15 9.49 · 10−15
6.47 · 10−10 3.41 1.43 · 10−11 4.60 · 10−16
T(20) = .
1.74 · 10−21 1.43 · 10−11 0.90 1.16 · 10−4
2.32 · 10−25 2.68 · 10−15 1.16 · 10−4 0.58