Loading [MathJax]/jax/output/HTML-CSS/jax.js
Skip to main content
Library homepage
 

Text Color

Text Size

 

Margin Size

 

Font Type

Enable Dyslexic Font
Mathematics LibreTexts

8.5: Computing Eigenvalues

( \newcommand{\kernel}{\mathrm{null}\,}\)

In practice, the problem of finding eigenvalues of a matrix is virtually never solved by finding the roots of the characteristic polynomial. This is difficult for large matrices and iterative methods are much better. Two such methods are described briefly in this section.

The Power Method

In Chapter [chap:3] our initial rationale for diagonalizing matrices was to be able to compute the powers of a square matrix, and the eigenvalues were needed to do this. In this section, we are interested in efficiently computing eigenvalues, and it may come as no surprise that the first method we discuss uses the powers of a matrix.

Recall that an eigenvalue λ of an n×n matrix A is called a dominant eigenvalue if λ has multiplicity 1, and

|λ|>|μ| for all eigenvalues μλ

Any corresponding eigenvector is called a dominant eigenvector of A. When such an eigenvalue exists, one technique for finding it is as follows: Let x0 in Rn be a first approximation to a dominant eigenvector λ, and compute successive approximations x1,x2, as follows:

x1=Ax0x2=Ax1x3=Ax2

In general, we define

xk+1=Axk for each k0

If the first estimate x0 is good enough, these vectors xn will approximate the dominant eigenvector λ (see below). This technique is called the power method (because xk=Akx0 for each k1). Observe that if z is any eigenvector corresponding to λ, then

z(Az)z2=z(λz)z2=λ

Because the vectors x1,x2,,xn, approximate dominant eigenvectors, this suggests that we define the Rayleigh quotients as follows:

rk=xkxk+1xk2 for k1

Then the numbers rk approximate the dominant eigenvalue λ.

025326 Use the power method to approximate a dominant eigenvector and eigenvalue of A=[1120].

The eigenvalues of A are 2 and 1, with eigenvectors [11] and [12]. Take x0=[10] as the first approximation and compute x1,x2,, successively, from x1=Ax0,x2=Ax1,. The result is

x1=[12], x2=[32], x3=[56], x4=[1110], x3=[2122], 

These vectors are approaching scalar multiples of the dominant eigenvector [11]. Moreover, the Rayleigh quotients are

r1=75,r2=2713,r3=11561,r4=451221,

and these are approaching the dominant eigenvalue 2.

To see why the power method works, let λ1,λ2,,λm be eigenvalues of A with λ1 dominant and let y1,y2,,ym be corresponding eigenvectors. What is required is that the first approximation x0 be a linear combination of these eigenvectors:

x0=a1y1+a2y2++amym with a10

If k1, the fact that xk=Akx0 and Akyi=λkiyi for each i gives

xk=a1λk1y1+a2λk2y2++amλkmym for k1

Hence

1λk1xk=a1y1+a2(λ2λ1)ky2++am(λmλ1)kym

The right side approaches a1y1 as k increases because λ1 is dominant (|λiλ1|<1 for each i>1). Because a10, this means that xk approximates the dominant eigenvector a1λk1y1.

The power method requires that the first approximation x0 be a linear combination of eigenvectors. (In Example [exa:025326] the eigenvectors form a basis of R2.) But even in this case the method fails if a1=0, where a1 is the coefficient of the dominant eigenvector (try x0=[12] in Example [exa:025326]). In general, the rate of convergence is quite slow if any of the ratios |λiλ1| is near 1. Also, because the method requires repeated multiplications by A, it is not recommended unless these multiplications are easy to carry out (for example, if most of the entries of A are zero).

QR-Algorithm

A much better method for approximating the eigenvalues of an invertible matrix A depends on the factorization (using the Gram-Schmidt algorithm) of A in the form

A=QR

where Q is orthogonal and R is invertible and upper triangular (see Theorem [thm:025166]). The QR-algorithm uses this repeatedly to create a sequence of matrices A1=A,A2,A3,, as follows:

  1. Define A1=A and factor it as A1=Q1R1.
  2. Define A2=R1Q1 and factor it as A2=Q2R2.
  3. Define A3=R2Q2 and factor it as A3=Q3R3.

In general, Ak is factored as Ak=QkRk and we define Ak+1=RkQk. Then Ak+1 is similar to Ak [in fact, Ak+1=RkQk=(Q1kAk)Qk], and hence each Ak has the same eigenvalues as A. If the eigenvalues of A are real and have distinct absolute values, the remarkable thing is that the sequence of matrices A1,A2,A3, converges to an upper triangular matrix with these eigenvalues on the main diagonal. [See below for the case of complex eigenvalues.]

025425 If A=[1120] as in Example [exa:025326], use the QR-algorithm to approximate the eigenvalues.

The matrices A1, A2, and A3 are as follows:

A1=[1120]=Q1R1 where Q1=15[1221] and R1=15[5102]A2=15[7942]=[1.41.80.80.4]=Q2R2 where Q2=165[7447] and R2=165[1311010]A3=113[275814]=[2.080.380.621.08]

This is converging to [201] and so is approximating the eigenvalues 2 and 1 on the main diagonal.

It is beyond the scope of this book to pursue a detailed discussion of these methods. The reader is referred to J. M. Wilkinson, The Algebraic Eigenvalue Problem (Oxford, England: Oxford University Press, 1965) or G. W. Stewart, Introduction to Matrix Computations (New York: Academic Press, 1973). We conclude with some remarks on the QR-algorithm.

Shifting. Convergence is accelerated if, at stage k of the algorithm, a number sk is chosen and AkskI is factored in the form QkRk rather than Ak itself. Then

Q1kAkQk=Q1k(QkRk+skI)Qk=RkQk+skI

so we take Ak+1=RkQk+skI. If the shifts sk are carefully chosen, convergence can be greatly improved.

Preliminary Preparation. A matrix such as

[000000]

is said to be in upper Hessenberg form, and the QR-factorizations of such matrices are greatly simplified. Given an n×n matrix A, a series of orthogonal matrices H1,H2,,Hm (called Householder matrices) can be easily constructed such that

B=HTmHT1AH1Hm

is in upper Hessenberg form. Then the QR-algorithm can be efficiently applied to B and, because B is similar to A, it produces the eigenvalues of A.

Complex Eigenvalues. If some of the eigenvalues of a real matrix A are not real, the QR-algorithm converges to a block upper triangular matrix where the diagonal blocks are either 1×1 (the real eigenvalues) or 2×2 (each providing a pair of conjugate complex eigenvalues of A).


This page titled 8.5: Computing Eigenvalues is shared under a CC BY-NC-SA 4.0 license and was authored, remixed, and/or curated by W. Keith Nicholson (Lyryx Learning Inc.) via source content that was edited to the style and standards of the LibreTexts platform.

Support Center

How can we help?