+ All Categories
Home > Documents > Eigenvalue Problems - University of Utah

Eigenvalue Problems - University of Utah

Date post: 18-Jan-2022
Category:
Upload: others
View: 1 times
Download: 0 times
Share this document with a friend
28
Eigenvalue Problems
Transcript
Page 1: Eigenvalue Problems - University of Utah

Eigenvalue

Problems

Page 2: Eigenvalue Problems - University of Utah

Last Time …

Social Network Graphs

Betweenness

Girvan-Newman Algorithm

Graph Laplacian

Spectral Bisection

πœ†2, 𝑀2

Page 3: Eigenvalue Problems - University of Utah

Today …

Small deviation into eigenvalue problems …

Page 4: Eigenvalue Problems - University of Utah

Formulation

Standard eigenvalue problem: Given a 𝑛 Γ— 𝑛 matrix 𝐴, find scalar πœ†and a nonzero vector π‘₯ such that

𝐴π‘₯ = πœ†π‘₯

πœ† is a eigenvalue, and π‘₯ is the corresponding eigenvector

Spectrum = πœ†(𝐴) = set of eigenvalues of 𝐴

Spectral radius = 𝜌 𝐴 = max πœ† ∢ πœ† ∈ πœ† 𝐴

Page 5: Eigenvalue Problems - University of Utah

Characteristic Polynomial

Equation 𝐴π‘₯ = πœ†π‘₯ is equivalent to

𝐴 βˆ’ πœ†πΌ π‘₯ = 0

Eigenvalues of 𝐴 are roots of the characteristic polynomial

det 𝐴 βˆ’ πœ†πΌ = 0

The characteristic polynomial is a powerful theoretical tool but

usually not useful computationally

Page 6: Eigenvalue Problems - University of Utah

Considerations

Properties of eigenvalue problem affecting choice of algorithm

Are all eigenvalues needed, or only a few?

Are only eigenvalues needed, or are corresponding eigenvectors also needed?

Is matrix real or complex?

Is matrix relatively small and dense, or large and sparse?

Does matrix have any special properties, such as symmetry?

Page 7: Eigenvalue Problems - University of Utah

Problem Transformations

Shift: If 𝐴π‘₯ = πœ†π‘₯ and 𝜎 is any scalar, then 𝐴 βˆ’ 𝜎𝐼 π‘₯ = πœ† βˆ’ 𝜎 π‘₯

Inversion: If 𝐴 is nonsingular and 𝐴π‘₯ = πœ†π‘₯, then πœ† β‰  0 and π΄βˆ’1π‘₯ =1

πœ†π‘₯

Powers: If 𝐴π‘₯ = πœ†π‘₯, then π΄π‘˜π‘₯ = πœ†π‘˜π‘₯

Polynomial: If 𝐴π‘₯ = πœ†π‘₯ and 𝑝(𝑑) is a polynomial, then 𝑝 𝐴 π‘₯ = 𝑝 πœ† π‘₯

Page 8: Eigenvalue Problems - University of Utah

Similarity Transforms

𝐡 is similar to 𝐴 if there is a nonsingular matrix 𝑇, such that

𝐡 = π‘‡βˆ’1𝐴𝑇

Then,

𝐡𝑦 = πœ†π‘¦ β‡’ π‘‡βˆ’1𝐴𝑇𝑦 = πœ†π‘¦ β‡’ 𝐴 𝑇𝑦 = πœ† 𝑇𝑦

Similarity transformations preserve eigenvalues and eigenvectors are easily recovered

Page 9: Eigenvalue Problems - University of Utah

Diagonal form

Eigenvalues of a diagonal matrix are the diagonal entries and the

eigenvectors are columns of the Identity matrix

The diagonal form is highly desirable in simplifying eigenvalue

problems for general matrices by similarity transformations

But not all matrices are diagonalizable by similarity transformations

Page 10: Eigenvalue Problems - University of Utah

Triangular form

Any matrix can be transformed into a triangular form by similarity

The eigenvalues are simply the diagonal values

Eigenvectors are not as obvious, but still easy to compute

Page 11: Eigenvalue Problems - University of Utah

Power iteration

Simplest method for computing one eigenvalue-eigenvector pair

π‘₯π‘˜ = 𝐴π‘₯π‘˜βˆ’1

Converges to multiple of eigenvector corresponding to dominant

eigenvalue

We have seen this before while computing the Page Rank

Proof of convergence ?

Page 12: Eigenvalue Problems - University of Utah

Convergence of Power iteration

Express starting vector π‘₯0 in terms of the eigenvectors of 𝐴

π‘₯0 =

𝑖=1

𝑛

𝛼𝑖 𝒗𝑖

Then,

π‘₯π‘˜ = 𝐴π‘₯π‘˜βˆ’1 = 𝐴2π‘₯π‘˜βˆ’2 = β‹― = 𝐴

π‘˜π‘₯0

𝑖=1

𝑛

πœ†π‘–π‘˜ 𝛼𝑖𝒗𝑖 = πœ†π‘›

π‘˜ 𝛼𝑛𝒗𝑛 +

𝑖=1

π‘›βˆ’1πœ†π‘–πœ†π‘›

π‘˜

𝛼𝑖𝒗𝑖

Since πœ†π‘–

πœ†π‘›< 1, successively higher powers go to zero

Page 13: Eigenvalue Problems - University of Utah

Power iteration with shift

Convergence rate of power iteration depends on the ratio πœ†π‘›βˆ’1

πœ†π‘›

It is possible to choose a shift, 𝐴 βˆ’ 𝜎𝐼, such that

πœ†π‘›βˆ’1 βˆ’ 𝜎

πœ†π‘› βˆ’ 𝜎<πœ†π‘›βˆ’1πœ†π‘›

so convergence is accelerated

Shift must be added back to result to obtain eigenvalue of original

matrix

Page 14: Eigenvalue Problems - University of Utah

Inverse iteration

If the smallest eigenvalues are required rather than the largest, we

can make use of the fact that the eigenvalues of π΄βˆ’1 are reciprocals of those of 𝐴, so the smallest eigenvalue of 𝐴 is the

reciprocal of the largest eigenvalue of π΄βˆ’1

This leads to the inverse iteration scheme

π΄π‘¦π‘˜ = π‘₯π‘˜βˆ’1π‘₯π‘˜ = π‘¦π‘˜/ π‘¦π‘˜ ∞

Inverse of 𝐴 is not computed explicitly, but some factorization of 𝐴 is

used to solve the system at each iteration

Page 15: Eigenvalue Problems - University of Utah

Shifted inverse iteration

As before, the shifting strategy using a scalar 𝜎 can greatly improve

convergence

It is particularly useful for computing the eigenvector corresponding

to the approximate eigenvalue

Inverse iteration is also useful for computing the eigenvalue closest

to a given value 𝛽, since if 𝛽 is used as the shift, then the desired

eigenvalue corresponds to the smallest eigenvalue of the shifted

matrix

Page 16: Eigenvalue Problems - University of Utah

Deflation

Once the dominant eigenvalue and eigenvector πœ†π‘›, 𝑀𝑛 have

been computed, the remaining eigenvalues can be computed using deflation, which effectively removes the known eigenvalue

Let 𝐻 be any nonsingular matrix such that 𝐻π‘₯ = 𝛼𝑒1

Then the similarity transform determined by 𝐻 transforms 𝐴 into,

π»π΄π»βˆ’1 = πœ†π‘› 𝑏𝑇

0 𝐡

Can now work with 𝐡 to compute the next eigenvalue,

Process can be repeated to find additional eigenvalues and eigenvectors

Page 17: Eigenvalue Problems - University of Utah

Deflation

Alternate approach: let 𝑒𝑛 be any vector such that 𝑒𝑛𝑇𝑀𝑛 = πœ†π‘›

Then 𝐴 βˆ’ 𝑀𝑛𝑒𝑛𝑇 has eigenvalues πœ†π‘›βˆ’1, … , πœ†1, 0

Possible choices for 𝑒𝑛

𝑒𝑛 = πœ†π‘›π‘€π‘›, if 𝐴 is symmetric and 𝑀𝑛 is normalized so that 𝑀𝑛 2 = 1

𝑒𝑛 = πœ†π‘›π‘¦π‘›, where 𝑦𝑛 is the corresponding left eigenvector (𝐴𝑇𝑦𝑛 = πœ†π‘›π‘¦π‘›)

𝑒𝑛 = π΄π‘‡π‘’π‘˜, if 𝑀𝑛 is normalized such that 𝑀𝑛 ∞ = 1 and the π‘˜π‘‘β„Ž

component of 𝑀𝑛 is 1

Page 18: Eigenvalue Problems - University of Utah

QR Iteration

Iteratively converges to a triangular or block-triangular form, yielding

all eigenvalues of 𝐴

Starting with 𝐴0 = 𝐴, at iteration π‘˜ compute QR factorization,

π‘„π‘˜π‘…π‘˜ = π΄π‘˜βˆ’1

And form the reverse product,

π΄π‘˜ = π‘…π‘˜π‘„π‘˜

Product of orthogonal matrices π‘„π‘˜ converges to matrix of

corresponding eigenvectors

If 𝐴 is symmetric, then symmetry is preserved by the QR iteration, so

π΄π‘˜ converges to a matrix that is both symmetric and triangular

diagonal

Page 19: Eigenvalue Problems - University of Utah

Preliminary reductions

Efficiency of QR iteration can be enhanced by first transforming the

matrix to be as close to triangular form as possible

Hessenberg matrix is triangular except for one additional nonzero

diagonal immediately adjacent to the main diagonal

Symmetric Hessenberg matrix is tridiagonal

Any matrix can be reduced to Hessenberg form in finite number of

steps using Householder transformations

Work per iteration is reduced from π’ͺ 𝑛3 to π’ͺ(𝑛2) for general

matrices and π’ͺ(𝑛) for symmetric matrices

Page 20: Eigenvalue Problems - University of Utah

Krylov subspace methods

Reduces matrix to Hessenberg or tridiagonal form using only matrix-vector products

For arbitrary starting vector π‘₯0, ifπΎπ‘˜ = π‘₯0 𝐴π‘₯0 β‹― 𝐴

π‘˜βˆ’1π‘₯0

thenπΎπ‘›βˆ’1𝐴𝐾𝑛 = 𝐢𝑛

where 𝐢𝑛 is upper Hessenberg

To obtain a better conditioned basis for span(𝐾𝑛), compute the QR factorization,

𝑄𝑛𝑅𝑛 = 𝐾𝑛

so that𝑄𝑛𝐻𝐴𝑄𝑛 = 𝑅𝑛𝐢𝑛𝑅𝑛

βˆ’1 ≑ 𝐻

with 𝐻 is upper Hessenberg

Page 21: Eigenvalue Problems - University of Utah

Krylov subspace methods

Equating π‘˜π‘‘β„Ž columns on each side of the equation 𝐴𝑄𝑛 = 𝑄𝑛𝐻 yields

π΄π‘žπ‘˜ = β„Ž1π‘˜π‘ž1 +β‹―+ β„Žπ‘˜π‘˜π‘žπ‘˜ + β„Žπ‘˜+1,π‘˜π‘žπ‘˜+1

relating π‘žπ‘˜+1 to preceding vectors π‘ž1, … , π‘žπ‘˜

Premultiplying by π‘žπ‘—π» and using orthonormality

β„Žπ‘—π‘˜ = π‘žπ‘—π»π΄π‘žπ‘˜ , 𝑗 = 1,… , π‘˜

These relationships yield the Arnoldi iteration

Page 22: Eigenvalue Problems - University of Utah

Arnoldi iteration

π‘₯0 : arbitrary nonzero starting vector

π‘ž1 = π‘₯0/ π‘₯0 2

for π‘˜ = 1,2, …

π‘’π‘˜ = π΄π‘žπ‘˜

for 𝑗 = 1 to π‘˜

β„Žπ‘—π‘˜ = π‘žπ‘—π»π‘’π‘˜

π‘’π‘˜ = π‘’π‘˜ βˆ’ β„Žπ‘—π‘˜π‘žπ‘—

β„Žπ‘˜+1,π‘˜ = π‘’π‘˜ 2

if β„Žπ‘˜+1,π‘˜ = 0 then stop

π‘žπ‘˜+1 = π‘’π‘˜/β„Žπ‘˜+1,π‘˜

Page 23: Eigenvalue Problems - University of Utah

Arnoldi iteration

Ifπ‘„π‘˜ = π‘ž1 β‹― π‘žπ‘˜

thenπ»π‘˜ = π‘„π‘˜

π»π΄π‘„π‘˜

is a upper Hessenberg matrix

Eigenvalues of π»π‘˜, called Ritz values, are approximate eigenvalues of 𝐴, and Ritz vectors given by π‘„π‘˜π‘¦, where 𝑦 is an eigenvector of π»π‘˜, are corresponding approximate eigenvectors of 𝐴

Eigenvalues of π»π‘˜ must be computed by another method, such as QR iteration, but this is an easier problem as π‘˜ β‰ͺ 𝑛

Page 24: Eigenvalue Problems - University of Utah

Arnoldi iteration

Is fairly expensive in both work and storage because each new

vector π‘žπ‘˜ must be orthogonalized against all previous columns of π‘„π‘˜, and all must be stored for that purpose

Is usually restarted periodically with a carefully chosen starting

vector

Ritz vectors and values produced are often good approximations to

eigenvalues and eigenvectors of 𝐴 after relatively few iterations

Page 25: Eigenvalue Problems - University of Utah

Lanczos iteration

Work and storage costs drop dramatically if the matrix is symmetric or Hermitian, since the recurrence has only three terms and π»π‘˜ is tridiagonal

π‘ž0, 𝛽0 = 0 and π‘₯0 = arbitrary nonzero starting vector

π‘ž1 = π‘₯0/ π‘₯0 2

for π‘˜ = 1,2, …

π‘’π‘˜ = π΄π‘žπ‘˜

π›Όπ‘˜ = π‘žπ‘˜π»π‘’π‘˜

π‘’π‘˜ = π‘’π‘˜ βˆ’ π›½π‘˜βˆ’1π‘žπ‘˜βˆ’1 βˆ’ π›Όπ‘˜π‘žπ‘˜

π›½π‘˜ = π‘’π‘˜ 2

if π›½π‘˜ = 0 then stop

π‘žπ‘˜+1 = π‘’π‘˜/π›½π‘˜

Page 26: Eigenvalue Problems - University of Utah

Lanczos iteration

π›Όπ‘˜ and π›½π‘˜ are diagonal and subdiagonal entries of symmetric

tridiagonal matrix π‘‡π‘˜

As with Arnoldi, Lanczos does not produce eigenvalues and

eigenvectors directly, but only the tridiagonal matrix π‘‡π‘˜, whose

eigenvalues and eigenvectors must be computed by another

method to obtain Ritz values and vectors

If 𝛽 = 0, then the invariant subspace has already been identified,

i.e., the Ritz values and vectors are already exact at that point

Page 27: Eigenvalue Problems - University of Utah

Lanczos iteration

In principle, if we let Lanczos run until π‘˜ = 𝑛, the resulting tridiagonal

matrix would be orthogonally similar to 𝐴

In practice, rounding errors cause loss of orthogonality

Problem can be overcome by reorthogonalizing vectors

In practice, this is usually ignored. The resulting approximations are

still good

Page 28: Eigenvalue Problems - University of Utah

Krylov subspace methods

Great advantage of Arnoldi and Lanczos is their ability to produce

good approximations to extreme eigenvalues for π‘˜ β‰ͺ 𝑛

They only require one matrix-vector product per step and little

auxiliary storage, so are ideally suited for large sparse matrices

If eigenvalues are needed in the middle of the spectrum, say near

𝜎, then the algorithms can be applied to 𝐴 βˆ’ 𝜎𝐼 βˆ’1, assuming it is

practical to solve systems of the form 𝐴 βˆ’ 𝜎𝐼 π‘₯ = 𝑦


Recommended