ECS130 Scientific Computing Handout E February 13, 2017

Similar documents
ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

Orthogonal iteration to QR

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Numerical Methods I Eigenvalue Problems

Orthogonal iteration revisited

Numerical methods for eigenvalue problems

Orthogonal iteration to QR

Eigenvalue Problems and Singular Value Decomposition

Notes on Eigenvalues, Singular Values and QR

Direct methods for symmetric eigenvalue problems

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem

Lecture 4 Eigenvalue problems

Solving large scale eigenvalue problems

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

Solving large scale eigenvalue problems

Numerical Methods I: Eigenvalues and eigenvectors

Matrices, Moments and Quadrature, cont d

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

EIGENVALUE PROBLEMS (EVP)

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Applied Linear Algebra in Geoscience Using MATLAB

The Eigenvalue Problem: Perturbation Theory

632 CHAP. 11 EIGENVALUES AND EIGENVECTORS. QR Method

Math 504 (Fall 2011) 1. (*) Consider the matrices

Course Notes: Week 1

The Singular Value Decomposition and Least Squares Problems

I. Multiple Choice Questions (Answer any eight)

Computational Methods. Eigenvalues and Singular Values

THE QR METHOD A = Q 1 R 1

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

be a Householder matrix. Then prove the followings H = I 2 uut Hu = (I 2 uu u T u )u = u 2 uut u

Conceptual Questions for Review

LARGE SPARSE EIGENVALUE PROBLEMS

Section 6.4. The Gram Schmidt Process

Eigenvalue problems III: Advanced Numerical Methods

AMS526: Numerical Analysis I (Numerical Linear Algebra)

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

Lecture 10 - Eigenvalues problem

MAA507, Power method, QR-method and sparse matrix representation.

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

A Review of Linear Algebra

Eigenvalues and eigenvectors

Iterative methods for symmetric eigenvalue problems

SUMMARY OF MATH 1600

Chapter 5. Eigenvalues and Eigenvectors

Applied Linear Algebra in Geoscience Using MATLAB

Algebra C Numerical Linear Algebra Sample Exam Problems

Accelerating computation of eigenvectors in the dense nonsymmetric eigenvalue problem

Eigenvalues and Eigenvectors

Orthonormal Transformations and Least Squares

EECS 275 Matrix Computation

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Krylov Subspaces. The order-n Krylov subspace of A generated by x is

Foundations of Matrix Analysis

Maths for Signals and Systems Linear Algebra in Engineering

From Matrix to Tensor. Charles F. Van Loan

Accelerating computation of eigenvectors in the nonsymmetric eigenvalue problem

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

Lecture 4 Basic Iterative Methods I

APPLIED NUMERICAL LINEAR ALGEBRA

6.4 Krylov Subspaces and Conjugate Gradients

Index. for generalized eigenvalue problem, butterfly form, 211

Numerical Solution of Linear Eigenvalue Problems

Lecture 4 Orthonormal vectors and QR factorization

The QR Algorithm. Marco Latini. February 26, 2004

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

A hybrid reordered Arnoldi method to accelerate PageRank computations

Solving linear equations with Gaussian Elimination (I)

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

Orthogonal Transformations

Worksheet for Lecture 25 Section 6.4 Gram-Schmidt Process

Eigenvalue problems. Eigenvalue problems

ANSWERS. E k E 2 E 1 A = B

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Eigenvalue and Eigenvector Problems

Problem # Max points possible Actual score Total 120

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Linear Systems. Class 27. c 2008 Ron Buckmire. TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5.4

Section 4.1 The Power Method

18.06SC Final Exam Solutions

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

Matrix Computations and Semiseparable Matrices

Eigenproblems II: Computation

Calculating determinants for larger matrices

The QR Decomposition

Krylov subspace projection methods

Transcription:

ECS130 Scientific Computing Handout E February 13, 2017 1. The Power Method (a) Pseudocode: Power Iteration Given an initial vector u 0, t i+1 = Au i u i+1 = t i+1 / t i+1 2 (approximate eigenvector) θ i+1 = u H i+1 Au i+1 (approximate eigenvalue) (b) Practical stopping criterion: θ i+1 θ i tol θ i. (c) Example: Let A = and λ(a) = {10,4,3}. Let u 0 = (1,0,0) T, then 261 209 49 530 422 98 800 631 144. i 1 2 3 10 θ i 994.49 13.0606 10.07191 10.0002 (d) Convergence analysis: Assume A = XΛX 1 with Λ = diag(λ 1,λ 2,...,λ n ) and λ 1 > λ 2... λ n. Then, we can show that u i = Ai u 0 A i u 0 x 1/ x 1, where x 1 = Xe 1 as i. θ i λ 1 as i. (e) The convergence rate depends on λ 2 λ 1. Therefore, if λ 2 λ 1 is close to 1, then the power method could be very slow convergent or doesn t converge at all. 1

2. Inverse Iteration (a) Purposes: Overcome the drawbacks of the power method (slow convergence) find an eigenvalue closest to a particular given number (called shift): σ (b) Observation: if λ is an eigenvalue of A, then λ σ is an eigenvalue of A σi, 1 λ σ is an eigenvalue of (A σi) 1. 1/(\lambda-\sigma) \sigma (c) Pseudocode Inverse Iteration Given an initial vector u 0 and a shift σ solve (A σi)t i+1 = u i for t i+1 u i+1 = t i+1 / t i+1 2 (approximate eigenvector) θ i+1 = u H i+1 Au i+1 (approximate eigenvalue) (d) Convergence analysis: Assume A = XΛX 1 with Λ = diag(λ 1,λ 2,...,λ n ) and λ k is the eigenvalue cloest to the shift σ. It can be shown that u i x k / x k as i, where x k = Xe k θ i converges to λ k i. Convergence rate depends on max j k λ k σ λ j σ. (e) Advantages: the ability to converge to any desired eigenvalue(the one nearest to the shift σ). By choosing σ very close to a desired eigenvalue, the method converges very quickly and thus not be as limited by the proximity of nearby eigenvalues as is the original power method. The method is particularly effective when we have a good approximation to an eigenvalue and want only its corresponding eigenvector. (f) Drawbacks: (a) expensive in general: solving (A σi)t i+1 = u i for u i+1. One LU factorization of A σi is required, which could be very expensive for large matrices, (b) Only compute one eigenpair. 2

3. Orthogonal iteration (subspace iteration, simultaneous iteration) (a) Purpose: compute a p-dimensional invaraint subspace, p > 1, rather than one eigenvector at a time. (b) Pseudocode: Orthogonal Iteration Given an initial n p orthogonal matrix Z 0 Y i+1 = AZ i Y i+1 = Z i+1 R i+1 (QR decomposition) The use of QR decomposition keeps the vectors spanning span{a i Z 0 } of full rank. (c) Convergence: under mild conditions, Z i converges to the invariant subspace spanned by the first p eigenvectors corresponding to the p dominant eigenvalues, where If we let B i = Z T i AZ i, then λ 1 λ 2 λ p > λ p+1 λ n. AZ i Z i B i 0 as i and eigenvalues of B i approximate the dominant eigenvalues of A. Convergence rate depends on λ p+1 / λ p. (d) Example: Let Z 0 = [e 1,e 2,e 3 ] and A = -0.4326 1.1892-0.5883-0.0956-0.6918-0.3999-1.6656-0.0376 2.1832-0.8323 0.8580 0.6900 0.1253 0.3273-0.1364 0.2944 1.2540 0.8156 0.2877 0.1746 0.1139-1.3362-1.5937 0.7119-1.1465-0.1867 1.0668 0.7143-1.4410 1.2902 1.1909 0.7258 0.0593 1.6236 0.5711 0.6686 Eigenvalues of A = -2.1659 +- 0.5560i, 2.1493, 0.2111 +- 1.9014i, -0.9548 i = 10: Eigenvalues of Z _10*A*Z_10: -1.4383 +- 0.3479i, 2.1500 i = 30: Eigenvalues of Z _30*A*Z_30: -2.1592 +- 0.5494i, 2.1118 i = 70: Eigenvalues of Z _70*A*Z_70: -2.1659 +- 0.5560i, 2.1493 (e) An important special case: Let p = n and Z 0 = I, then A i = Z T i AZ i converges to the Schur form of A provided that (1) all eigenvalues of A have distinct absolute values and (2) all the principal submatrices of S have full rank, where we assume A = SΛS 1. (f) Example: the same test matrix, numerical results of orthogonal iteration with Z 0 = I: 3

A_10 = -1.6994-0.2201-0.8787 1.4292 0.3847 0.0112 0.0007 1.1325-1.2186 1.2245-0.0867-0.0648 0.2637-1.9636-0.1598 2.3959-0.8136-0.4311-0.0364-0.2346 0.5527-0.4393-1.9263-1.2496-0.4290 1.3482 1.1484 0.6121-0.5937-0.2416 0.0003-0.0013-0.0003 0.0011-0.0014-0.9554 A_30 = -2.4055-1.0586-1.3420 0.0991-1.1210 0.1720 0.0517 0.9645-1.6519 0.8512 0.7215 0.7654 0.2248-1.9947-0.7656-1.1876-0.2736 0.1552 0.0029 0.0263-0.0682 0.1381-2.3094-0.6765 0.0147-0.0808-0.0569 1.5462 0.3082 0.8476 0.0000 0.0000 0.0000 0.0000 0.0000-0.9548 A_70 = -2.0800 1.6092-0.6426 1.1025-0.1553 0.0734-0.1690-1.6820 1.7425-0.0311 0.9229-0.3087 0.0677 1.2521 1.5795-0.1458 1.2092 0.7088 0.0000 0.0000-0.0005 0.5575-1.7386-1.0520 0.0000 0.0004 0.0003 2.1489-0.1353-0.3250 0.0000 0.0000 0.0000 0.0000 0.0000-0.9548... As we see A k is converging to a quasi-upper triangular matrix as k increasing. 4

4. QR iteration (a) Our goal is to reorganize orthogonal iteration to incorporate shifting and inverting as in the inverse iteration. This will make it more efficient and eliminate the assumption that eigenvalues differ in magnitude. (b) Pseudocode QR Iteration, without shift A 0 = A A i = Q i R i (QR decomposition) A i+1 = R i Q i (c) Properties Observe that A i+1 = R i Q i = Q T i Q i R i Q i = Q T i A i Q i A i+1 isorthogonallysimilartoa 0 = A. ThereforeA i+1 andahavesameeigenvalues: A i+1 = (Q 0 Q 1 Q i 1 Q i ) T A(Q 0 Q 1 Q i 1 Q i ). Note that Q 0 Q i 1 Q i is an orthogonal matrix since all Q j are. A i computed by QR iteration is identical to the matrix Z T i AZ i implicitly computed by orthogonal iteration. In fact, we have Proposition. Let A i be the matrix computed by QR iteration. Then A i = Z T i AZ i, where Z i is the matrix computed from orthogonal iteration starting with Z 0 = I. Therefore, A i converges to Schur form if all the eigenvalues have different absolute values. (d) Example. The same test matrix, numerical results of QR iteration A_10 = -1.6994 0.2201-0.8787-1.4292-0.3847 0.0112-0.0007 1.1325 1.2186 1.2245-0.0867 0.0648 0.2637 1.9636-0.1598-2.3959 0.8136-0.4311 0.0364-0.2346-0.5527-0.4393-1.9263 1.2496 0.4290 1.3482-1.1484 0.6121-0.5937 0.2416 0.0003 0.0013-0.0003-0.0011 0.0014-0.9554 A_30 = -2.4055-1.0586 1.3420-0.0991 1.1210 0.1720 0.0517 0.9645 1.6519-0.8512-0.7215 0.7654-0.2248 1.9947-0.7656-1.1876-0.2736-0.1552-0.0029-0.0263-0.0682 0.1381-2.3094 0.6765-0.0147 0.0808-0.0569 1.5462 0.3082-0.8476 0.0000 0.0000 0.0000 0.0000 0.0000-0.9548 5

A_70 = -2.0800 1.6092-0.6426 1.1025-0.1553-0.0734-0.1690-1.6820 1.7425-0.0311 0.9229 0.3087 0.0677 1.2521 1.5795-0.1458 1.2092-0.7088 0.0000 0.0000-0.0005 0.5575-1.7386 1.0520 0.0000 0.0004 0.0003 2.1489-0.1353 0.3250 0.0000 0.0000 0.0000 0.0000 0.0000-0.9548 Note that the results are identical to the orthogonal iteration. 6

5. QR iteration with shifts QR Algorithm (a) Purpose: accelerate the convergence of QR iteration by using shifts (b) Pseudo-code QR Iteration with Shifts A 0 = A Choose a shift σ i A i σ i I = Q i R i (QR decomposition) A i+1 = R i Q i +σ i I (c) Property: A i and A i+1 are orthogonally similar: A i+1 = Q T i A i Q i. Therefore, A i+1 and A are orthogonally similar, and A i+1 and A have the same eigenvalues. (d) How to choose the shifts σ i? If σ i is an exact eigenvalue of A, then it can be shown that [ ] A a A i+1 = R i Q i +σ i I =. 0 σ i This means that the algorithm converges in one iteration. If more eigenvalues are wanted, we can apply the algorithm again to the n 1 by n 1 matrix A. In practice, a common choice of the σ i is σ i = A i (n,n). A motivation of this choice is by observing that the convergence of the QR iteration (without a shift), the (n,n) entry of A i usually converges to an eigenvalue of A first. (e) Example. The same test matrix as before. The following is the numerical result of QR iteration with a shift. With the shift σ 0 as an exact eigenvalue σ 0 = 2.1493, then A_1 = -1.4127 1.4420 1.0845-0.6866-0.1013-0.2042-1.2949-0.2334 1.4047-1.3695 1.5274-0.7062 0.5473 0.1343-0.7991-0.6716 1.1585 0.0736-0.2630 0.0284 0.5440-1.4616-1.5892 0.9205-1.6063-0.3898 0.3410 0.1623-0.9576-0.5795 0.0000 0.0000 0.0000 0.0000 0.0000 2.1493 7

With the shifts σ i = A i (n,n). A_7 = -2.4302 2.0264-0.2799-0.2384 0.3210-0.0526-0.1865-1.4295-1.3515 0.0812 0.8577-0.0388-0.1087-0.8991 0.4491 0.4890-1.8463-1.2034-0.0008 0.0511-0.5997-0.7839-0.8088-0.5188-0.0916-0.8273 1.6940 0.0645-0.6698-0.0854 0.0000 0.0000 0.0000 0.0000 0.0000 2.1493 We observe that by 7th iteration, we have found an eigenvalue of A. (f) Note that the QR decomposition in the algorithm takes O(n 3 ) flops. Even if the algorithm took n iterations to converge, the overall cost of the algorithm will be O(n 4 ). This is too expensive (today, the complexity of algorithms for all standard matrix computation problems is at O(n 3 ).) However, if the matrix is initially reduced to upper Hessenberg form, then the QR decomposition of a Hessenberg form costs O(n 2 ) flops. As a result, the overall cost of the algorithm is reduced to O(n 3 ). This is referred to as the Hessenberg QR algorithm, the method of choice for dense eigenvalue problem today, say Matlab s eigensolver eig use LAPACK s implementation of the QR algorithm. (g) QR algorithm is one of the top 10 algorithms in the 20th century. 8