Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Similar documents
Today: eigenvalue sensitivity, eigenvalue algorithms Reminder: midterm starts today

Eigenvalue and Eigenvector Problems

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Lecture 10 - Eigenvalues problem

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

MATH 583A REVIEW SESSION #1

Linear Algebra - Part II

The Eigenvalue Problem: Perturbation Theory

The Singular Value Decomposition

Linear Algebra: Matrix Eigenvalue Problems

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

Quantum Computing Lecture 2. Review of Linear Algebra

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors

Numerical Methods I Eigenvalue Problems

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Lecture 13 Eigenvalue Problems

235 Final exam review questions

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

LinGloss. A glossary of linear algebra

Symmetric and anti symmetric matrices

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Eigenvalues and eigenvectors

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Numerical Linear Algebra Homework Assignment - Week 2

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Matrices and Linear Algebra

c c c c c c c c c c a 3x3 matrix C= has a determinant determined by

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

Math 3191 Applied Linear Algebra

Math Matrix Algebra

Introduction to Matrix Algebra

1 Linear Algebra Problems

SECTIONS 5.2/5.4 BASIC PROPERTIES OF EIGENVALUES AND EIGENVECTORS / SIMILARITY TRANSFORMATIONS

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

Chap 3. Linear Algebra

Linear Algebra Review. Vectors

Lecture 15, 16: Diagonalization

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

Review problems for MA 54, Fall 2004.

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Matrix Vector Products

Spectral Theorem for Self-adjoint Linear Operators

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Foundations of Matrix Analysis

1 Last time: least-squares problems

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

Linear algebra and applications to graphs Part 1

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

Eigenvalues and Eigenvectors

Review of some mathematical tools

1 Inner Product and Orthogonality

Eigenpairs and Diagonalizability Math 401, Spring 2010, Professor David Levermore

Up to this point, our main theoretical tools for finding eigenvalues without using det{a λi} = 0 have been the trace and determinant formulas

Calculating determinants for larger matrices

Review of linear algebra

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

CS 143 Linear Algebra Review

Chapter 3. Determinants and Eigenvalues

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

Math 205, Summer I, Week 4b:

π 1 = tr(a), π n = ( 1) n det(a). In particular, when n = 2 one has

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Review of Linear Algebra

CS 246 Review of Linear Algebra 01/17/19

MATH 423 Linear Algebra II Lecture 20: Geometry of linear transformations. Eigenvalues and eigenvectors. Characteristic polynomial.

Notes on basis changes and matrix diagonalization

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

A = 3 B = A 1 1 matrix is the same as a number or scalar, 3 = [3].

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

MATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization.

Linear Algebra for Machine Learning. Sargur N. Srihari

Math 310 Final Exam Solutions

A VERY BRIEF LINEAR ALGEBRA REVIEW for MAP 5485 Introduction to Mathematical Biophysics Fall 2010

Notes on Linear Algebra

Draft. Lecture 14 Eigenvalue Problems. MATH 562 Numerical Analysis II. Songting Luo. Department of Mathematics Iowa State University

Background Mathematics (2/2) 1. David Barber

The Spectral Theorem for normal linear maps

Matrix Algebra: Summary

Matrix Theory, Math6304 Lecture Notes from September 27, 2012 taken by Tasadduk Chowdhury

MATH 221, Spring Homework 10 Solutions

Announcements Monday, November 06

Eigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems.

Computational math: Assignment 1

Maths for Signals and Systems Linear Algebra in Engineering

Definitions for Quizzes

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

Math Homework 8 (selected problems)

Linear Algebra. Workbook

Transcription:

Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic

Unit V: Eigenvalue Problems Chapter V.2: Fundamentals 2 / 31

Eigenvalues and Eigenvectors Eigenvalues and eigenvectors of real-valued matrices can be complex Hence in this Unit we will generally work with complex-valued matrices and vectors, A C n n, v C n For A C n n, we shall consider the eigenvalue problem: find (λ, v) C C n such that Av = λv, v 2 = 1 Note that for v C n, v 2 ( n k=1 v k 2) 1/2, where is the modulus of a complex number 3 / 31

Eigenvalues and Eigenvectors This problem can be reformulated as: (A λi)v = 0 We know this system has a solution if and only if (A λi) is singular, hence we must have det(a λi) = 0 p(z) det(a zi) is a degree n polynomial, called the characteristic polynomial of A The eigenvalues of A are exactly the roots of the characteristic polynomial 4 / 31

Characteristic Polynomial By the fundamental theorem of algebra, we can factorize p(z) as p(z) = c n (z λ 1 )(z λ 2 ) (z λ n ), where the roots λ i C need not be distinct Note also that complex eigenvalues of a matrix A R n n must occur as complex conjugate pairs That is, if λ = α + iβ is an eigenvalue, then so is its complex conjugate λ = α iβ 5 / 31

Characteristic Polynomial This follows from the fact that for a polynomial p with real coefficients, p(z) = p(z) for any z C: p(z) = n c k (z) k = k=0 n c k z k = k=0 n c k z k = p(z) k=0 Hence if w C is a root of p, then so is w, since 0 = p(w) = p(w) = p(w) 6 / 31

Companion Matrix We have seen that every matrix has an associated characteristic polynomial Similarly, every polynomial has an associated companion matrix The companion matrix, C n, of p P n is a matrix which has eigenvalues that match the roots of p Divide p by its leading coefficient to get a monic polynomial, i.e. with leading coefficient equal to 1 (this doesn t change the roots) p monic (z) = c 0 + c 1 z + + c n 1 z n 1 + z n 7 / 31

Companion Matrix Then p monic is the characteristic polynomial of the n n companion matrix 0 0 0 c 0 1 0 0 c 1 C n = 0 1 0 c 2....... 0 0 1 c n 1 8 / 31

Companion Matrix We show this for the n = 3 case: Consider which has companion matrix p 3,monic (z) c 0 + c 1 z + c 2 z 2 + z 3, C 3 0 0 c 0 1 0 c 1 0 1 c 2 Recall that for a 3 3 matrix, we have a 11 a 12 a 13 det a 21 a 22 a 23 = a 11 a 22 a 33 + a 12 a 23 a 31 + a 13 a 21 a 32 a 31 a 32 a 33 a 13 a 22 a 31 a 11 a 23 a 32 a 12 a 21 a 33 9 / 31

Companion Matrix Substituting entries of C 3 then gives det(zi C 3 ) = c 0 + c 1 z + c 2 z 2 + z 3 = p 3,monic (z) This link between matrices and polynomials is used by Matlab s roots function roots computes all roots of a polynomial by using algorithms considered in this Unit to find eigenvalues of the companion matrix 10 / 31

Eigenvalue Decomposition Let λ be an eigenvalue of A C n n ; the set of all eigenvalues is called the spectrum of A The algebraic multiplicity of λ is the multiplicity of the corresponding root of the characteristic polynomial The geometric multiplicity of λ is the number of linearly independent eigenvectors corresponding to λ For example, for A = I, λ = 1 is an eigenvalue with algebraic and geometric multiplicity of n (Char. poly. for A = I is p(z) = (z 1) n, and e i C n, i = 1, 2,..., n are eigenvectors) 11 / 31

Eigenvalue Decomposition Theorem: The geometric multiplicity of an eigenvalue is less than or equal to its algebraic multiplicity If λ has geometric multiplicity < algebraic multiplicity, then λ is said to be defective We say a matrix is defective if it has at least one defective eigenvalue 12 / 31

Eigenvalue Decomposition For example, A = 2 1 0 0 2 1 0 0 2 has one eigenvalue with algebraic multiplicity of 3, geometric multiplicity of 1 >> A = [2 1 0; 0 2 1; 0 0 2]; >> [V,D] = eig(a) V = 1.000000000000000-1.000000000000000 1.000000000000000 0 0.000000000000000-0.000000000000000 0 0 0.000000000000000 D = 2 0 0 0 2 0 0 0 2 13 / 31

Eigenvalue Decomposition Let A C n n be a nondefective matrix, then it has a full set of n linearly independent eigenvectors v 1, v 2,..., v n C n Suppose V C n n contains the eigenvectors of A as columns, and let D = diag(λ 1,..., λ n ) Then Av i = λ i v i, i = 1, 2,..., n is equivalent to AV = VD Since we assumed A is nondefective, we can invert V to obtain A = VDV 1 This is the eigendecomposition of A This shows that for a non-defective matrix, A is diagonalized by V 14 / 31

Eigenvalue Decomposition We introduce the conjugate transpose, A C n m, of a matrix A C m n (A ) ij = A ji, i = 1, 2,..., m, j = 1, 2,..., n A matrix is said to be hermitian if A = A (this generalizes matrix symmetry) A matrix is said to be unitary if AA = I (this generalizes the concept of an orthogonal matrix) Also, for v C n, v 2 = v v 15 / 31

Eigenvalue Decomposition The operator in Matlab actually performs conjugate transpose: >> [1+i i; -2*i 3] ans = 1.0000-1.0000i 0 + 2.0000i 0-1.0000i 3.0000 To get the raw transpose, do. : >> [1 i; -2*i 3]. 1.0000 + 1.0000i 0-2.0000i 0 + 1.0000i 3.0000 16 / 31

Eigenvalue Decomposition In some cases, the eigenvectors of A can be chosen such that they are orthonormal { vi 1, i = j v j = 0, i j In such a case, the matrix of eigenvectors, Q, is unitary, and hence A can be unitarily diagonalized A = QDQ 17 / 31

Eigenvalue Decomposition Theorem: A hermitian matrix is unitarily diagonalizable, and its eigenvalues are real But hermitian matrices are not the only matrices that can be unitarily diagonalized... A C n n is normal if A A = AA Theorem: A matrix is unitarily diagonalizable if and only if it is normal 18 / 31

Gershgorin s Theorem Due to the link between eigenvalues and polynomial roots, in general one has to use iterative methods to compute eigenvalues However, it is possible to gain some information about eigenvalue locations more easily from Gershgorin s Theorem Let D(c, r) {x C : x c r} denote a disk in the complex plane centered at c with radius r For a matrix A C n n, D(a ii, R i ) is called a Gershgorin disk, where R i = n a ij, j=1 j i 19 / 31

Gershgorin s Theorem Theorem: All eigenvalues of A C n n are contained within the union of the n Gershgorin disks of A Proof: See lecture 20 / 31

Gershgorin s Theorem Note that a matrix is diagonally dominant if a ii > n a ij, for i = 1, 2,..., n j=1 j i It follows from Gershgorin s Theorem that a diagonally dominant matrix cannot have a zero eigenvalue, hence must be invertible For example, the finite difference discretization matrix of the differential operator + I is diagonally dominant In 2-dimensions, ( + I)u = u xx u yy + u Each row of the corresponding discretization matrix contains diagonal entry 4/h + 1, and four off-diagonal entries of 1/h 21 / 31

Sensitivity of Eigenvalue Problems We shall now consider the sensitivity of the eigenvalues to perturbations in the matrix A Suppose A is nondefective, and hence A = VDV 1 Let δa denote a perturbation of A, and let E V 1 δav, then V 1 (A + δa)v = V 1 AV + V 1 δav = D + E 22 / 31

Sensitivity of Eigenvalue Problems For a nonsingular matrix X, the map A X 1 AX is called a similarity transformation of A Theorem: A similarity transformation preserves eigenvalues Proof: We can equate the characteristic polynomials of A and X 1 AX (denoted p A (z) and p X 1 AX (z), respectively) as follows: p X 1 AX (z) = det(zi X 1 AX ) = det(x 1 (zi A)X ) = det(x 1 )det(zi A)det(X ) = det(zi A) = p A (z), where we have used the identities det(ab) = det(a)det(b), and det(x 1 ) = 1/det(X ) 23 / 31

Sensitivity of Eigenvalue Problems The identity V 1 (A + δa)v = D + E is a similarity transformation Therefore A + δa and D + E have the same eigenvalues Let λ k, k = 1, 2,..., n denote the eigenvalues of A, and λ denote an eigenvalue of A + δa Then for some w C n, ( λ, w) is an eigenpair of (D + E), i.e. (D + E)w = λw 24 / 31

Sensitivity of Eigenvalue Problems This can be rewritten as w = ( λi D) 1 Ew This is a promising start because: we want to bound λ λ k for some k ( λi D) 1 is a diagonal matrix with entries 1/( λ λ k ) on the diagonal 25 / 31

Sensitivity of Eigenvalue Problems Taking norms yields w 2 ( λi D) 1 2 E 2 w 2, or ( λi D) 1 1 2 E 2 Note that the norm of a diagonal matrix is given by its largest entry (in abs. val.) 1 max v 0 Dv v (D 11 v 1, D 22 v 2,..., D nn v n ) = max v 0 v { } max D v ii max i=1,2,...,n v 0 v = max D ii i=1,2,...,n 1 This holds for any induced matrix norm, not just the 2-norm 26 / 31

Sensitivity of Eigenvalue Problems Hence ( λi D) 1 2 = 1/ λ λ k, where λ k of A closest to λ is the eigenvalue Therefore it follows from ( λi D) 1 1 2 E 2 that λ λ k = ( λi D) 1 1 2 E 2 = V 1 δav 2 V 1 2 δa 2 V 2 = cond(v ) δa 2 This result is known as the Bauer-Fike Theorem 27 / 31

Sensitivity of Eigenvalue Problems Hence suppose we compute the eigenvalues, λ i, of the perturbed matrix A + δa Then Bauer-Fike tells us that each λ i must reside in a disk of radius cond(v ) δa 2 centered on some eigenvalue of A If V is poorly conditioned, then even for small perturbations δa, the disks can be large: sensitivity to perturbations If A is normal then cond(v ) = 1, in which case the Bauer-Fike disk radius is just δa 2 28 / 31

Sensitivity of Eigenvalue Problems Matlab example: Eigenvalue perturbation and Bauer-Fike Theorem 3 2 imag(λ) 1 0 1 2 4 3 2 1 0 1 2 3 real(λ) 29 / 31

Sensitivity of Eigenvalue Problems Note that a limitation of Bauer-Fike is that it does not tell us which disk λ i will reside in Therefore, this doesn t rule out the possibility of, say, all λ i clustering in just one Bauer-Fike disk In the case that A and A + δa are hermitian, we have a stronger result 30 / 31

Sensitivity of Eigenvalue Problems Weyl s Theorem: Let λ 1 λ 2 λ n and λ 1 λ 2 λ n be the eigenvalues of hermitian matrices A and A + δa, respectively. Then max λ i λ i δa 2. i=1,...,n Hence in the hermitian case, each perturbed eigenvalue must be in the disk 2 of its corresponding unperturbed eigenvalue! 2 In fact, eigenvalues of a hermitian matrix are real, so disk here is actually an interval in R 31 / 31