Spectral Theorems in Euclidean and Hermitian Spaces

Similar documents
Spectral Theorems in Euclidean and Hermitian Spaces

4.2. ORTHOGONALITY 161

Basics of Hermitian Geometry

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Singular Value Decomposition (SVD) and Polar Form

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Linear algebra 2. Yoav Zemel. March 1, 2012

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

Linear Algebra Lecture Notes-II

Spectral Theorem for Self-adjoint Linear Operators

The Cartan Dieudonné Theorem

1 Last time: least-squares problems

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Linear Algebra 2 Spectral Notes

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

6 Inner Product Spaces

Diagonalizing Matrices

Chapter 5. Basics of Euclidean Geometry

Math Linear Algebra II. 1. Inner Products and Norms

Quantum Computing Lecture 2. Review of Linear Algebra

The Spectral Theorem for normal linear maps

Real symmetric matrices/1. 1 Eigenvalues and eigenvectors

NOTES ON BILINEAR FORMS

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Numerical Linear Algebra

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Linear Algebra. Min Yan

9.1 Eigenvectors and Eigenvalues of a Linear Map

LINEAR ALGEBRA REVIEW

Chapter 6 Inner product spaces

A PRIMER ON SESQUILINEAR FORMS

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

MATHEMATICS 217 NOTES

Review of Linear Algebra

Basics of Hermitian Geometry

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Linear Algebra: Matrix Eigenvalue Problems

Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

Chapter 6: Orthogonality

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Elementary linear algebra

MATH 583A REVIEW SESSION #1

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

f(a)f(b) = ab for all a, b E. WhenE = F, an affine isometry f: E E is also called a rigid motion.

Math 108b: Notes on the Spectral Theorem

1. General Vector Spaces

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Lecture 7: Positive Semidefinite Matrices

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

Linear Algebra using Dirac Notation: Pt. 2

(v, w) = arccos( < v, w >

Linear Algebra Review. Vectors

No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question.

Review problems for MA 54, Fall 2004.

Chapter 3 Transformations

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Last name: First name: Signature: Student number:

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

1 Inner Product and Orthogonality

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

SYLLABUS. 1 Linear maps and matrices

There are two things that are particularly nice about the first basis

October 25, 2013 INNER PRODUCT SPACES

Math 593: Problem Set 10

Eigenvalues and eigenvectors

Practice Exam. 2x 1 + 4x 2 + 2x 3 = 4 x 1 + 2x 2 + 3x 3 = 1 2x 1 + 3x 2 + 4x 3 = 5

Definitions for Quizzes

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

Categories and Quantum Informatics: Hilbert spaces

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam June 8, 2012

Algebra I Fall 2007

Homework 11 Solutions. Math 110, Fall 2013.

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

LinGloss. A glossary of linear algebra

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

1 Planar rotations. Math Abstract Linear Algebra Fall 2011, section E1 Orthogonal matrices and rotations

Recall the convention that, for us, all vectors are column vectors.

Lecture 3: Review of Linear Algebra

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Schur s Triangularization Theorem. Math 422

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

Analysis Preliminary Exam Workshop: Hilbert Spaces

Lecture 3: Review of Linear Algebra

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

Mathematical Methods wk 1: Vectors

Mathematical Methods wk 1: Vectors

Chapter 7: Symmetric Matrices and Quadratic Forms

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

GQE ALGEBRA PROBLEMS

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

Linear Algebra. Session 12

MATH 115A: SAMPLE FINAL SOLUTIONS

Transcription:

Chapter 9 Spectral Theorems in Euclidean and Hermitian Spaces 9.1 Normal Linear Maps Let E be a real Euclidean space (or a complex Hermitian space) with inner product u, v u, v. In the real Euclidean case, recall that, is bilinear, symmetric and positive definite (i.e., u, u > 0forall u = 0). In the complex Hermitian case, recall that, is sesquilinear, which means that it linear in the first argument, semilinear in the second argument (i.e., u, µv = µu, v), v, u = u, v, andpositivedefinite (as above). 361

362 CHAPTER 9. SPECTRAL THEOREMS In both cases we let u = u, u and the map u u is a norm. Recall that every linear map, f : E E, hasanadjoint f which is a linear map, f : E E, suchthat for all u, v E. f(u),v = u, f (v), Since, is symmetric, it is obvious that f = f. Definition 9.1. Given a Euclidean (or Hermitian) space, E, alinearmapf : E E is normal iff f f = f f. Alinearmapf : E E is self-adjoint if f = f, skewself-adjoint if f = f,andorthogonal if f f = f f =id.

9.1. NORMAL LINEAR MAPS 363 Our first goal is to show that for every normal linear map f : E E (where E is a Euclidean space), there is an orthonormal basis (w.r.t., ) suchthatthematrix of f over this basis has an especially nice form: It is a block diagonal matrix in which the blocks are either one-dimensional matrices (i.e., single entries) or twodimensional matrices of the form λ µ µ λ This normal form can be further refined if f is self-adjoint, skew-self-adjoint, or orthogonal. As a first step, we show that f and f have the same kernel when f is normal. Proposition 9.1. Given a Euclidean space E, if f : E E is a normal linear map, then Ker f =Kerf.

364 CHAPTER 9. SPECTRAL THEOREMS The next step is to show that for every linear map f : E E, thereissomesubspacew of dimension 1 or 2suchthatf(W ) W. When dim(w )=1,W is actually an eigenspace for some real eigenvalue of f. Furthermore, when f is normal, there is a subspace W of dimension 1 or 2 such that f(w ) W and f (W ) W. The difficulty is that the eigenvalues of f are not necessarily real. One way to get around this problem is to complexify both the vector space E and the inner product,. First, we need to embed a real vector space E into a complex vector space E C.

9.1. NORMAL LINEAR MAPS 365 Definition 9.2. Given a real vector space E, lete C be the structure E E under the addition operation (u 1,u 2 )+(v 1,v 2 )=(u 1 + v 1,u 2 + v 2 ), and multiplication by a complex scalar z = x+iy defined such that (x + iy) (u, v) =(xu yv, yu + xv). The space E C is called the complexification of E. It is easily shown that the structure E C is a complex vector space. It is also immediate that (0, v)=i(v, 0), and thus, identifying E with the subspace of E C consisting of all vectors of the form (u, 0), we can write (u, v) =u + iv. Given a vector w = u + iv, itsconjugate w is the vector w = u iv.

366 CHAPTER 9. SPECTRAL THEOREMS Observe that if (e 1,...,e n )isabasisofe (a real vector space), then (e 1,...,e n )isalsoabasisofe C (recall that e i is an abreviation for (e i, 0)). Given a linear map f : E E, themapf can be extended to a linear map f C : E C E C defined such that f C (u + iv) =f(u)+if(v). For any basis (e 1,...,e n )ofe, thematrixm(f) representing f over (e 1,...,e n )isidenticaltothematrix M(f C )representingf C over (e 1,...,e n ), where we view (e 1,...,e n )asabasisofe C. As a consequence, det(zi M(f)) = det(zi M(f C )), which means that f and f C have the same characteristic polynomial (which has real coefficients). We know that every polynomial of degree n with real (or complex) coefficients always has n complex roots (counted with their multiplicity), and the roots of det(zi M(f C )) that are real (if any) are the eigenvalues of f.

9.1. NORMAL LINEAR MAPS 367 Next, we need to extend the inner product on E to an inner product on E C. The inner product, on a Euclidean space E is extended to the Hermitian positive definite form, C on E C as follows: u 1 + iv 1,u 2 + iv 2 C = u 1,u 2 + v 1,v 2 + i(u 2,v 1 u 1,v 2 ). Then, given any linear map f : E E,itiseasilyverified that the map fc defined such that f C(u + iv) =f (u)+if (v) for all u, v E, istheadjoint of f C w.r.t., C.

368 CHAPTER 9. SPECTRAL THEOREMS Assuming again that E is a Hermitian space, observe that Proposition 9.1 also holds. Proposition 9.2. Given a Hermitian space E, for any normal linear map f : E E, a vector u is an eigenvector of f for the eigenvalue λ (in C) iffu is an eigenvector of f for the eigenvalue λ. The next proposition shows a very important property of normal linear maps: eigenvectors corresponding to distinct eigenvalues are orthogonal. Proposition 9.3. Given a Hermitian space E, for any normal linear map f : E E, if u and v are eigenvectors of f associated with the eigenvalues λ and µ (in C) where λ = µ, then u, v =0.

9.1. NORMAL LINEAR MAPS 369 We can also show easily that the eigenvalues of a selfadjoint linear map are real. Proposition 9.4. Given a Hermitian space E, the eigenvalues of any self-adjoint linear map f : E E are real. There is also a version of Proposition 9.4 for a (real) Euclidean space E and a self-adjoint map f : E E. Proposition 9.5. Given a Euclidean space E, if f : E E is any self-adjoint linear map, then every eigenvalue of f C is real and is actually an eigenvalue of f. Therefore, all the eigenvalues of f are real. Given any subspace W of a Hermitian space E, recall that the orthogonal W of W is the subspace defined such that W = {u E u, w =0, for all w W }.

370 CHAPTER 9. SPECTRAL THEOREMS Recall that E = W W (construct an orthonormal basis of E using the Gram Schmidt orthonormalization procedure). The same result also holds for Euclidean spaces. As a warm up for the proof of Theorem 9.9, let us prove that every self-adjoint map on a Euclidean space can be diagonalized with respect to an orthonormal basis of eigenvectors. Theorem 9.6. Given a Euclidean space E of dimension n, for every self-adjoint linear map f : E E, there is an orthonormal basis (e 1,...,e n ) of eigenvectors of f such that the matrix of f w.r.t. this basis is a diagonal matrix with λ i R. λ 1... λ 2............ λ n,

9.1. NORMAL LINEAR MAPS 371 One of the key points in the proof of Theorem 9.6 is that we found a subspace W with the property that f(w ) W implies that f(w ) W. In general, this does not happen, but normal maps satisfy astrongerpropertywhichensuresthatsuchasubspace exists. The following proposition provides a condition that will allow us to show that a normal linear map can be diagonalized. It actually holds for any linear map. Proposition 9.7. Given a Hermitian space E, for any linear map f : E E, if W is any subspace of E such that f(w ) W and f (W ) W, then f(w ) W and f (W ) W. The above Proposition also holds for Euclidean spaces. Although we are ready to prove that for every normal linear map f (over a Hermitian space) there is an orthonormal basis of eigenvectors, we now return to real Euclidean spaces.

372 CHAPTER 9. SPECTRAL THEOREMS If f : E E is a linear map and w = u + iv is an eigenvector of f C : E C E C for the eigenvalue z = λ + iµ, whereu, v E and λ, µ R, since and f C (u + iv) =f(u)+if(v) f C (u + iv) =(λ + iµ)(u + iv) = λu µv + i(µu + λv), we have f(u) =λu µv and f(v) =µu + λv, from which we immediately obtain f C (u iv) =(λ iµ)(u iv), which shows that w = u iv is an eigenvector of f C for z = λ iµ. Using this fact, we can prove the following proposition:

9.1. NORMAL LINEAR MAPS 373 Proposition 9.8. Given a Euclidean space E, for any normal linear map f : E E, if w = u + iv is an eigenvector of f C associated with the eigenvalue z = λ + iµ (where u, v E and λ, µ R), if µ = 0 (i.e., z is not real) then u, v =0and u, u = v, v, which implies that u and v are linearly independent, and if W is the subspace spanned by u and v, then f(w )=W and f (W )=W. Furthermore, with respect to the (orthogonal) basis (u, v), the restriction of f to W has the matrix λ µ. µ λ If µ =0, then λ is a real eigenvalue of f and either u or v is an eigenvector of f for λ. IfW is the subspace spanned by u if u = 0, or spanned by v = 0if u =0, then f(w ) W and f (W ) W.

374 CHAPTER 9. SPECTRAL THEOREMS Theorem 9.9. (Main Spectral Theorem) Given a Euclidean space E of dimension n, for every normal linear map f : E E, there is an orthonormal basis (e 1,...,e n ) such that the matrix of f w.r.t. this basis is a block diagonal matrix of the form A 1... A 2............ A p such that each block A j is either a one-dimensional matrix (i.e., a real scalar) or a two-dimensional matrix of the form A j = λj µ j µ j λ j where λ j,µ j R, with µ j > 0.

9.1. NORMAL LINEAR MAPS 375 After this relatively hard work, we can easily obtain some nice normal forms for the matrices of self-adjoint, skewself-adjoint, and orthogonal, linear maps. However, for the sake of completeness, we state the following theorem. Theorem 9.10. Given a Hermitian space E of dimension n, for every normal linear map f : E E, there is an orthonormal basis (e 1,...,e n ) of eigenvectors of f such that the matrix of f w.r.t. this basis is a diagonal matrix where λ j C. λ 1... λ 2............ λ n Remark: Thereisaconverse to Theorem 9.10, namely, if there is an orthonormal basis (e 1,...,e n )ofeigenvectors of f, thenf is normal.

376 CHAPTER 9. SPECTRAL THEOREMS 9.2 Self-Adjoint, Skew-Self-Adjoint, and Orthogonal Linear Maps Theorem 9.11. Given a Euclidean space E of dimension n, for every self-adjoint linear map f : E E, there is an orthonormal basis (e 1,...,e n ) of eigenvectors of f such that the matrix of f w.r.t. this basis is a diagonal matrix where λ i R. λ 1... λ 2............ λ n Theorem 9.11 implies that if λ 1,...,λ p are the distinct real eigenvalues of f and E i is the eigenspace associated with λ i,then E = E 1 E p, where E i and E j are othogonal for all i = j.

9.2. SELF-ADJOINT AND OTHER SPECIAL LINEAR MAPS 377 Theorem 9.12. Given a Euclidean space E of dimension n, for every skew-self-adjoint linear map f : E E, there is an orthonormal basis (e 1,...,e n ) such that the matrix of f w.r.t. this basis is a block diagonal matrix of the form A 1... A 2............ A p such that each block A j is either 0 or a two-dimensional matrix of the form A j = 0 µj µ j 0 where µ j R, with µ j > 0. In particular, the eigenvalues of f C are pure imaginary of the form ±iµ j, or 0.

378 CHAPTER 9. SPECTRAL THEOREMS Theorem 9.13. Given a Euclidean space E of dimension n, for every orthogonal linear map f : E E, there is an orthonormal basis (e 1,...,e n ) such that the matrix of f w.r.t. this basis is a block diagonal matrix of the form A 1... A 2............ A p such that each block A j is either 1, 1, or a twodimensional matrix of the form cos θj sin θ A j = j sin θ j cos θ j where 0 <θ j <π. In particular, the eigenvalues of f C are of the form cos θ j ± i sin θ j, or 1, or 1.

9.2. SELF-ADJOINT AND OTHER SPECIAL LINEAR MAPS 379 It is obvious that we can reorder the orthonormal basis of eigenvectors given by Theorem 9.13, so that the matrix of f w.r.t. this basis is a block diagonal matrix of the form A 1............ A r I q... I p where each block A j is a two-dimensional rotation matrix A j = ±I 2 of the form with 0 <θ j <π. A j = cos θj sin θ j sin θ j cos θ j The linear map f has an eigenspace E(1,f)=Ker(f id) of dimension p for the eigenvalue 1, and an eigenspace E( 1,f)=Ker(f +id) of dimension q for the eigenvalue 1.

380 CHAPTER 9. SPECTRAL THEOREMS If det(f) = +1 (f is a rotation), the dimension q of E( 1,f)mustbeeven,andtheentriesin I q can be paired to form two-dimensional blocks, if we wish. Remark: Theorem 9.13 can be used to prove a sharper version of the Cartan-Dieudonné Theorem. Theorem 9.14. Let E be a Euclidean space of dimension n 2. For every isometry f O(E), if p =dim(e(1,f)) = dim(ker (f id)), then f is the composition of n p reflections and n p is minimal. The theorems of this section and of the previous section can be immediately applied to matrices.

9.3. NORMAL AND OTHER SPECIAL MATRICES 381 9.3 Normal, Symmetric, Skew-Symmetric, Orthogonal, Hermitian, Skew-Hermitian, and Unitary Matrices First, we consider real matrices. Definition 9.3. Given a real m n matrix A,thetranspose A of A is the n m matrix A =(a ij )defined such that a ij = a ji for all i, j, 1 i m, 1 j n. Arealn n matrix A is 1. normal iff 2. symmetric iff AA = A A, A = A, 3. skew-symmetric iff A = A, 4. orthogonal iff AA = A A = I n.

382 CHAPTER 9. SPECTRAL THEOREMS Theorem 9.15. For every normal matrix A, there is an orthogonal matrix P and a block diagonal matrix D such that A = PDP, where D is of the form D = D 1... D 2............ D p such that each block D j is either a one-dimensional matrix (i.e., a real scalar) or a two-dimensional matrix of the form D j = λj µ j µ j λ j where λ j,µ j R, with µ j > 0.

9.3. NORMAL AND OTHER SPECIAL MATRICES 383 Theorem 9.16. For every symmetric matrix A, there is an orthogonal matrix P and a diagonal matrix D such that A = PDP, where D is of the form where λ i R. D = λ 1... λ 2............ λ n

384 CHAPTER 9. SPECTRAL THEOREMS Theorem 9.17. For every skew-symmetric matrix A, there is an orthogonal matrix P and a block diagonal matrix D such that A = PDP, where D is of the form D = D 1... D 2............ D p such that each block D j is either 0 or a two-dimensional matrix of the form D j = 0 µj µ j 0 where µ j R, with µ j > 0. In particular, the eigenvalues of A are pure imaginary of the form ±iµ j, or 0.

9.3. NORMAL AND OTHER SPECIAL MATRICES 385 Theorem 9.18. For every orthogonal matrix A, there is an orthogonal matrix P and a block diagonal matrix D such that A = PDP, where D is of the form D = D 1... D 2............ D p such that each block D j is either 1, 1, or a twodimensional matrix of the form cos θj sin θ D j = j sin θ j cos θ j where 0 <θ j <π. In particular, the eigenvalues of A are of the form cos θ j ± i sin θ j, or 1, or 1. We now consider complex matrices.

386 CHAPTER 9. SPECTRAL THEOREMS Definition 9.4. Given a complex m n matrix A, the transpose A of A is the n m matrix A = (a ij ) defined such that a ij = a ji for all i, j, 1 i m, 1 j n. Theconjugate A of A is the m n matrix A =(b ij )definedsuchthat b ij = a ij for all i, j, 1 i m, 1 j n. Given an n n complex matrix A, theadjoint A of A is the matrix defined such that A = (A )=(A). Acomplexn n matrix A is 1. normal iff 2. Hermitian iff 3. skew-hermitian iff AA = A A, A = A, A = A, 4. unitary iff AA = A A = I n.

9.3. NORMAL AND OTHER SPECIAL MATRICES 387 Theorem 9.10 can be restated in terms of matrices as follows. We can also say a little more about eigenvalues (easy exercise left to the reader). Theorem 9.19. For every complex normal matrix A, there is a unitary matrix U and a diagonal matrix D such that A = UDU. Furthermore, if A is Hermitian, D is a real matrix, if A is skew-hermitian, then the entries in D are pure imaginary or null, and if A is unitary, then the entries in D have absolute value 1.

388 CHAPTER 9. SPECTRAL THEOREMS