LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Similar documents
LINEAR ALGEBRA BOOT CAMP WEEK 2: LINEAR OPERATORS

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Linear Algebra. Workbook

Numerical Linear Algebra

Review problems for MA 54, Fall 2004.

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7

EXERCISES ON DETERMINANTS, EIGENVALUES AND EIGENVECTORS. 1. Determinants

Math 108b: Notes on the Spectral Theorem

LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Chapter 6: Orthogonality

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Linear algebra and applications to graphs Part 1

GQE ALGEBRA PROBLEMS

Linear Algebra: Matrix Eigenvalue Problems

Lecture 7: Positive Semidefinite Matrices

Linear Algebra 2 Spectral Notes

Supplementary Notes on Linear Algebra

Elementary linear algebra

1 Linear Algebra Problems

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

ALGEBRA 8: Linear algebra: characteristic polynomial

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

MAT Linear Algebra Collection of sample exams

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Lecture 11: Finish Gaussian elimination and applications; intro to eigenvalues and eigenvectors (1)

1. General Vector Spaces

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Diagonalizing Matrices

Exercise Sheet 1.

LINEAR ALGEBRA SUMMARY SHEET.

Introduction to Matrix Algebra

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Math Linear Algebra II. 1. Inner Products and Norms

MATH 115A: SAMPLE FINAL SOLUTIONS

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

A PRIMER ON SESQUILINEAR FORMS

MATH 583A REVIEW SESSION #1

235 Final exam review questions

Quantum Computing Lecture 2. Review of Linear Algebra

x i e i ) + Q(x n e n ) + ( i<n c ij x i x j

Math 25a Practice Final #1 Solutions

Functional Analysis Exercise Class

Symmetric and self-adjoint matrices

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Final A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017

Ph.D. Katarína Bellová Page 1 Mathematics 2 (10-PHY-BIPMA2) EXAM - Solutions, 20 July 2017, 10:00 12:00 All answers to be justified.

4. Linear transformations as a vector space 17

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

Symmetric matrices and dot products

Chapter 4 Euclid Space

Foundations of Matrix Analysis

I. Multiple Choice Questions (Answer any eight)

Worksheet for Lecture 25 Section 6.4 Gram-Schmidt Process

Eigenvalues and Eigenvectors

LINEAR ALGEBRA QUESTION BANK

Definition (T -invariant subspace) Example. Example

Vector Spaces and Linear Transformations

Mathematical Methods wk 2: Linear Operators

Math Abstract Linear Algebra Fall 2011, section E1 Practice Final. This is a (long) practice exam. The real exam will consist of 6 problems.

Linear Algebra: Graduate Level Problems and Solutions. Igor Yanovsky

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Linear Algebra Massoud Malek

Linear algebra 2. Yoav Zemel. March 1, 2012

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

Linear algebra I Homework #1 due Thursday, Oct Show that the diagonals of a square are orthogonal to one another.

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

Chapter 7: Symmetric Matrices and Quadratic Forms

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

implies that if we fix a basis v of V and let M and M be the associated invertible symmetric matrices computing, and, then M = (L L)M and the

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam June 8, 2012

Linear Algebra Lecture Notes-II

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

NOTES ON BILINEAR FORMS

Math 291-2: Lecture Notes Northwestern University, Winter 2016

MATH 369 Linear Algebra

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

1.4 Solvable Lie algebras

Linear Algebra- Final Exam Review

Math Linear Algebra Final Exam Review Sheet

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

INTRODUCTION TO LIE ALGEBRAS. LECTURE 2.

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

7. Symmetric Matrices and Quadratic Forms

Online Exercises for Linear Algebra XM511

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

Math Linear Algebra

MATH 221, Spring Homework 10 Solutions

Algebra Exam Syllabus

There are two things that are particularly nice about the first basis

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

Transcription:

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator L : V V on a finite dimensional vector space is called selfadjoint if L = L and skew-adjoint if L = L. When F = R we use the terms symmetric and skew-symmetric, and when F = C we use the terms Hermitian and skew-hermitian. Definition 2. A positive definite Hermitian inner product is just what we have been calling an inner product all along. Definition 3. A transformation L is a positive definite Hermitian transformation if it is Hermitian (or symmetric in the real case) and satisfies (Lx, x) > 0 for x V \ {0}. Definition 4. A linear map L : V V is said to be completely reducible if every invariant subspace has a complementary invariant subspace. Theorem 5. (If you didn t prove this last week, you should prove it now) The following are equivalent: (1) L(x) = x for all x V. (2) (L(x), L(y)) = (x, y) for all x, y V. (3) L L = 1 (4) LL = 1 (5) L takes orthonormal sets to orthonormal sets. Corollary 6. L is an isometry (an isomorphism satisfying (2) above) if and only if L = L 1. Definition 7. If V = W = R n, the isometries are called orthogonal matrices. If V = W = C n, the isometries are called unitary matrices. (The sets O n of orthogonal matrices and U n of unitary matrices are groups, but we won t use that here). Definition 8. A linear operator L : V V is normal if LL = L L. Warmup. I should have put this on last week s worksheet. (1) Consider a 3 3 real symmetric matrix with determinant 6. Assume (1, 2, 3) and (0, 3, 2) are eigenvectors with eigenvalues 1 and 2, respectively. (a) Give (with proof) an eigenvector of the form (1, x, y) for some real x, y which is linearly independent of the two vectors above. (b) What is the eigenvalue of this eigenvector? F03-9

2 LINEAR ALGEBRA BOOT CAMP WEEK 4 Preliminaries. (2) (Self-and Skew-Adjoint Operators are Completely Reducible) Prove that if L : V V is self- or skew-adjoint then for each invariant subspace M V the orthogonal complement M is also invariant. (3) (Isometries are Completely Reducible) Prove that if L : V V is an isometry then for each invariant subspace M V the orthogonal complement M is also invariant. (4) Suppose that L is self-adjoint. Prove that L = 0 if and only if (L(x), x) = 0 for all x V. Hint: look at (L(x + y), x + y). (5) Suppose that F = C. Prove that L = 0 if and only if (L(x), x) = 0 for all x V. Hint: use the previous hint with x + y and x + iy. (6) Suppose that T is a normal operator. Prove that if v is an eigenvector for T then v is also an eigenvector for T. S07-5 (7) Prove the following characterization of normal operators L : V V. The following are equivalent: (a) LL = L L. (b) L(x) = L (x) for all x V. (c) AB = BA, where A = 1 2 (L + L ) and B = 1 2 (L L ).

LINEAR ALGEBRA BOOT CAMP WEEK 4 3 The Spectral Theorem. In this section we will prove various versions of The Spectral Theorem in several different ways. You already saw a proof of The Spectral Theorem over R last week (#37), but we will see a few more here. In each case, we will use the same basic outline: find an eigenvector x 0 for L, show that (span x) is L-invariant, lather, rinse, and repeat. (8) Prove the Spectral Theorem over C: Theorem 9. Let L : V V be a self-adjoint (i.e. Hermitian) linear operator on a complex inner product space. Then there exists an orthonormal basis of eigenvectors with real eigenvalues. S11-3 (9) Show that if L is a complex skew-adjoint operator, then il is self-adjoint. Conclude that the Spectral Theorem also holds for complex skew-adjoint operators. In this case, what do the eigenvalues look like? (10) Suppose that T : V V is Hermitian and that the matrix representation of T 2 in the standard basis has trace zero. Prove that T = 0. S03-10 (11) Prove that commuting Hermitian operators are simultaneously orthogonally diagonalizable (i.e. there is an orthonormal basis consisting of eigenvectors for both operators). S02-11, S10-3 (12) Let Y be an arbitrary set of commuting matrices in M n (C). Prove that there exists a non-zero vector v which is a common eigenvector for all elements of Y. S06-10 (13) Let A be a Hermitian n n complex matrix. Show that if (Av, v) 0 for all v C n then there exists an n n matrix T so that A = T T. S05-3

4 LINEAR ALGEBRA BOOT CAMP WEEK 4 (14) In this exercise we ll prove the Spectral Theorem over R several times. F01-9, S04-10, S06-9 Theorem 10. Let L : V V be a symmetric linear operator on a real inner product space. Then there exists an orthonormal basis of eigenvectors with real eigenvalues. Note that the proof is identical to the proof in the complex case once we know that we can always find an eigenvector. So we ll actually prove the following: Theorem 11. Let L : V V be a symmetric linear operator on a real inner product space. Then L has an eigenvector with a real eigenvalue. There are several reasonable ways to do this, and this seems like a good place to list a few. (a) Let S = {x V : (x, x) = 1}. Choose x S such that (Lx, x) = sup S (Ly, y). Prove that if (x, y) = 0 then (Lx, y) = 0. Conclude that x is an eigenvector for L with eigenvalue (Lx, x). (You did this last week, so skip it if you want.) (b) With S, x as in (a), the real-valued function f(y) = (Ly, y) y 2 has a maximum at y = x and the maximum value is λ = (Lx, x). So for any y V, the function φ(t) = f(x + ty) has a maximum at t = 0. Compute the derivative φ (t) and use its value at t = 0 to show that x is an eigenvector with eigenvalue λ. (c) Prove that L 2 + bl + c is invertible when L is symmetric and b 2 < 4c. (Hint: Show that ((L 2 + bl + c)v, v) > 0 for all 0 v V. Use this to conclude that L has an eigenvector with a real eigenvalue. (d) Choose an orthonormal basis for V and let A = [L] M n (R) in this basis. Thinking of A as an element of M n (C), show that any eigenvalue of A must be real. Conclude that L has an eigenvector with a real eigenvalue. (15) Let L : V V be a self-adjoint linear operator. Let µ C and ε > 0 be given. Suppose that there is a unit vector x V such that L(x) µx ɛ. Prove that L has an eigenvalue λ with λ µ ε. F11-9

LINEAR ALGEBRA BOOT CAMP WEEK 4 5 (16) Suppose that A is a symmetric n n real matrix with eigenvalues λ 1,..., λ k (here k n and λ j λ k if j k). Find the sets S07-4 { X = x R n : lim { L = } k (xt A 2k x) 1/k exists lim k (xt A 2k x) 1/k : x X }. (17) (a) Let A = (a ij ) be an n n real symmetric matrix with i,j a ijx i x j 0 for every x = (x 1,..., x n ) R n. Prove that if tr(a) = 0 then A = 0. (b) Let T be a linear transformation in a complex vector space V with a positive definite Hermitian inner product. Suppose that T T = 4T 3I. Prove that T is positive definite Hermitian and find all possible eigenvalues of T. S10-4 (18) For each of the following three fields (separately), is it true that every symmetric matrix A M 2 (F) is diagonalizable? (a) F = R (b) F = C (c) F = F 3 = Z/3Z, the field with three elements. Supply proofs/counterexamples (or cite the relevant theorems) for all parts. F17-6 (19) Prove the Spectral Theorem for Normal Operators. F02-10 Theorem 12. Let V be a complex inner product space. Then the linear operator L : V V is normal if and only if there is an orthonormal basis of eigenvectors. For the direction = : instead of invoking the Fundamental Theorem of Algebra to find an eigenvector, do the following. (This sort of reasoning will be useful later.) (a) Decompose L = A + ib = 1 2 (L + L ) + i 1 2i (L L ). Find an eigenvalue α for A and, within the α-eigenspace for A, find an eigenvector x for B with eigenvalue β. Conclude that α + iβ is an eigenvalue for L. (b) Show that span{x} is L- and L -invariant, and show that M = (span{x}) is also L- and L -invariant. Conclude that (L M ) = L M and finish the proof. (20) Show that if A M n (C) is normal then A = P (A) for some polynomial P (x) C[x]. S14-6

6 LINEAR ALGEBRA BOOT CAMP WEEK 4 (21) Mimic the proof of the Spectral Theorem for Normal Operators and prove the following. Hint: if A = 1 2 (L + LT ) and B = 1 2 (L LT ) then L = A + B, where A and B commute, A is symmetric, and B is skew-symmetric. Use that B 2 is symmetric. Theorem 13. Let L : V V be a normal operator on a real inner product space. Then there is an orthonormal basis where k + 2l = n and where λ i, α j, β j R. e 1,..., e k, x 1, y 1,..., x l, y l, L(e i ) = λ i e i, L(x j ) = α j x j + β j y j, L(y j ) = β j x j + α j y j, (22) Even when the operator L : V V has no special properties which ensure an orthonormal basis of eigenvectors, we can still get a reasonably nice basis such that [L] is upper triangular. Prove Schur s Theorem: Theorem 14 (Schur s Theorem). Let L : V V be an operator on a complex inner product space. Then there is an orthonormal basis e 1,..., e n such that a 11 a 12 a 1n 0 a 22 a 2n L = [e 1 e n ]...... [e 1 e n ]. 0 0 a nn W02-11, W06-10 Quadratic Forms. A quadratic form is a function Q : R n R defined by Q(x) = (Ax, x), where A is a symmetric, real matrix. We say that Q is positive (resp. negative) definite if Q(x) > 0 (resp. Q(x) < 0) for all x 0. If > is replaced by (or < by ) then we say Q is positive (negative) semidefinite. A symmetric matrix is positive/negative definite/semidefinite if the corresponding quadratic form has that property. (23) Let M be an n n real matrix. Suppose that M is orthogonal and symmetric. (a) Prove that if M is positive definite then M is the identity. (b) Does the answer change if M is only positive semidefinite? S17-3

LINEAR ALGEBRA BOOT CAMP WEEK 4 7 (24) Let A, B be two positive definite 2 2 matrices. Prove or disprove: (a) A + B is also positive definite. (b) AB + BA is also positive definite. S18-5 (25) Let Q(x) be a quadratic form on R n. Show that there is an orthogonal basis where where 0 k l n. Q(z) = z 2 1... z 2 k + z2 k+1 +... + z2 l, (26) Prove the following special case of Descartes Rule of Signs: Theorem 15. Let p(t) = t n + a n 1 t n 1 +... + a 1 t + a 0 = (t λ 1 ) (t λ n ), where a i, λ i R. (a) All roots of p are negative if and only if a i > 0 for all i. (b) All roots of p are positive if and only if a n 1 < 0 and sign(a i ) alternates. (27) Let f(x, y, z) = 9x 2 + 6y 2 + 6z 2 + 12xy 10xz 2yz. Is there a point (x, y, z) such that f(x, y, z) < 0? S15-7 (28) A bilinear form on a vector space V is a function B : V V F such that x B(x, y) and y B(x, y) are both linear. Show that the quadratic form Q always looks like Q(x) = B(x, x) where B is a symmetric bilinear form (i.e. B(x, y) = B(y, x)). Orthogonal Transformations. (29) Suppose A is a real n n orthogonal matrix. (a) Show that A is similar to a matrix which consists of 2 2 blocks down the diagonal along with some diagonal elements that are ±1 with the 2 2 blocks being rotation matrices of the form ( ) cos θ sin θ. sin θ cos θ (b) Use part (a) to show that if n is odd, then there is a nonzero vector v such that A 2 v = v. S13-9

8 LINEAR ALGEBRA BOOT CAMP WEEK 4 (30) An orthogonal n n matrix A is called elementary if the corresponding linear transformation L A : R n R n fixes an (n 2)-dimensional subspace. Prove that every orthogonal matrix M is a product of at most (n 1) elementary orthogonal matrices. S16-7 (31) Let T : R 3 R 3 be the rotation by 60 counterclockwise about the plane perpendicular to the vector (1, 1, 1) and let S : R 3 R 3 be the reflection about the plane perpendicular to the vector (1, 0, 1). Determine the matrix representation of S T in the standard basis. You do not have to multiply the resulting matrices but you must determine any inverses that arise. F01-8 (32) Let T : R 3 R 3 be the rotation by 60 counterclockwise about the plane perpendicular to the vector (1, 1, 2). (a) Determine the matrix representation of T in the standard basis. Find all eigenvalues and eigenspaces of T. (b) What are the eigenvalues and eigenspaces of T if R 3 is replaced by C 3? F01-8, S04-8 (33) Find the matrix representation in the standard basis for either rotation by an angle θ in the plane perpendicular to the subspace spanned by the vectors (1, 1, 1, 1) and (1, 1, 1, 0) in R 4. You do not have to multiply the resulting matrices but you must determine any inverses that arise. S02-9 Extra Problems. (34) Let H be an n n Hermitian matrix with non-zero determinant. Use H to define an Hermitian form [, ] by the formula [x, y] = x t Hy. Let W be a complex subspace of C n such that [w 1, w 2 ] = 0 for all w 1, w 2 W. Show that dim W n/2. Also give an example of an H for which dim W = n/2 if n is even and dim W = (n 1)/2 if n is odd. F12-8 (35) Let x 1,..., x k be vectors in an inner product space V. Show that the k k matrix G(x 1,..., x k ) whose ij entry is (x i, x j ) is self-adjoint and that all its eigenvalues are nonnegative.

LINEAR ALGEBRA BOOT CAMP WEEK 4 9 (36) Show that the three subsets of Hom(V, V ) defined by M 1 = span I, M 2 = {L : L is skew-adjoint}, M 2 = {L : tr L = 0 and L is self-adjoint}, are subspaces over R, are mutually orthogonal with respect to the real inner product (L, K) =Re(tr(L K)), and yield a direct sum decomposition of Hom(V, V ). (37) Suppose that L : V V is normal. Without invoking the spectral theorem, show the following. (a) ker(l) = ker(l k ) for any k 1. (b) im(l) = im(l k ) for any k 1. (c) ker(l λi) = ker((l λi) k ) for any k 1. (d) Show that the minimal polynomial of L has no multiple roots. (38) Let L : V V be a linear operator on an inner product space that preserves right angles (i.e. (L(x), L(y)) = 0 if (x, y) = 0). (a) Show that if x = y and (x, y) = 0 then L(x) = L(y). (b) Show that L = λu, where U is an isometry. (39) Let V be a real inner product space and F : V V a bijective map that preserves distances, i.e. for all x, y V we have F (x) F (y) = x y. (a) Show that G(x) = F (x) F (0) also preserves distances and that G(0) = 0. (b) Show that G(x) = x for all x V. (c) Show that (G(x), G(y)) = (x, y) for all x, y V. (d) If e 1,..., e n is an orthonormal basis, show that G(e 1 ),..., G(e n ) is an orthonormal basis. (e) Show that G(x) = (x, e 1 )G(e 1 ) +... + (x, e n )G(e n ) and conclude that G is linear. (f) Conclude that F (x) = L(x) + F (0) for some linear isometry L. (40) Consider the Killing Form on Hom(V, V ), where 2 dim V <, defined by K(L, K) = tr L tr K tr(lk). (a) Show that K(L, K) = K(K, L). (b) Show that K K(L, K) is linear. (c) Assume in addition that V is an inner product space. Show that K(L, L) > 0 if L is skew-adjoint and nonzero. Show that K(L, L) < 0 if L is self-adjoint and nonzero. (d) Show that K is nondegenerate: if L 0 then there exists a K such that K(L, K) 0.