X i, AX i X i. (A λ) k x = 0.

Similar documents
Notes on the matrix exponential

MATHEMATICS 217 NOTES

1. General Vector Spaces

Foundations of Matrix Analysis

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

Linear algebra and applications to graphs Part 1

THE MINIMAL POLYNOMIAL AND SOME APPLICATIONS

MATH 583A REVIEW SESSION #1

Lecture Summaries for Linear Algebra M51A

MATH 5640: Functions of Diagonalizable Matrices

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

LINEAR ALGEBRA BOOT CAMP WEEK 2: LINEAR OPERATORS

NORMS ON SPACE OF MATRICES

Linear Algebra. Workbook

Linear Algebra Highlights

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Differential equations

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

2 b 3 b 4. c c 2 c 3 c 4

Recall : Eigenvalues and Eigenvectors

Math 489AB Exercises for Chapter 2 Fall Section 2.3

Linear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8.

0.1 Rational Canonical Forms

The Jordan canonical form

4. Linear transformations as a vector space 17

LINEAR ALGEBRA MICHAEL PENKAVA

The following definition is fundamental.

ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices

Linear Algebra Lecture Notes-II

GRE Subject test preparation Spring 2016 Topic: Abstract Algebra, Linear Algebra, Number Theory.

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued)

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

Topic 1: Matrix diagonalization

Eigenvalues and Eigenvectors

Control Systems. Linear Algebra topics. L. Lanari

Symmetric and self-adjoint matrices

Further linear algebra. Chapter IV. Jordan normal form.

Elementary linear algebra

Jordan Normal Form. Chapter Minimal Polynomials

Matrix Theory. A.Holst, V.Ufnarovski

LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS

Online Exercises for Linear Algebra XM511

av 1 x 2 + 4y 2 + xy + 4z 2 = 16.

Linear Algebra: Matrix Eigenvalue Problems

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Math Linear Algebra Final Exam Review Sheet

Spectral radius, symmetric and positive matrices

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Chapter 1. Matrix Calculus

Linear Algebra Practice Problems

ELEMENTARY LINEAR ALGEBRA

Topics in linear algebra

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

Dimension. Eigenvalue and eigenvector

Knowledge Discovery and Data Mining 1 (VO) ( )

(f + g)(s) = f(s) + g(s) for f, g V, s S (cf)(s) = cf(s) for c F, f V, s S

Math 554 Qualifying Exam. You may use any theorems from the textbook. Any other claims must be proved in details.

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Linear Algebra. Min Yan

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Introduction to Quantitative Techniques for MSc Programmes SCHOOL OF ECONOMICS, MATHEMATICS AND STATISTICS MALET STREET LONDON WC1E 7HX

Chapter 4 Euclid Space

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

Eigenvalues and Eigenvectors A =

A Brief Outline of Math 355

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS

A proof of the Jordan normal form theorem

Eigenpairs and Diagonalizability Math 401, Spring 2010, Professor David Levermore

Linear Algebra in Actuarial Science: Slides to the lecture

Spectral Theorem for Self-adjoint Linear Operators

Putzer s Algorithm. Norman Lebovitz. September 8, 2016

Mathematics 1. Part II: Linear Algebra. Exercises and problems

ELEMENTARY LINEAR ALGEBRA

Algebra Exam Syllabus

Q N id β. 2. Let I and J be ideals in a commutative ring A. Give a simple description of

CS 246 Review of Linear Algebra 01/17/19

Given a finite-dimensional vector space V over a field K, recall that a linear

Massachusetts Institute of Technology Department of Economics Statistics. Lecture Notes on Matrix Algebra

235 Final exam review questions

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Mathematical Methods wk 2: Linear Operators

π 1 = tr(a), π n = ( 1) n det(a). In particular, when n = 2 one has

GQE ALGEBRA PROBLEMS

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

Linear Algebra Practice Final

Chapter 3. Determinants and Eigenvalues

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Lecture 10 - Eigenvalues problem

Math Matrix Algebra

Problem 1A. Suppose that f is a continuous real function on [0, 1]. Prove that

SPRING 2006 PRELIMINARY EXAMINATION SOLUTIONS

Chapter 0 Preliminaries

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.

Transcription:

Chapter 4 Spectral Theory In the previous chapter, we studied spacial operators: the self-adjoint operator and normal operators. In this chapter, we study a general linear map A that maps a finite dimensional vector space V over a field F into itself. Since for a general mapping, a Euclidean structure has little use, we shall ingore it. Instead, we shall assume that the field is closed; i.e., every polynomial can be factorized into linear factors. For convenience, we use the notation AX := {Ax x X}. A subspace X is called invariant under A if AX X. We also call a linear map a linear operator, since the underlying physical vector spaces could have a much broader meaning than merely vector spaces. The basic question in linear algebra here is to find a basis such that the matrix of a given map under the basis is as diagonal as possible. That is to say, we want to find a finest decomposition V = i X i, AX i X i. Here finest means that X i has as small dimension as possible. Since AX i X i, the study of A on V is then reduced to the study of A on each X i, which should be significantly smaller than V. Clearly, the best scenario would be dim(x i ) = 1. In this case, X i = span{x} for some x and since Ax X i, there exists a scalar λ such that Ax = λx. If λ F and x V \ {0} satisfies Ax = λx, we call λ an eigenvalue, x an eigenvector associated with λ, and (λ, x) an eigenpair. A problem of finding eigenpairs is called an eigenvalue problem. If we can find eigenpairs (λ 1, u 1 ),, (λ n, u n ) of a map A such that {u 1,, u n } is a basis, then the linear map A can be describes by A(x i u i ) = i λ ix i u i, and the corresponding matrix of the map A under the basis is diag(λ 1,, λ n ). There are cases where we cannot find such a basis; namely, there do not exist n linearly independent eigenvectors. The spectral theory tells us to look for generalized eigenvectors, namely, solutions to (A λ) k x = 0. The spectral theory says that we can find a basis that consists of eigenvectors and generalized eigenvectors. As we have seen before, a good representation of a linear operator allows us to find quickly many other related linear operators such as e A, sin A, etc. 123

124 CHAPTER 4. SPECTRAL THEORY 4.1 Examples of Linear Operators Example 4.1. Let F = R and V n be the set of all polynomials of degree n. If we denote by e i the ith power function t R t i, then V n = span{e 0,, e n }. Set V = n=1 V n. Let A : V V be the map defined by Af = f(0)e 0 + bf(1)e 1 f V where b R is a fixed number and f(t) is the value of f at t. It is easy to show that A is a linear map from V n to V n itself, as long as n 1. (1) Let u i = e i e 1 for i 2. Then u i (0) = u i (1) = 0 so that For n 2, set U = span{u 2,, u n }. Then Au i = 0. V n = V 1 U, AV 1 V 1, AU = {0}. Hence, to study A, we need only concentrate our attention to the behavior of A on V 1. (2) Now consider A on V 1. Under the basis {e 0, e 1 } its associated matrix is ( ) 1 0 A = since A(e b b 0 e 1 ) := (Ae 0 Ae 1 ) = (e 0 e 1 )A. Suppose b 1. Then we can find two eigenpairs, (1, (b 1)e 0 be 1 ) and (b, e 1 ). Hence, under the basis {(b 1)e 0 be 1, e 1 } the matrix associated with A takes the form ( ) 1 0 Ã = 0 b Suppose b = 1. Then we can find only one eigenpair (1, e 1 ), all the others are of the form (1, ce 1 ) (c 0), so that we cannot find a basis consisting of eigenvectors. On the other hand, denoting I the identity map, we have (A I)e 0 = e 1, (A I)e 1 = 0, (A I) 2 f = 0 f V 1. Example 4.2. With same vector space as in the previous example, consider the linear differential operator B = d dt + I : f V f + f. It is easy see that B is a linear map from V n to V n itself, for each n 0. If we focus our attention to V 1. The corresponding matrix under the basis {e 1, e 0 } is ( ) 1 0. 1 1 This is the same matrix as in the previous example with b = 1. It is important to note that in the previous case the basis is {e 0, e 1 } whereas in the current case the basis is {e 1, e 0 }.

4.1. EXAMPLES OF LINEAR OPERATORS 125 In V n (n 2), the matrix of the linear operator B under the basis {e 0,, e n } is 1 1 0 0 0 1 2 0...... 0 1 n 0 0 0 1 If we use the basis { 1 n! e n,, 1 0! e 0}, the matrix becomes 1 0 0 1 1 0...... 0 1 1 Example 4.3. Consider F = R, W = span{sin πt, cos πt}, and the differential operator B := d dt + I. It is a linear operator from W to itself. Under the basis {sin πt, cos πt}, the matrix of B is ( ) 1 π. π 1 One can show that this matrix does not have any real eigenvalues. Indeed, the eigenvalues are 1 ± πi. If F = C, then under the basis {sin(πt) + i cos(πt), sin(πt) i cos(πt)} the matrix becomes ( 1 + πi 0 ) 0 1 πi Example 4.4. Let U = M 2 (R), a four dimensional vector space. Fix a matrix A = (a ij ) U, we can define a linear operator A : B U AB U. Let e 1 = ( ) 1 0, e 0 0 2 = ( ) 0 1, e 0 0 3 = ( ) 0 0, e 1 0 4 = ( ) 0 0. 0 1 Then {e 1,, e 4 } is a basis of U. Under this basis, the matrix associated with the linear operator A is a 11 0 a 12 0 0 a 11 0 a 12 a 21 0 a 22 0 0 a 21 0 a 22 If we use the basis {e 1, e 3, e 2, e 4 } the matrix of the operator A becomes a 11 a 12 0 0 a 21 a 22 0 0 0 0 a 11 a 12 0 0 a 21 a 22

126 CHAPTER 4. SPECTRAL THEORY Example 4.5. Let V = C0([0, 2 π]) be the space of all twice continuous differentiable functions on [0, 2π] that vanish at 0 and π. Consider the linear operator A = d2 dx. The eigenvalue problem is to find λ C 2 and u V such that u (x) = λu(x) x (0, π), u(0) = u(π) = 0. One can obtain a complete set of eigenpairs { n 2, sin(nx)} n=1. If one uses {sin(nx)} n=0 as the basis for V, then the operator d2 dx 2 can be written as A a n sin(nx) = d2 dx 2 an sin(nx) = n 2 a n sin(nx). Also, for each integer n 1 and real number t, e ta sin(nx) = k=0 1 k! (ta)k sin(nx) = ( n 2 t) k sin(nx) = e n2t sin(nx). k! k=1 Now we consider the differential equation w t = Aw for w : t [0, ) w(t) V. Supplied with an initial data w(0) = u 0 = a n sin(nx) V, the solution is w(t) = e ta u 0 = e n2t a n sin(nx). Here, one first compute the solution in a formal level, then uses a rigorous mathematics to verify the result. Exercise 1. For the operator in example 4.1, find the matrix associated with the operator A under the basis {e 0, e 1,, e n } for V n. Exercise 2. Assume that V = m i=1 X i and AX i X i. Let U i = {x 1 i,, xni i } be a basis for X i, i = 1,, m. Show that under the basis {U 1,, U m }, the matrix associated with the map A takes the form A 1 0... 0 A m where A i is an n i n i matrix associated with the map A on X i. Exercise 3. Suppose A is the matrix of A under the basis {u 1,, u n }. What is the matrix B of A under the basis {u n,, u 1 }? Does B has any relation with A T? If we write B = P 1 AP, what is P? Exercise 4. In example 4.4, find the matrix for the linear map: B U BA U. Extend the result to the case U = M 3 (F). Exercise 5. Carry out the same formal analysis as in example 4.5 for A = 2 x + 2 2 y on the space V of 2 smooth functions on [0, π] [0, π] with zero boundary data. Also, find solutions to initial value problem w t = Aw for w : t [0, ) V with initial data w(0) = w 0 = m,n a mn sin(nx) sin(my).

4.2. FUNCTIONS OF MATRICES 127 4.2 Functions of Matrices When A is a square matrix, we can form a polynomial of the matrix p(a) = a 0 I + a 1 A + + a n A n. For real or complex matrices, we can also define transcendental functions: 1 exp(a) = n! An. Then sin(a) = n=1 k=1 ( 1) k (2k + 1)! A2k+1, exp(ia) = cos A + i sin A = ln(i + A) = k=0 ( 1) k 2k + 1 A2k+1. n=1 For diagonal matrices, the calculation is very simple. Suppose Λ = diag(λ 1,, λ n ) Λ k = diag(λ k 1,, λ k n), exp(λ) = diag(e λ1,, e λn ) 1 n! (ia)n, Approximate any continuous function f : R R by polynomials, we also can extend it to be functions of square matrices: f(λ) = diag(f(λ 1 ),, f(λ n )). For general square matrix A, the direct computation of these functions from their series expansion is time consuming. However, if A is diagonalizable, the computation will be very easy if we diagonalize it. Suppose A = P 1 ΛP Then, since A k = P Λ k P, exp(a) = i=0 1 n! An = P 1 e Λ P. f(a) = P 1 f(λ)p. Exercise 6. (i) Using the serious expansion show that for any square matrix A, d e (t+τ)a e ta exp(ta) := lim = Ae ta = e ta A. dt τ 0 τ (ii) Show that for any A M n (C), B M n,p (C) and C M p,n (C), the functions X(t) = e ta B and Y (t) = Ce ta solve the system of differential equations ( ) 1 2 2 5 dx dt = AX, dy dt = Y A. (iii) For A = work out the details of (i) and (ii). In particular, (a) write down the system ( ) x y of equations, in scalar form for X = and similarly for Y = (u v). (b) write down, in scalar z w form, the general solutions you get.

128 CHAPTER 4. SPECTRAL THEORY 4.3 Spectrum The eigenvalue problem can be written as (A λi)x = 0 or (A λi)x = 0, (4.1) where A is the matrix of the operator A under a basis, say {e 1,, e n }, and x F n is the coordinates of x under the basis. Using determinants, we know that a necessary and sufficient condition for λ to be an eigenvalue is det(a λi) = 0. Definition 4.1. (1) Let A M n (F). The characteristic polynomial of A is p A (z) := det(zi A) = z n Tr(A) z n 1 + + ( 1) n det(a). (2) The spectrum of A is the set σ(a) := {λ F p A (λ) = 0} An element λ σ(a) is called an eigenvalue of A, and its multiplicity as eigenvalue is its multiplicity as the root to the equation p A ( ) = 0. In the sequel, a polynomial is called monic if its highest order term has coefficient 1. A monic polynomial cannot be identically zero. To define the characteristic polynomial of a linear operator, we need Lemma 4.1. If B is similar to A, i.e., B = P 1 AP for some invertiable matrix P, then their characteristic polynomials are identical: p A ( ) = p B ( ). Definition 4.2. (1) Let A be a linear operator mapping an n-dimensional vector space V into itself. Let A be the corresponding matrix of A under an arbitrarily chosen basis. Then p A ( ) is called the characteristic polynomial of the map A, written as p A (z) = z n + a 1 z n 1 + + a n. Here a 1 is called the trace of A and ( 1) n a n the determinant of A. (2) Similarly, σ(a) is called the spectrum of A, written as σ(a). (3) For λ σ(a), the set E(λ) := ker ((λi A) i) i=1 is called the eigenspace of A associated with the eigenvalue λ. Here ker(b), the kernel of B, is the set ker(b) := {x V Bx = 0}. One notices from the Lemma that p A is well-defined. It is intrinsic, in the sense that p A ( ) is independent of the choices of bases for the underlying vector space. We leave it as an exercise to show that an eigenspace is a subspace. In the infinite space dimensional cases, characteristic polynomial of an operator is not defined, neither are the trace and determinant. Nevertheless, the spectrum and eigenspace of an operator mapping a vector space into itself is well defined.

4.3. SPECTRUM 129 Corollary 4.1. Given a linear operator A from a finite dimensional vector space to itself, σ(a) = F \ ρ(a), ρ(a) := {λ F λi A is an isomorphism }. In addition, E(λ) = {0} if λ ρ(a). Remark 4.1. One notices that the above formula does not involve any coordinate systems. Indeed, it can be used as the definition of spectrum for linear operators that map an infinite dimensional vector space into itself. For normed vector spaces, one needs a stronger condition for ρ(a): ρ(a) := {λ C λi A is an isomorphism and (λi A) 1 is bounded }. Then the set σ(a) := C \ ρ(a) consists of two parts: The part σ p (A), called point spectrum, in which each λ admits at least one eigenvector, and the part σ c (A), called continuum spectrum in which λ does not admits any eigenvector, namely (λi A) is one-to-one, but (λi A) 1 is not bounded. The following result will be proven later. Lemma 4.2. If p(z) = n i=0 (z λ i) is the characteristic polynomial of a matrix A M n (F), then for any polynomial f(s) = m i=0 a is i with coefficients in F, the characteristic polynomial of f(a) := m i=1 a ia i is p f (z) = n i=1 (z f(λ i)). Example 4.6. Suppose the characteristic polynomial for A M 3 (R) is p(z) = z(z 2 + 1). What is the characteristic polynomial for A 2, e A, f(a) where f : C C is analytic (i.e., f(z) = n c nz n is uniformly convergent in z C.) Solutions. The eigenvalues of A is 0, i, and i. Hence, the eigenvalues of f(a) is f(0), f(i), f( i). Thus, the characteristic polynomial for f(a) is p f (z) = (z f(0))(z f(i))(z f( i)). In particular, p A 2(z) = z(z + 1) 2, p A 3(z) = z(z 2 + 1), p e A(z) = (z 1)(z e i )(z e i ) = z(z 2 2z cos 1 + 1). Exercise 7. Prove Lemma 4.1 and Corollary 4.1. Exercise 8. Show that an eigenspace of a linear map is a subspace. Also show that in an n dimensional vector space, there exists k n such that ( E(λ) = ker (λi A) k). Exercise 9. Suppose the characteristic polynomial of A M 3 (R) is p(z) = z 3 2z 2 + z. What is the characteristic polynomial for A 2, A 3, e A, sin A, f(a)? Exercise 10. Fix a matrix A M 3 (F). We define a linear operator A : M 3 (F) M 3 (F) by A(B) = AB for all B M 3 (F). Find the characteristic polynomial of the map A. Exercise 11. Let F = R and V n be the set of polynomials of degree n. Let A = d2 dx 2 + 2 d dx + I be the operator mapping f V n to A(f) = f +2f +f. Find the the characteristic polynomial of A for V 1, V 2, and V 3 respectively.

130 CHAPTER 4. SPECTRAL THEORY 4.4 The Cayley-Hamilton Theorem Given a matrix A M n (F), if we regard M n (F) as a n 2 dimensional vector space over F, then the n 2 + 1 matrices I, A, A 2,, A n2 are linearly dependent Hence, there is a non-trivial linear combination that is equal to 0I. The following theorem says that indeed such a combination can be obtained by using only the first n + 1 matrices in the list. Theorem 4.1. (Cayley-Hamilton Theorem) Let A be a linear operator from an n-dimensional vector space to itself and be the characteristic polynomial of A. Then p(z) = z n + a 1 z n 1 + + a n p(a) := A n + a 1 A n 1 + + a n I = 0I. We remark that ( det(zi ) A) is a scalar, whereas p(a) is a matrix, so that p(b) det(bi A). For a b example, let A =. Then its characteristic polynomial is c d p(z) = det(zi A) = z 2 (a + d)z + (ad bc). The Cayley-Hamilton theorem says that there is the identity p(a) = 0I, i.e. ( ) ( ) ( ) ( ) a b a b a b 1 0 (a + d) + (ad bc) = c d c d c d 0 1 ( ) 0 0. 0 0 Proof. Fix a basis and let A be the matrix of A under the basis. Let Q(z) = (q ij (z)) n n be the cofactor matrix of zi A, i.e. q ij (z) = ( 1) i+j det( (zi A) ji ) where (zi A)ji is the n 1 by n 1 matrix obtained by deleting from zi A the jth row and ith column. By using the Laplace expansion, we have Q(z)(zI A) = det(zi A)I = p(z)i. Writing Q(z) = Q 0 + zq 1 + + z n 1 Q n 1 we then have (Q 0 + zq 1 + + z n 1 Q n 1 )(zi A) = (z n + a 1 z n 1 + + a n )I Since F has infinitely many elements, the coefficients on both sides have to be the same. Hence, This implies that Q n 1 = I, Q n 2 Q n 1 A = a 1 I,, Q 0 A = a n I. p(a) = A n + a 1 IA n 1 + + a n I = Q n 1 A n + (Q n 2 Q n 1 A)A n 1 + + ( Q 0 A) = 0I.

4.4. THE CAYLEY-HAMILTON THEOREM 131 Definition 4.3. Let A be a linear operator from a vector space V to itself. A monic polynomial q of minimum degree that annihilates A (e.g. q(a) = 0I) is called the minimal polynomial of A We point out that the minimal polynomial is unique. Indeed, if q 1 and q 2 are polynomials satisfying q 1 (A) = 0I = q 2 (A), then by the following lemma their greatest common divisor r(z) also satisfies r(a) = 0I. In particular, the minimal polynomial is a factor of the characteristic polynomial. Lemma 4.3. (1) Suppose q 1 (z), q 2 (z) are polynomials with coefficients in a field F, not necessarily closed. If r(z) = (q 1 (z), q 2 (z)) is the greatest common divisor, then there exist polynomials a(z) and b(z) such that a(z)p(z) + b(z)q(z) = r(z). Proof. We use a degree reduction process. If deg(p) = deg(r), then set b = 0 and a = 1/p 0 where p 0 is the leading coefficients of p, and we are done. If deg(p) > deg(r), we find the remainder q of q p so that q = q ãp, deg( q) < deg(p), and the greatest common divisor ( q, p) is still r. If deg( q) > deg(r), we replace {p, q} by { q, p} and continue the previous step; otherwise we must have q = q 0 r where q 0 is the leading coefficients of q. After escrowing the remainder equations, we obtain r = ap + bq. ( ) a b Example 4.7. Let A = M c d 2 (R). Find the minimal polynomial of A. Solution. Set α = Tr(A) = a + d and β = det(a) = ad bc. Then the characteristic polynomial of A is p(z) = z 2 αz + β. The minimal polynomial is p(z), except the case A = ai in which the minimal polynomial is q(z) = z a. ( ) 1 2 Example 4.8. Find A 5 + 2A 4 + A 3 + A 2 + A + I where A =. 1 1 Solution. The characteristic polynomial of A is p(z) = z 2 +1. Noting that q(z) := z 5 +2z 4 +z 3 ( +z 2 +z+ ) 3 2 1 = (z 2 +1)(z 3 +2z 2 1)+z+2. It follows from Cayley-Hamilton theorem that q(a) = A+2I =. 1 1 Exercise 12. (1) Let P 1,, P m, Q 1,, Q p be n n matrices, and P (z) = P j z j, Q(z) = Q k z k, R(z) = j k i ( j+k=i P j Q k )z i. Suppose an n n matrix A commutes with each of Q 1,, Q p, show that R(A) = P (A)Q(A). (2) For n = 2, find an example of P and Q such that R(B) P (B)Q(B) for some n n matrix B. Exercise 13. Show that in the field C, the minimal polynomial of a real square matrix has real coefficients. Exercise 14. Show that in the finite dimensional case, a minimal polynomial of a linear operator exists and is unique. Exercise 15. Find the minimal polynomials of the following matrices: 1 0 0 1 0 0 1 0 0 2 3 0, 1 1 0, 1 1 0, 4 5 6 0 1 1 1 1 1 1 0 0 4 1 0. 0 0 1

132 CHAPTER 4. SPECTRAL THEORY Exercise 16. For A = ( ) 1 2 find A 1 1 5 + 2A 4 + 3A 3 + 4A 2 + 5A + 6I. Exercise 17. Show that if two matrices are similar, then their minimal polynomials are identical. Exercise 18. Suppose the characteristic polynomial of A is p(z) = Π i (z λ i ) ni where λ 1,, λ m are distinct. Show that the minimal polynomial q of A has the form q(z) = Π i (z λ) mi with 1 m i n i. Exercise 19. Given a monic polynomial p(z) = z n a 1 t n 1 + a n, show that the minimal and characteristic polynomials of following matrix are the same and equal to p: 0 0 0 a n 1 1 0 0 a n 2........... 0 a 1 0 1 a 0 Exercise 20. Suppose the minimal polynomial of A M 5 (R) is p(z) = (z 1) 2 (z 2 + 1). What is the characteristic polynomial for A, A 2, A 3, A 5, e A, sin A?

4.5. THE SPECTRAL THEOREM 133 4.5 The Spectral Theorem Theorem 4.2. (The Spectral Theorem) Let A be a linear operator from a finite dimensional vector space V over a closed field to V itself. Then V = ker((λi A) k ), AE(λ) E(λ) λ σ(a) E(λ), E(λ) := k 1 That is, V is a direct sum of eigenspaces which are invariant under A; in other words, every vector can be written as a sum of eigenvectors, genuine or generalized. We remark that the theorem may not be true when the field is not closed. That s why in some cases, we need to extend vector spaces over the field R to the field C. The spectral theorem can be extended to infinite dimensional normed vector spaces over C in which the finiteness condition on space dimension is replaced by a requirement that the operator A be a compact operator, i.e., the set {Ax x = 1} is compact. We shall use the following lemma to prove the theorem. Lemma 4.4. Suppose p 1 (z),, p k (z) are polynomials that are pairwise prime to each other, i.e. the greatest common divisor (p i, p j ) = 1 for all i j. Then ( ) ker p i (A) = ker(p i (A)). i i Proof. First we consider the case k = 2. Let a(z) and b(z) be polynomials such that 1 = a(z)p 1 (z)+ b(z)p 2 (z). Then I = a(a)p 1 (A) + b(a)p 2 (A) so that for each x V, we can write x = x 2 + x 1, x 2 := a(a)p 1 (A)x, x 1 := b(a)p 2 (A)x. Now if x ker(p 1 (A)p 2 (A)), then p 1 (A)x 1 = b(a)p 2 (A)p 1 (A)x = 0 so that x 1 ker(p 1 (A)). Similarly, x 2 ker(p 2 (A)). As ker(p i (A)) ker(p 1 (A)p 2 (A)) for i = 1, 2, we see that ker(p 1 (A)p 2 (A)) = ker(p 1 (A)) + ker(p 1 (A)). To show that the sum is direct, let x ker(p 1 (A)) ker(p 2 (A)), then p 1 (A)x = 0 = p 2 (A)x so that x = a(a)p 1 (A)x + b(a)p 2 (A)x = 0. Hence, ker(p 1 (A)) ker(p 2 (A)) = {0}. Thus, ker(p 1 (A)p 2 (A)) = ker(p 1 (A)) ker(p 1 (A)). Now in the general case k 3, we can use the fact that p j is prime to p j+1 p k for j = 1,, k 1 to obtain ker(p 1 (A) p k (A)) = ker(p 1 (A)) ker(p 2 (A) p k (A)) ( ) = ker(p 1 (A)) ker(p 2 (A)) ker(p 3 (A) p k (A)) = = ker(p 1 (A)) ker(p 2 (A)) ker(p k (A)). Proof of the Spectral Theorem. Let p(z) be the characteristic polynomial of A. By the Cayley- Hamilton theorem, p(a) = 0I, i.e. p(a)x = 0 for all x V, so that V = ker(p(a)). As F is closed, we can write p(z) = k i=1 (z λ i) ni where λ i λ j for i j. The assertion of the theorem then follows from the lemma by setting p i (z) = (z λ i ) ni, i = 1, k.

134 CHAPTER 4. SPECTRAL THEORY Exercise 21. (1) Let Z 2 = {m + ni m, n Z}, with summation and multiplication as that of complex numbers. Show that in Z 2, 1 + 2i is prime to both 2 + i and 2 i. However, 1 + 2i is not prime to (2 + i)(2 i) = 5 since 5 = (1 + 2i)(1 2i). (2) Let p 1 (z),, p n (z) be polynomials with coefficients in a field. Show that if p 1 (z) is prime to each of p 2 (z),, p k (z), then p 1 (z) is prime to p 2 (z) p k (z). Exercise 22. Let p(z) = z, q(z) = (z 1) 2, and r(z) = (z 2) 3. Find polynomials a(z), b(z), and c(z) such that a(z)p(z) + b(z)q(z) + c(z)r(z) = 1. Exercise 23. Let {e 1, e 2, e 3 } be the basis of a vector space and A be a linear operator defined by A( i x i e i ) = (x 1 + 2x 2 + 3x 3 )e 1 + (2x 3 + x 3 )e 2 + 5x 3 e 3 (x 1 x 2 x 3 ) F n. Find the spectral decomposition.

4.6. NILPOTENT OPERATOR 135 4.6 Nilpotent Operator Now we study the action of A on its eigenspace E(λ). By working on A λi, we can assume that λ = 0. Since E(0) is invariant, we can assume that E(0) = V. Definition 4.4. A linear operator from a vector space V to itself is called nilpotent if A n = 0I for some positive integer n. Let A be a nilpotent operator over a vector space V. Let k be the positive integer such that A k+1 = 0I A k. Then V = ker(a k+1 ) ker(a k ) ker(a 0 ) = {0}. Here the first inequality follows from A k 0I, whereas the rest inequalities follow from the following: Lemma 4.5. For each j = 1,, k, A maps ker(a j+1 )/ ker(a j ) injectively to ker(a j )/ ker(a j 1 ). Consequently, if define ( ) n j := dim ker(a j+1 )/ ker(a j ), j = 0, 1,, k. then n k n k 1 n 0. In addition, dim(ker(a j+1 )) = n j + dim(a j ), j dim(v ) = dim(ker(a k+1 )) = n k + n k 1 + + n 0. Proof. If x ker(a j+1 ), A j (Ax) = 0, i.e, Ax ker(a j ). Hence, A maps ker(a j+1 ) to ker(a j ). Next, suppose y x mod (ker(a j )). Then writing y = x + z where z ker(a j ), we have Ay Ax = Az ker(a j 1 ), i.e., Ay Ax mod (ker(a j 1 )). Thus is, A maps ker(a j+1 )/ ker(a j ) to ker(a j )/ ker(a j 1 ). Finally, we show that the map is injective. Suppose Ax 0 mod (ker(a j 1 )). Then 0 = A j 1 (Ax) = A j x so that x 0 mod (ker(a j )). Thus, the map is injective. Then Now we construct a basis for V as follows. Select x k 1,, xk n k such that x k i mod (ker(a k )), i = 1,, n k, form a basis for ker(a k+1 )/ ker(a k ). V = ker(a k+1 ) = ( ) n k i=1 span{xk i } ker(a k ). Suppose we have constructed a basis x j i mod (ker(a j )), i = 1,, n j for ker(a j+1 )/ ker(a j ). Set x j 1 i = Ax j i, i = 1,, n j. Then by the lemma, these vectors are linearly independent in ker(a j )/ ker(a j 1 ). Hence, we can expand it to x j 1 i mod (A j 1 ), i = 1,, n j 1 to become a basis for ker(a j )/ ker(a j 1 ). There holds ( ) ker(a j ) = nj 1 i=1 span{xj 1 i } ker(a j 1 ) Recursively finding these bases for j = k, k 1,, 1, we then obtain {x j i 1,, 0} which is a basis for V since ker(a 0 ) = {0}. i = 1,, n j, j = k, k

136 CHAPTER 4. SPECTRAL THEORY Since x j 1 i = Ax j i for i n j, we can list all the elements in the basis of V as follows: x k 1,, xk n k, Ax k 1,, Axk n k, xn k 1 k +1,, xk 1 n k 1, A k x k 1, A k x k n k, A k 1 xn k 1 k +1, Ak 1 x nk 1 x 0 n 1+1,, x 0 n 0 We can order them in the vertical direction. For each i = 1,, n 0, there is an integer k i such that x j i = Aki j x ki i for j = 0,, k i 1 and A ki+1 x ki i = 0. Denoting n 0 = m, and x ki i = x i, we then have the following. Theorem 4.3. Let A be a nilpotent operator mapping a finite dimensional vector space into itself. Then there exists a basis of the form {A k1 x 1, A k1 1 x 1,, x 1, A k2 x 2, A k2 1 x 2,, x 2,, A km x m,, x m } where A ki+1 x i = 0 for i = 1,, m. In particular, under this basis, the matrix A associated with the operator is given by the Jordan form 0 1 0 J k1 0 0 1 A =..., where J k :=...... 0 J km 0 1 0 0 where k = k 1 k 2 k m 1. k k Exercise 24. Show that in finite dimensional case, a linear operator A is nilpotent if and only if σ(a) = {0}. Exercise 25. When the space dimension is 5, list all the possible Jordan matrices for nilpotent operators. Group them according to the similarity of matrices. Exercise 26. Suppose A M 6 (R) satisfies A 2 = 0I. List all possible Jordan normal forms for A. Exercise 27. Let V be the space of all polynomials of degree n with real coefficients. Find the Jordan form for the operator A = d2 dx 2 : f V f V. What is the minimal polynomial of A?

4.7. THE JORDAN FORM 137 4.7 The Jordan form With the spectral theorem and the theory for nilpotent operators, we can now make similarity transformations to associate each square matrix with a canonical form. Definition 4.5. A Jordan block J k (λ) is a k by k matrix of the form 1 0 0 0 1 0 0 1 0 0 1 J k (λ) = λi k + J k, I k =...... J k :=...... 1 0 0 1 0 0 1 0 0 k k k k namely, I k is the identity matrix of rank k, and J k = (δ i+1,j ) k k. When k = 1, J 1 = (0). A Jordan matrix is a direct sum of Jordan blocks: J k1 (λ 1 ) 0 J k2 (λ 2 ) J =... 0 J km (λ m ) in which the values λ i need not be distinct. Given a matrix A M n (F), if J = P 1 AP is a Jordan matrix, then J is also called the Jordan canonical form or Jordan normal form of the matrix A. We remark that one can require the Jordan blocks to be arranged in a special order so that the Jordan canonical form of a square matrix is unique. We shall not do it here. Theorem 4.4. (The Jordan Theorem of Normal Form) Any square matrix on a closed field is similar to a Jordan matrix. For every linear operator from a finite dimensional vector space over a closed field into itself, there exists a basis such that the matrix of the operator under the basis is a Jordan matrix. We leave the proof as an exercise. From the process of finding basis for nilpotent operators, we see that the final canonical form depends only on the dimensional of ker(a j ), j = 1,, or ker((a λi) j ) on E(λ). Indeed, we have the following: Theorem 4.5. Let V be a finite dimensional vector space over a closed field F and A and B be two linear maps from V into itself. Then A is similar to B, i.e., there exists an automorphism P such that B = P 1 AP if and only if ( ) dim ker((λi A) m ) ( ) = dim ker((λ B) m ) λ F, m 1. Instead of giving a proof to the Jordan theorem, we provide here the basic steps in finding the Jordan normal form for given a square matrix A in M n (F). (1) Find the characteristic polynomial p(z) = det(zi A) and all the eigenvalues: λ i with multiplicity m i, i = 1,, ˆn. Then for each i, set B = A λ i I and perform the following step.

138 CHAPTER 4. SPECTRAL THEORY (2)(i) Find N 0, N 1,, N k by solving iteratively the following B 1 (0) = N 0 = ker(b 1 ) = N 0, B 1 (N j 1 ) = N j N 0 = ker(b j+1 ) = N j ker(b j ), BN j N j 1. In finite steps, one finds N k+1 = {0}. This implies that E(λ i ) = j 1 ker(b j ) = ker(b k+1 ) = N k N k 1 N 0 Denote by m i the multiplicity of the eigenvalue λ i and by n j the dimension of N j. Then 0 = n k+1 < n k n k 1 n 0, dim(e(λ i )) = m i = j dim(n j ) = n k + + n 0 (ii) Find a basis U k = {x 1,, x nk } for N k. For each j = k 1, k 2,, 0, generate a basis U j for N j by adding vectors x nj+1,, x nj 1 to an initial collection BU j+1. (iii) Combine U 0,, U k and order them to a basis for E(λ i ) in the following form P i = (B k1 x 1,, x 1,, B kn 0 xn0,, x n0 ). Then AP i = P i J i where J i = diag(j k1,, J kn0 ) consists of n 0 Jordan blocks. (3) Finally combining all the basis P i for E(λ i ), i = 1,, ˆn, we obtain the similarity transform matrix P = (P 1 Pˆn ) and the Jordan form J = diag(j 1,, J ˆn ). Example 4.9. Find the Jordan form and the associated transformation for the matrix 1 0 0 0 0 1 1 0 0 0 A = 1 1 1 0 0 2 2 0 1 0 3 3 0 0 2 Solution. (1) The characteristic polynomial is p(z) = (z 1) 4 (z 2). Hence, λ 1 = 2 is a simple eigenvalue, and λ 2 = 1 is an eigenvalue of multiplicity 4. (2) For λ 1 = 2, we find an eigenvector ξ 1 = (0 0 0 0 1) T. (3) For λ 2 = 1, solving (A I)x = 0 gives us, ξ 2 = (0 0 1 0 0) T, ξ 3 = (0 0 0 1 0) T, and N 0 = ker(a I) = span{ξ 2, ξ 3 }. (4) Solving (A I)x = c 1 ξ 2 + c 2 ξ 3 N 0 gives us, ξ 4 = (0 1 0 0 3) T and N 1 = span{ξ 4 }, ker((a I) 2 ) = N 1 N 0, (A I)ξ 4 = ξ 2 + 2ξ 3. (5) Solving (A I)x = x 4 N 1 gives ξ 5 = (1 1 0 0 3) T and N 2 = span{ξ 5 }, ker((a I) 3 ) = N 2 ker((a I) 2 ), (A I)ξ 5 = ξ 4. Set (6) Now we go back. Set U 2 = {ξ 5 }, U 1 = {(A I)ξ 5 } = {ξ 4 }, U 0 = {(A I)ξ 4, ξ 3 }. Hence, we set u 3 = ξ 5, u 2 = (A I)u 3 = ξ 4, u 1 = (A I) 2 u 3 = ξ 2 + 2ξ 3, (A I)u 1 = 0, u 4 = ξ 3, (A I)u 4 = 0 u 5 = ξ 1, (A 2I)ξ 1 = 0

4.7. THE JORDAN FORM 139 We see that A(u 1, u 2, u 3, u 4, u 5 ) = (u 1 + u 2, u 2 + u 3, u 3, u 4, 2u 5 ). That is, AP = P J or P 1 AP = J where 1 1 0 0 0 0 0 1 0 0 0 1 1 0 0 0 1 1 0 0 P = (u 1 u 2 u 3 u 4 u 5 ) = 1 0 1 0 0 2 0 0 1 0, J = 0 0 1 0 0 0 0 0 1 0 0 3 3 0 1 0 0 0 0 2 Exercise 28. Prove the Jordan theorem of normal form for square matrices. Exercise 29. Prove Theorem 4.5. Also state its analogous for matrices. Exercise 30. Let A M 5 (R). Find possible Jordan forms in the following cases: (i) A is idempotent: A 2 = A; (ii) A is tripotent: A 3 = A; (iii) A k = A for some k 4. Exercise 31. For the following matrices, find P such that P 1 AP is in a Jordan form: 1 0 0 0 0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 1 1 1 0 0 0 1 1 1 0 0 2 2 0 1 0, 1 1 0 0 0 1 1 1 0 0 2 2 0 1 0, 1 1 0 0 0 1 1 1 0 0 2 2 0 1 0, 1 0 0 0 5 0 1 0 0 10 0 0 1 0 10 3 3 0 0 1 3 4 0 0 2 5 5 0 0 2 0 0 0 1 5

140 CHAPTER 4. SPECTRAL THEORY 4.8 Real Jordan Canonical Forms In addition to the Jordan canonical form, there are several other matrix factorizations that can be useful in various circumstances. Here we consider only a variant of the Jordan form when the matrix has only real entries. (Recall that R is not closed.) When A is a real matrix, all the non-real eigenvalues must occur in conjugate pairs. Moreover, ker((a λi) j ) = ker((a λi) j ) λ C, k 1. Hence, the structure of A on E(λ) is the same as that on E( λ) = E(λ). Suppose λ = α + ωi is an eigenvalue (ω > 0) and {u 1,, u k } is a set of (generalized) eigenvectors on which the corresponding map A : x C n Ax C n takes a Jordan block form: (Au 1 Au k ) = λ(u 1 u k ) + (0 u 1 u k 1 ) = (u 1 u k )J k (λ). Write u i = u 1 i + iu2 i. The invariant space span{u 1,, u k } span{ū 1,, ū k ) has a real basis {u 1 1, u2 1,, u1 k, u2 k }. Under this basis, we can calculate, since A is real, (Au 1 i, Au2 i ) = (αu1 i ωu2 i, ωu i + αu 2 i ) + (u1 i 1, u2 i 1 ), i = 1,, k where u 1 0 = u2 0 := 0. In the matrix form, this can be written as A(u 1 1 u2 1 u1 k u2 k ) = (u1 u 2 u 1 k u2 k )C k(α, ω) where C k (α, ω) is the real Jordan block C I 0 C I C k (α, ω) :=...... C I 0 C 2k 2k ( ) α ω, C = C(α, ω) := ω α (4.2) and I is the 2 2 identity matrix. Converting each pair of complex Jordan blocks into a real Jordan block, we conclude as follows: Theorem 4.6. Each real matrix A M n (R) is similar to, via a real similarity transformation P 1 AP, a real Jordan canonical form : C k1 (α 1, ω 1 ) 0... Ckp(αp, ωp) J m1 (λ 1 )... 0 J mq (λ q ) where λ i, i = 1,, q are real eigenvalues, and α j + ω j i, j = 1,, p (ω j > 0) are non-real eigenvalues. Example 4.10. Find the real Jordan canonical form and the similarity transformation for A = Solution. The characteristic polynomial is p(z) = z 2 2z + 10 so eigenvalues are λ = 1 ± 3i. ( ) 0 10. 1 2

4.8. REAL JORDAN CANONICAL FORMS 141 The eigenvector for 1 + 3i is x = (1 3i 1) T. Hence we take ( ( 1 3 u 1 = Re(x) =, u 1) 2 = Im(x) = 0 ), P = (u 1 u 2 ) = ( ) 1 3. 1 0 We find that Au 1 = u 1 3u 2, Au 2 = 3u 1 + u 2, or AP = P J, P 1 AP = J = C 1 (1, 3) = ( 1 ) 3 3 1 Example 4.11. Find the similarity transformation and the real Jordan normal form for 1 1 0 0 A = 1 1 0 0 1 1 1 1 1 1 1 1 Solution. (i) The characteristic polynomial is p(z) = det(zi A) = [(z 1) 2 + 1] 2 ; the eigenvalues are 1 ± i of multiplicity 2. (ii) For λ = 1 + i, solving (A λi)x = 0 gives (iii) Solving (A λi)x = x 1 gives ker(a λi) = span{x 1 }, x 1 = (0 0 1 i) T. ker((a λi) 2 ) = span{x 2, x 1 }, x 2 = (1 i 0 i) T, (A λi)x 2 = x 1, (iv) Extracting the real and imaginary parts, we obtain ( ) P = Re(x 1 ) Im(x 1 ) Re(x 2 ) Im(x 2 ), AP = P J, P 1 AP = J, 0 0 1 0 1 1 1 0 P = 0 0 0 1 1 0 0 0, J = 1 1 0 1 0 0 1 1. 0 1 0 1 0 0 1 1 Exercise 32. Suppose the minimal polynomial of A M n (R) is p = z n z. Find its real Jordan normal form. Exercise 33. Find the similarity transformation and the real Jordan normal form for the following matrices ( ) 1 1 1 1 1 2 1 1 1 5, 1 1 1 1 1 1 0 0 1 1, 2 1 0 1 0 0 1 2 0 0 1 1 0 0 2 1 Exercise 34. Let F = R and V be the set of all polynomials of order < 5. Set A = d2 dx 2 + 2 d dx + I. Find the corresponding matrix for A 10 under the basis {e 0,, e 4 } where e i (x) = x i for i = 0,, 4.

142 CHAPTER 4. SPECTRAL THEORY