October 4, 2017 EIGENVALUES AND EIGENVECTORS. APPLICATIONS

Similar documents
September 26, 2017 EIGENVALUES AND EIGENVECTORS. APPLICATIONS

EIGENVALUES AND EIGENVECTORS 3

m We can similarly replace any pair of complex conjugate eigenvalues with 2 2 real blocks. = R

1. General Vector Spaces

MATH 583A REVIEW SESSION #1

Linear Algebra: Matrix Eigenvalue Problems

Lecture 7: Positive Semidefinite Matrices

A Brief Outline of Math 355

Eigenvalues and Eigenvectors

Math Matrix Algebra

Eigenvalues and Eigenvectors. Review: Invertibility. Eigenvalues and Eigenvectors. The Finite Dimensional Case. January 18, 2018

Eigenvalues and Eigenvectors

Calculating determinants for larger matrices

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors

Study Guide for Linear Algebra Exam 2

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Recall : Eigenvalues and Eigenvectors

Econ Slides from Lecture 7

DM554 Linear and Integer Programming. Lecture 9. Diagonalization. Marco Chiarandini

MATH 423 Linear Algebra II Lecture 20: Geometry of linear transformations. Eigenvalues and eigenvectors. Characteristic polynomial.

Topic 1: Matrix diagonalization

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Chapter 5 Eigenvalues and Eigenvectors

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

MATHEMATICS 217 NOTES

Linear Algebra Primer

Linear Algebra 2 Spectral Notes

Definition (T -invariant subspace) Example. Example

av 1 x 2 + 4y 2 + xy + 4z 2 = 16.

MTH 464: Computational Linear Algebra

Math 113 Homework 5. Bowei Liu, Chao Li. Fall 2013

Math Linear Algebra Final Exam Review Sheet

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

Generalized eigenvector - Wikipedia, the free encyclopedia

Linear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8.

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

21 Linear State-Space Representations

October 25, 2013 INNER PRODUCT SPACES

and let s calculate the image of some vectors under the transformation T.

4. Determinants.

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

1 Invariant subspaces

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

4.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial

7 Planar systems of linear ODE

Eigenvalues and Eigenvectors

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

c c c c c c c c c c a 3x3 matrix C= has a determinant determined by

5 Eigenvalues and Diagonalization

4. Linear transformations as a vector space 17

Section 8.2 : Homogeneous Linear Systems

Chapter 7. Linear Algebra: Matrices, Vectors,

Foundations of Matrix Analysis

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

2 Eigenvectors and Eigenvalues in abstract spaces.

Jordan normal form notes (version date: 11/21/07)

Generalized Eigenvectors and Jordan Form

Chapter 4 & 5: Vector Spaces & Linear Transformations

Eigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems.

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

Lecture Summaries for Linear Algebra M51A

Matrices and Linear Algebra

Solutions to Final Exam

Solutions Problem Set 8 Math 240, Fall

Chapter Two Elements of Linear Algebra

A matrix over a field F is a rectangular array of elements from F. The symbol

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Generalized eigenspaces

Linear Algebra 2 Final Exam, December 7, 2015 SOLUTIONS. a + 2b = x a + 3b = y. This solves to a = 3x 2y, b = y x. Thus

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

6 Inner Product Spaces

235 Final exam review questions

9.1 Eigenanalysis I Eigenanalysis II Advanced Topics in Linear Algebra Kepler s laws

Eigenspaces and Diagonalizable Transformations

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

CS 246 Review of Linear Algebra 01/17/19

Chapters 5 & 6: Theory Review: Solutions Math 308 F Spring 2015

Lecture 10 - Eigenvalues problem

MATH 221, Spring Homework 10 Solutions

Eigenvalues and Eigenvectors A =

Knowledge Discovery and Data Mining 1 (VO) ( )

LINEAR ALGEBRA BOOT CAMP WEEK 2: LINEAR OPERATORS

Math 108b: Notes on the Spectral Theorem

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

Math Ordinary Differential Equations

EE5120 Linear Algebra: Tutorial 6, July-Dec Covers sec 4.2, 5.1, 5.2 of GS

(f + g)(s) = f(s) + g(s) for f, g V, s S (cf)(s) = cf(s) for c F, f V, s S

Eigenvalues, Eigenvectors, and Diagonalization

Matrix Theory. A.Holst, V.Ufnarovski

MIT Final Exam Solutions, Spring 2017

CAAM 335 Matrix Analysis

Solving a system by back-substitution, checking consistency of a system (no rows of the form

Notes on the matrix exponential

vibrations, light transmission, tuning guitar, design buildings and bridges, washing machine, Partial differential problems, water flow,...

Math 489AB Exercises for Chapter 2 Fall Section 2.3

Linear Algebra Primer

Transcription:

October 4, 207 EIGENVALUES AND EIGENVECTORS. APPLICATIONS RODICA D. COSTIN Contents 4. Eigenvalues and Eigenvectors 3 4.. Motivation 3 4.2. Diagonal matrices 3 4.3. Example: solving linear differential equations 3 4.4. Eigenvalues and eigenvectors: definition 4 4.5. The characteristic equation 5 4.6. Geometric interpretation of eigenvalues and eigenvectors 6 4.7. Diagonal matrices 6 4.8. Similar matrices have the same eigenvalues 7 4.9. Projections 7 4.0. Trace, determinant and eigenvalues 8 4.. The eigenvalue zero 9 4.2. Eigenvectors corresponding to different eigenvalues are independent 9 4.3. Diagonalization of matrices with linearly independent eigenvectors 0 4.4. Eigenspaces 4.5. Real matrices with complex eigenvalues; decomplexification 3 4.6. Jordan normal form 5 4.7. Block matrices 8 5. Solving linear equations 2 5.. Solutions of linear differential equations with constant coefficients 2 5.2. The case when M is diagonalizable. 2 5.3. Non-diagonalizable matrix 25 5.4. Fundamental facts on linear differential systems 28 5.5. Eigenvalues and eigenvectors of the exponential of a matrix 29 5.6. Higher order linear differential equations; companion matrix 30 5.7. Systems of second order equations 33 5.8. Stability in differential equations 34 5.9. Difference equations (Discrete dynamical systems) 39 5.0. Solutions of linear difference equations 40 5.. Stability 4 5.2. Example: Fibonacci numbers 4 5.3. Positive matrices 42 5.4. Markov chains 42

2 RODICA D. COSTIN 6. More functional calculus 44 6.. Discrete equations versus differential equations 44 6.2. Functional calculus for digonalizable matrices 45 6.3. Commuting matrices 48

4.. Motivation. EIGENVALUES AND EIGENVECTORS. APPLICATIONS 3 4. Eigenvalues and Eigenvectors 4.2. Diagonal matrices. Perhaps the simplest type of linear transformations are those whose matrix is diagonal (in some basis). Consider for example the matrices [ ] [ ] a 0 b 0 () M =, N = 0 a 2 0 b 2 It can be easily checked that and M = αm + βn = [ a 0 0 a 2 [ ] αa + βb 0 0 αa 2 + βb 2 ] [, M k a k = 0 0 a k 2 ] [ ] a b, MN = 0 0 a 2 b 2 Diagonal matrices behave like the bunch of numbers on their diagonal! The linear transformation consisting of multiplication by the matrix M in () dialates a times vectors in the direction of e and a 2 times vectors in the direction of e 2. In this chapter we will see that most linear transformations do have diagonal matrices in a special basis, whose elements are called the eigenvectors of the transformation. We will learn how to find these bases. Along the special directions of the eigenvectors the transformation just dialation by a factor, called eigenvalue. 4.3. Example: solving linear differential equations. Consider the simple equation du dt = λu which is linear, homogeneous, with constant coefficients, and unknown function u(t) R (or in C). Its general solution is, as it is well known, u(t) = Ce λt. Consider now a similar equation, but where the unknown u(t) is a vector valued function: (2) du dt = Mu, u(t) Rn, M is an n n constant matrix Inspired by the one dimensional case we look for exponential solutions. Substituting in (2) u(t) = e λt v (where λ is a number and v is a constant vector, both be determined) and dividing by e λt, we obtain that the scalar λ and the vector v must satisfy (3) λv = Mv or (4) (M λi)v = 0

4 RODICA D. COSTIN If the null space of the matrix M λi is zero, then the only solution of (4) is v = 0 which gives the (trivial!) solution u(t) 0. If however, we can find special values of λ for which N (M λi) is not null, then we found a nontrivial solution of (2). Such values of λ are called eigenvalues of the matrix M, and vectors v N (M λi), v 0, are called eigenvectors corresponding to the eigenvalue λ. Of course, the necessary and sufficient condition for N (M λi) {0} is that (5) det(m λi) = 0 Example. Let us calculate the exponential solutions for [ ] 3 (6) M = 0 2 Looking for eigenvalues of M we solve equation (5), which for (6) is ([ ] [ ]) 3 0 λ 3 det λ = = ( λ) (2 λ) 0 2 0 0 2 λ with solutions λ = and λ 2 = 2. We next determine an eigenvector corresponding to the eigenvalue λ = λ = : looking for a nozero vector v such that (M λ I)v = 0 we solve [ ] [ ] [ ] 0 3 x 0 = 0 3 0 giving x 2 = 0 and x arbitrary; therefore the first eigenvector is any scalar multiple of v = (0, ) T. Similarly, for the eigenvalue λ = λ 2 = 2 we solve (M λ 2 I)v 2 = 0: [ ] [ ] [ ] 3 3 y 0 = 0 0 0 which gives y 2 = y and y arbitrary, and the second eigenvector is any scalar multiple of v 2 = (, ) T. We found two particular solutions of (2), (6), namely u (t) = e t (0, ) T and u 2 (t) = e 2t (, ) T. These are functions belonging to the null space of the linear operator Lu = du dx Mu, therefore any linear combination of these two solutions also belongs to the null space: any C u (t) + C 2 u 2 (t) is also a solution, for and constants C, C 2. A bit later we will show that these are all the solutions. 4.4. Eigenvalues and eigenvectors: definition. Denote the set of n n (square) matrices with entries in F (= R or C) x 2 M n (F ) = {M M = [M ij ] i,j=,...n, M ij F } A matrix M M n (F ) defines an endomorphism the vector space F n (over the scalars F ) by usual multiplication x Mx. y 2

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 5 Note that a matrix with real entries can also act on C n, since for any x C n also Mx C n. But a matrix with complex non real entries cannot act on R n, since for x R n the image Mx may not belong to R n (while certainly Mx C n ). Definition. Let M be an n n matrix acting on the vector space V = F n. A scalar λ F is an eigenvalue of M if for some nonzero vector v V, v 0 we have (7) Mv = λv The vector v is called eigenvector corresponding to the eigenvalue λ. Of course, if v is an eigenvector corresponding to λ, then so is any scalar multiple cv (for c 0). 4.5. The characteristic equation. Equation (7) can be rewritten as M v λv = 0, or (M λi)v = 0, which means that the nonzero vector v belongs to the null space of the matrix M λi, and in particular this matrix is not invertible. Using the theory of matrices, we know that this is equivalent to det(m λi) = 0 The determinant has the form M λ M 2... M n M 2 M 22 λ... M 2n det(m λi) =... M n M n2... M nn λ This is a polynomial in λ, having degree n. To understand why this is the case, consider first n = 2 and n = 3. For n = 2 the characteristic polynomial is M λ M 2 M 22 λ = (M λ) (M 22 λ) M 2 M 2 M 2 = λ 2 (M + M 22 )λ + (M M 22 M 2 M 2 ) which is a quadratic polynomial in λ; the dominant coefficient is. For n = 3 the characteristic polynomial is M λ M 2 M 3 M 2 M 22 λ M 23 M 3 M 23 M 33 λ and expanding along say, row, = ( ) + (M λ) M 22 λ M 23 M M 23 M 33 λ +( )+2 M 2 M 23 2 M 3 M 33 λ +( ) +3 M M 2 M 22 λ 3 M 3 M 23 = λ 3 + (M + M 22 + M 33 )λ 2 +...

6 RODICA D. COSTIN which is a cubic polynomial in λ; the dominant coefficient is. It is easy to show by induction that det(m λi) is polynomial in λ, having degree n, and that the coefficient of λ n is ( ) n. Definition 2. The polynomial det(m λi) is called the characteristic polynomial of the matrix M, and the equation det(m λi) = 0 is called the characteristic equation of M. Remark. Some authors refer to the characteristic polynomial as det(λi M); the two polynomial are either equal or a multiple of each other, since det(λi M) = ( ) n det(m λi). 4.6. Geometric interpretation of eigenvalues and eigenvectors. Let M be an n n matrix, and T : R n R n defined by T (x) = Mx be the corresponding linear transformation. If v is an eigenvector corresponding to an eigenvalue λ of M: Mv = λv, then T expands or contracts v (and any vector in its direction) λ times (and it does not change its direction!). If the eigenvalue/vector are not real, a similar fact is true, only that multiplication by a complex (not real) scalar cannot be easily called an expansion or a contraction (there is no ordering in complex numbers), see the example of rotations, 4.5.. The special directions of the eigenvectors are called principal axes of the linear transformation (or of the matrix). 4.7. Diagonal matrices. Let D be a diagonal matrix: d 0... 0 0 d 2... 0 (8) D =... 0 0... d n To find its eigenvalues, calculate d λ 0... 0 0 d 2 λ... 0 det(d λi) = = (d λ )(d 2 λ 2 )... (d n λ n )... 0 0... d n λ The eigenvalues are precisely the diagonal elements, and the eigenvector corresponding to d j is e j (as it is easy to check). The principal axes of diagonal matrices the coordinate axes. Vectors in the direction of one of these axes preserve their direction and are stretched or compressed: if x = ce k then Dx = d k x. Diagonal matrices are easy to work with: what was noted for the 2 2 matrices in 4. is true in general, and one can easily check that any power D k is the diagonal matrix having d k j on the diagonal.

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 7 If p(x) is a polynomial p(t) = a k t k + a k t k +... + a t + a 0 then for any square matrix M one can define p(m) as (9) p(m) = a k M k + a k M k +... + a M + a 0 I If D is a diagonal matrix (8) then p(d) is the diagonal matrix having p(d j ) on the diagonal. (Check!) Diagonal matrices can be viewed as the collection of their eigenvalues! Exercise. Show that the eigenvalues of an upper (or lower) triangular matrix are the elements on the diagonal. 4.8. Similar matrices have the same eigenvalues. It is very easy to work with diagonal matrices and a natural question arises: which linear transformations have a diagonal matrix in a well chosen basis? This is the main topic we will be exploring for many sections to come. Recall that if the matrix M represents the linear transformation L : V V in some basis B V of V, and the matrix M represents the same linear transformation L, only in a different basis B V, then the two matrices are similar: M = S MS (where S the the matrix of change of basis). Eigenvalues are associated to the linear transformation (rather than its matrix representation): Proposition 3. Two similar matrices have the same eigenvalues: if M, M, S are n n matrices, and M = S MS then the eigenvalues of M and of M are the same. This is very easy to check, since suppose λ is an eigenvalue of M: Mv = λv for some v 0 Then S Mv = λs v, so (S MS)S v = λs v which means that S v is an eigenvector of S MS corresponding to the same eigenvalue λ. 4.9. Projections. Recall that projections do satisfy P 2 = P (we saw this for projections in dimension two, and we will prove it in general). Proposition 4. Let P be a square matrix satisfying P 2 = P. Then the eigenvalues of P can only be 0 or. Proof. Let λ be an eigenvalue; this means that there is a nonzero vector v so that P v = λv. Applying P to both sides of the equality we obtain P 2 v = P (λv) = λp v = λ 2 v. Using the fact that P 2 v = P v = λv it follows that λv = λ 2 v so (λ λ 2 )v = 0 and since v 0 then λ λ 2 = 0 so λ {0, }. Example. Consider the projection of R 3 onto the x x 2 plane. Its matrix P = 0 0 0 0 0 0 0

8 RODICA D. COSTIN is diagonal, with eigenvalues,, 0. 4.0. Trace, determinant and eigenvalues. Definition 5. Let M be an n n matrix, M = [M ij ] i,j=,...,n. The trace of M is the sum of its elements on the principal diagonal: n Tr M = The following theorem shows that what we noticed in 4.5 for n = 2 is true for any n: Theorem 6. Let M be an n n matrix, an let λ,..., λ n be its n eigenvalues (complex, not necessarily distinct). Then j= M jj (0) det M = λ λ 2... λ n and () Tr M = λ + λ 2 +... + λ n In particular, the traces of similar matrices are equal, and so are their determinants. Sketch of the proof. The coefficients of any polynomial can be written in terms of its roots, and, in particular, it is not hard to see that (2) p(x) (x λ )(x λ 2 )... (x λ n ) = x n (λ + λ 2 +... + λ n )x n +... + ( ) n (λ λ 2... λ n ) In particular, p(0) = ( ) n λ λ 2... λ n. The characteristic polynomial factors as det(m λi) = ( ) n (λ λ )... (λ λ n ) ( ) n p(λ) (recall that the dominant coefficient of the characteristic polynomial is ( ) n ) and (0) follows. To show () we expand the determinant det(m λi) using minors and cofactors keeping track of the coefficient of λ n. As seen on the examples in 4.5, only the first term in the expansion contains the power λ n, and continuing to expand to lower and lower dimensional determinants, we see that the only term containing λ n is (M λ)(m 22 λ)... (M nn λ) = ( ) n λ n ( ) n (M + M 22 +... + M nn )λ n + lower powers of λ which compared to (2) gives (). These are called Vieta s formulas.

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 9 4.. The eigenvalue zero. As an immediate consequence of Theorem 6, we can recognize invertible matrices by looking at their eigenvalues: Corollary 7. A matrix M is invertible if and only if all its eigenvalues are nonzero. Note that a matrix M has an eigenvalue equal to zero if and only if its null space N (M) is nontrivial. Moreover, the matrix M has dim N (M) eigenvectors linearly independent which correspond to the eigenvalue zero. 4.2. Eigenvectors corresponding to different eigenvalues are independent. Theorem 8. Let M be an n n matrix. Let λ,... λ k a set of distinct eigenvalues of M and v,..., v k be corresponding eigenvectors. Then the set v,..., v k is linearly independent. In particular, if M has entries in F = R or C, and all the eigenvalues of M are in F and are distinct, then the set of corresponding eigenvectors form a basis for F n. Proof. Assume, to obtain a contradiction, that the eigenvectors are linearly dependent: there are c,..., c k F not all zero such that (3) c v +... + c k v k = 0 Step I. We can assume that all c j are not zero, otherwise we just remove those v j from (3) and we have a similar equation with a smaller k. If after this procedure we are left with k =, then this implies c v = 0 which contradicts the fact that not all c j are zero or the fact that v 0. Otherwise, for k 2 we continue as follows. Step II. Then we can solve (3) for v k : (4) v k = c v +... + c k v k where c j = c j/c k. Applying M to both sides of (4) we obtain (5) λ k v k = c λ v +... + c k λ k v k Multiplying (4) by λ k and subtracting from (5) we obtain (6) 0 = c (λ λ k )v +... + c k (λ k λ k )v k Note that all c j (λ j λ k ) are non-zero (since all c are non-zero, and λ j λ k ). If k=2, then this implies v = 0 which is a contradiction. If k > 2 we go to Step I. with a lower k. The procedure decreases k, therefore it must end, and we have a contradiction.

0 RODICA D. COSTIN 4.3. Diagonalization of matrices with linearly independent eigenvectors. Suppose that the M be an n n matrix has n independent eigenvectors v,..., v n. Note that, by Theorem 8, this is the case if we work in F = C and all the eigenvalues are distinct (recall that this happens with probability one). Also this is the case if we work in F = R and all the eigenvalues are real and distinct. Let S be the matrix with columns v,..., v n : S = [v,..., v n ] which is invertible, since v,..., v n are linearly independent. Se k = v k for all k =,..., n. Since Mv j = λ j v j then (7) MS = M[v,..., v n ] = [Mv,..., Mv n ] = [λ v,..., λ n v n ] Note that To identify the matrix on the right side of (7) note that S(λ j e j ) = λ j v j so λ 0... 0 0 λ 2... 0 [λ v,..., λ n v n ] = S[λ e,..., λ n e n ] = SΛ, where Λ =... 0 0... λ n Relation (7) is therefore MS = SΛ, or S MS = Λ = diagonal Note that the matrix S which diagonalizes a matrix is not unique. For example, we can replace any eigenvector by a scalar multiple of it. Also, we can use different orders for the eigenvectors (this will result on a diagonal matrix with the same values on the diagonal, but in different positions). Example. Consider the matrix (6) for which we found the eigenvalues λ = and λ 2 = 2 and the corresponding eigenvectors v = (0, ) T. v 2 = (, ) T. Taking [ ] 0 S = we have S MS = [ 0 0 2 Not all matrices are diagonalizable, certainly those with distinct eigenvalues are, and some matrices with multiple eigenvalues. Example 2. The matrix [ ] 0 (8) N = 0 0 has eigenvalues λ = λ 2 = 0 but only one (up to a scalar multiple) eigenvector v = e. ]

EIGENVALUES AND EIGENVECTORS. APPLICATIONS Multiple eigenvalues are not guaranteed to have an equal number of independent eigenvectors! N is not diagonalizable. Indeed, assume the contrary, to arrive at a contradiction. Suppose there exists an invertible matrix S so that S NS = Λ where Λ is diagonal, hence it has the eigenvalues of N on its diagonal, and therefore it is the zero matrix: S NS = 0, which multiplied by S to the left and S to the right gives N = 0, which is a contradiction. Some matrices with multiple eigenvalues may still be diagonalized; next section explores when this is the case. 4.4. Eigenspaces. Consider an n n matrix M with entries in F, with eigenvalues λ,..., λ n in F. Definition 9. The set V λj = {x F n Mx = λ j x} is called the eigenspace of M associated to the eigenvalue λ j. Exercise. Show that V λj is the null space of the transformation M λ j I and that V λj is a subspace of F n. Note that all the nonzero vectors in V λj are eigenvectors of M corresponding to the eigenvalues λ j. Definition 0. A subspace V is called an invariant subspace for M if M(V ) V (which means that if x V then Mx V ). The following Remark gathers important features of eigenspaces: it shows that diagonalizable matrices are precisely those for which the dimension of each eigenspace coincides with the multiplicity of the corresponding eigenvalue. (One implication is obvious, which one?) Remark.. Each V λj is an invariant subspace for M. 2. V λj V λl = {0} if λ j λ l. 3. Denote by λ,..., λ k the distinct eigenvalues of M and by r j the multiplicity of the eigenvalue λ j, for each j =,..., k; it is clear that k det(m λi) = (λ j λ) r j and r +... + r k = n j= Then dim V λj r j 4. M is diagonalizable in F n if and only if dim V λj = r j for all j =,..., k and then V λ... V λk = F n Proof. Showing of. and 2. is left to the reader.

2 RODICA D. COSTIN The idea in proving 3. is: since eigenspaces are invariant, we could diagonalize M only one eigenspace, say V λ, obtaining, in the characteristic polynomial, a factor (λ λ ) dim V λ. Let us transform is idea into a proof. Suppose, for some eigenvalue, say λ, of multiplicity r, we have dim V λ = d > r (to arrive to a contradiction). Let v,... v d be a basis for V λ, which we complete to a basis of F n : v,... v d, w,..., w n d. Consider the matrix with these columns, S = [v,... v d, w,..., w n d ] and attempt the calculation at the beginning of 4.3: (9) MS = M[v,... v d, w,..., w n d ] = [Mv,... Mv d, Mw,..., Mw n d ] [λ v,... λ v d, Mw,..., Mw n d ] = [λ Se,... λ Se d, Mw,..., Mw n d ] = S[λ e,... λ e d, S Mw,..., S Mw n d ] := S M Therefore S MS = M, so M and M have the same eigenvalues. Expanding det( M λi) along the first row, then second etc. we see that the characteristic polynomial of M has a factor (λ λ ) d, which is a contradiction, since d > r! The statement 4. easily follows from the preceding ones. Example. Consider the matrix 2 0 0 (20) M := 0 Its characteristic polynomials is 2 det(m λi) = λ 3 + 3 λ 2 4 = (λ + ) (λ 2) 2 so λ = and λ 2 = λ 3 = 2 is a double eigenvalue. The eigenspace V λ is one dimensional, spanned by an eigenvector, which, after a simple calculation turns out to be v = (0,, ) T. If the eigenspace V λ2 is two-dimensional (which is not guaranteed) then the matrix M is diagonalizable. A simple calculation shows that there are two independent eigenvectors corresponding to the eigenvalue λ 2 = 2, for example v 2 = (, 0, ) T and v 3 = (2,, 0) T (the null space of M λ 2 I is two-dimensional). Let 0 2 S = [v, v 2, v 3 ] = 0 then S MS = 0 0 0 0 2 0 0 0 2

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 3 4.5. Real matrices with complex eigenvalues; decomplexification. 4.5.. Complex eigenvalues of real matrices. For an n n matrix with real entries, if we want to have n guaranteed eigenvalues, then we have to accept working in C n. Otherwise, if we want to restrict ourselves to working only with real vectors, then we have to accept that we may have fewer (real) eigenvalues, or perhaps none. Complex eigenvalues of real matrices come in pairs: if λ is an eigenvalue of M, then so is its complex conjugate λ (since the characteristic equation has real coefficients). Also, if v is an eigenvector corresponding to the eigenvalue λ, then v is eigenvector corresponding to the eigenvalue λ (check!). The real and imaginary parts of v span a plane where the linear transformation acts by rotation, and a possible dilation. Simple examples are shown below. Example : rotation in the xy-plane. Consider a rotation matrix [ ] cos θ sin θ (2) R θ = sin θ cos θ To find its eigenvalues calculate det(r θ λi) = cos θ λ sin θ sin θ cos θ λ = (cos θ λ)2 +sin 2 θ = λ 2 2 cos θ+ hence the solutions of the characteristic equations det(r θ λi) = 0 are λ,2 = cos θ ± i sin θ = e ±iθ. It is easy to see that v = (i, ) T is the eigenvector corresponding to λ = e iθ and v 2 = ( i, ) T is the eigenvector corresponding to λ 2 = e iθ. Example 2: complex eigenvalues in R 3. Consider the matrix M = Its characteristic polynomial is 2 3 5 2 3 0 2 3 + 2 3 0 0 0 4 det(m λi) = λ 3 2 λ 2 + 4 λ 6 = (λ + 4) ( λ 2 2 λ + 4 ) and its eigenvalues are: λ,2 = ± i 3 = 2e ±iπ/3 and λ 3 = 4, and corresponding eigenvectors v,2 = ( ± 2i,, 0) T, and v 3 = e 3. The matrix S = [v, v 2, v 3 ] diagonalized the matrix: S MS is the diagonal matrix, having the eigenvalues on the diagonal, but all these are complex matrices. To understand how the matrix acts on R 3, we consider the real and imaginary parts of v : let x = Rv = 2 (v + v 2 ) = (,, 0) T and y = Iv = 2i (v v 2 ) = (2, 0, 0) T. Since the eigenspaces are invariant under M, then so is Sp(x, y ), over the complex and even over the real numbers (since M has real elements). The span over the real numbers is the xy-plane, and it is invariant under M. The figure shows the image of the unit circle in the xy-plane under the matrix M: it is an ellipse.

4 RODICA D. COSTIN Figure. The image of the unit circle in the xy-plane. Along the direction of the third eigenvector (the z-axis) the matrix multiples any c e 3 by 4. In the basis x, y, v 3 the matrix of the linear transformation has its simplest form: using S R = [x, y, v 3 ] we obtain the matrix of the transformation in this new basis as 3 0 S R MS R = 3 0 0 0 4 and the upper 2 2 block represents the rotation and dilation 2R π/3. 4.5.2. Decomplexification. Suppose the n n matrix M has real elements, eigenvalues λ,..., λ n and n independent eigenvectors v,..., v n. Then M is diagonalizable: if S = [v,..., v n ] then S MS = Λ where Λ is a diagonal matrix with λ,..., λ n on its diagonal. Suppose that some eigenvalues are not real. Then the matrices S and Λ are not real either, and the diagonalization of M must be done in C n. Suppose that we want to work in R n only. Recall that the nonreal eigenvalues and eigenvectors of real matrices come in pairs of complex-conjugate ones. In the complex diagonal form Λ one can replace diagonal 2 2 blocks [ ] λj 0 0 λ j by a 2 2 matrix which is not diagonal, but has real entries. To see how this is done, suppose λ C \ R and λ 2 = λ, v 2 = v. Splitting into real and imaginary parts, write λ = α +iβ and v = x +iy. Then from M(x + iy ) = (α + iβ )(x + iy ) identifying the real and imaginary parts, we obtain Mx + imy = (α x β y) + i(β x + α y)

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 5 In the matrix S = [v, v 2,..., v n ] composed of independent eigenvectors of M, replace the first two columns v, v 2 = v by x, y (which are vectors in R n ): using the matrix S = [x, y, v 3,..., v n ] instead of S we have M S = S Λ where α β 0... 0 β α 0... 0 Λ = 0 0 λ 3... 0.... 0 0 0... λ m We can similarly replace any pair of complex conjugate eigenvalues with 2 2 real blocks. Exercise. Show that each 2 2 real block obtained through decomplexification has the form [ ] α β = ρ R β α θ for a suitable ρ > 0 and R θ rotation matrix (2). 4.6. Jordan normal form. We noted in 4.4 that a matrix is similar to a diagonal matrix if and only if the dimension of each eigenspace V λj equals the order of multiplicity of the eigenvalue λ j. Otherwise, there are fewer than n independent eigenvectors; such a matrix is called defective. 4.6.. Jordan blocks. Defective matrices can not be diagonalized, but we will see that they are similar to block diagonal matrices, called Jordan normal forms; these are upper triangular, have the eigenvalues on the diagonal, in selected placed above the diagonal, and zero in the rest. After that, in section 4.6.3 it is shown how to construct the transition matrix S, which conjugates a defective matrix to its Jordan form; its columns are made of generalized eigenvectors. The Jordan blocks which appear on the diagonal of a Jordan normal form are as follows. Jordan blocks are just [λ]. 2 2 Jordan blocks have the form [ ] λ (22) J 2 (λ) = 0 λ For example, the matrix (8) is a Jordan block J 2 (0). 3 3 Jordan blocks have the form λ 0 (23) J 3 (λ) = 0 λ 0 0 λ In general, a k k Jordan block, J k (λ), is a matrix having the same number, λ, on the diagonal, above the diagonal and 0 everywhere else.

6 RODICA D. COSTIN Note that Jordan blocks J k (λ) have the eigenvalue λ with multiplicity k, and the dimension of the eigenspace is one. Example of a matrix in Jordan normal form: 3 0 0 0 0 0 2 0 0 0 0 0 2 0 0 0 0 2 0 0 0 0 2 which is block-diagonal, having two Jordan blocks and one 3 3 Jordan block along its diagonal. The eigenvalue 3 is simple, while 2 has multiplicity four. The eigenspace corresponding to 2 is two-dimensional (e 2 and e 3 are eigenvectors). Note how Jordan blocks act on the vectors of the basis. For (22): J 2 (λ)e = λe, so e is an eigenvector. Also (24) J 2 (λ)e 2 = e + λe 2 which implies that (J 2 (λ) λi) 2 e 2 = (J 2 (λ) λi)e = 0. Similarly, for (30): J 3 (λ)e = λe so e is an eigenvector. Then (25) J 3 (λ)e 2 = e + λe 2 implying that (J 3 (λ) λi) 2 e 2 = (J 3 (λ) λi)e = 0. Finally, (26) J 3 (λ)e 3 = e 2 + λe 3 implying that (J 3 (λ) λi) 3 e 3 = (J 3 (λ) λi) 2 e 2 = 0. This illuminates the idea behind the notion of generalized eigenvectors defined in the next section. 4.6.2. The generalized eigenspace. Defective matrices are similar to a matrix which is block-diagonal, having Jordan blocks on its diagonal. An appropriate basis is formed using generalized eigenvectors: Definition. A generalized eigenvector of M corresponding to the eigenvalue λ is a vector x 0 so that (27) (M λi) k x = 0 for some positive integer k. Examples. ) Eigenvectors are generalized eigenvectors (take k = in (27)). 2) Vectors in the standard basis are generalized eigenvectors for Jordan blocks.

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 7 Definition 2. The generalized eigenspace of M corresponding to the eigenvalue λ is the subspace E λ = {x (M λi) k x = 0 for some k Z + } Sometimes we want to refer to only at the distinct eigenvalues of a matrix, this set is called the spectrum : Definition 3. The spectrum σ(m) of a matrix M is the set of its eigenvalues. Theorem 4. For any n n matrix M the following hold: (i) V λ E λ ; (ii) E λ is a subspace; (iii) E λ is an invariant subspace under M; (iv) E λ E λ2 = 0 for λ λ 2. (v) dim E λ =the multiplicity of λ. (vi)the set of eigenvectors and generalized eigenvectors of M span the whole space C n : λ σ(m) E λ = C n The proofs of (i)-(iv) are simple exercises. The proofs of (v), (vi) are not included here. 4.6.3. How to find a basis for each E λ that can be used to conjugate a matrix to a Jordan normal form. Example. The matrix [ ] + a (28) M = a is defective: it has eigenvalues a, a and only one independent eigenvector, (, ) T. It is therefore similar to J 2 (a). To find a basis x, x 2 in which the matrix takes this form, let x = (, ) T (the eigenvector); to find x 2 we solve (M ai)x 2 = x (as seen in (24) and in (25)). The solutions are x 2 (, 0) T + N (M ai), and any vector in this space works, for example x 2 = (, 0) T. For [ ] (29) S = [x, x 2 ] = 0 we have S MS = J 2 (a). Example 2. The matrix M = 2 3 2 0 3

8 RODICA D. COSTIN has eigenvalues 2, 2, 2 and only one independent eigenvector v = (,, ) T. Let x = v = (,, ) T. Solving (M 2I)x 2 = x we obtain x 2 = (,, 0) T (plus any vector in N (M 2I) = V λ ). Next solve (M 2I)x 3 = x 2 which gives x 3 = (0,, ) T (plus any vector in the null space of M 2I). For S = [x, x 2, x 3 ] we have (30) S MS = 2 0 0 2 0 0 2 In general, if λ is an eigenvalue of M for which dimv λ is less than the multiplicity of λ, we do the following. Choose a basis for V λ. For each eigenvector v in this basis set x = v and solve recursively (3) (M λi)x k+ = x k, k =, 2,... Note that each x satisfies (27) for k =, x 2 satisfies (27) for k = 2, etc. At some step k the system (M λi)x k + = x k will have no solution; we found the generalized eigenvectors x,..., x k (which will give a k k Jordan block). We then repeat the procedure for a different eigenvector in the chosen basis for V λ, and obtain a new set of generalized eigenvectors, corresponding to a new Jordan block. Note: Jordan form is not unique. 4.6.4. Real Jordan normal form. If a real matrix has multiple complex eigenvalues and is defective, then its Jordan form can be replaced with an upper block diagonal matrix in a way similar to the diagonal case illustrated in 4.5.2, by replacing the generalized eigenvectors with their real and imaginary parts. For example, a real matrix which can be brought to the complex Jordan normal form α + iβ 0 0 0 α + iβ 0 0 0 0 α iβ 0 0 0 α iβ can be conjugated (by a real matrix) to the real matrix α β 0 β α 0 0 0 α β 0 0 β α 4.7. Block matrices.

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 9 4.7.. Multiplication of block matrices. It is sometimes convenient to work with matrices split in blocks. We have already used this when we wrote M[v,..., v n ] = [Mv,..., Mv n ] More generally, if we have two matrices M, P with dimensions that allow for multiplication (i.e. the number of columns of M equals the number of rows of P ) and they are split into blocks: M = M M 2 M 2 M 22, P = P P 2 P 2 P 22 then MP = M P + M 2 P 2 M P 2 + M 2 P 22 M 2 P + M 22 P 2 M 2 P 2 + M 22 P 22 if the number of columns of M equals the number of rows of P. Exercise. Prove that the block multiplication formula is correct. More generally, one may split the matrices M and P into many blocks, so that the number of block-columns of M equal the number of block-rows of P and so that all products M jk P kl make sense. Then MP can be calculated using blocks by a formula similar to that using matrix elements. In particular, if M, P are block diagonal matrices, having the blocks M jj, P jj on the diagonal, then MP is a block diagonal matrix, having the blocks M jj P jj along the diagonal. For example, if M is a matrix in Jordan normal form, then it is block diagonal, with Jordan blocks M jj along the diagonal. Then the matrix M 2 is block diagonal, having Mjj 2 along the diagonal, and all powers M k are block diagonal, having Mjj k along the diagonal. Furthermore, any linear combination of these powers of M, say c M +c 2 M 2 is block diagonal, having the corresponding c M jj + c 2 Mjj 2 along the diagonal. 4.7.2. Determinant of block matrices. Proposition 5. Let M be a square matrix, having a triangular block form: [ ] [ ] A B A 0 M = or M = 0 D C D where A and D are square matrices, say A is k k and D is l l. Then det M = det A det D. Moreover, if a,..., a k are the eigenvalues of A, and d,..., d l are the eigenvalues of D, then the eigenvalues of M are a,..., a k, d,..., d l. The proof is left to the reader as an exercise. 2 2 Hint: bring A, D to Jordan normal form, then M to an upper triangular form.

20 RODICA D. COSTIN For a more general 2 2 block matrix, with D invertible 3 [ ] A B M = C D the identity [ A C ] [ B D I 0 D C I ] = [ A BD C B 0 D together with Proposition 5 implies that [ ] A B det = det(a BD C D C) det D = det(ad BD CD) For larger number of blocks, there are more complicated formulas. ] 3 References: J.R. Silvester, Determinants of block matrices, Math. Gaz., 84(50) (2000), pp. 460-467, and P.D. Powell, Calculating Determinants of Block Matrices, http://arxiv.org/pdf/2.4379v.pdf

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 2 5. Solving linear equations 5.. Solutions of linear differential equations with constant coefficients. In 4.3 we saw an example which motivated the notions of eigenvalues and eigenvectors. General linear first order systems of differential equations with constant coefficients can be solved in a quite similar way. Consider du (32) dt = Mu where M is an m m constant matrix and u in an m-dimensional vector. As in 4.3, it is easy to check that u(t) = e λt v is a solution of (32) if λ is an eigenvalue of M, and v is a corresponding eigenvector. The goal is to find the solution to any initial value problem: find the solution of (32) satisfying (33) u(0) = u 0 for any given vector u 0. 5.2. The case when M is diagonalizable. Assume that M has m independent eigenvectors v,..., v m, corresponding to the eigenvalues λ,..., λ m. Then (32) has the solutions u j (t) = e λ jt v j for each j =,..., m. These solutions are linearly independent. Indeed, assume that for some constants c,..., c m we have c u (t) +... + c m u m (t) = 0 for all t. Then, in particular, for t = 0 it follows that c v +... + c m v m = 0 which implies that all c j are zero (since v,..., v m were assumed independent). 5.2.. Fundamental matrix solution. Since equation (32) is linear, then any linear combination of solutions is again a solution: (34) u(t) = a u (t) +... + a m u m (t) The matrix = a e λ t v +... + a m e λmt v m, a j arbitrary constants (35) U(t) = [u (t),..., u m (t)] is called a fundamental matrix solution. Formula (34) can be written more compactly as (36) u(t) = U(t)a, where a = (a,..., a m ) T The initial condition (33) determines the constants a, since (33) implies U(0)a = u 0. Noting that U(0) = [v,..., v m ] = S therefore a = S u 0 and the initial value problem (32), (33) has the solution u(t) = U(t)S u 0 General results in the theory of differential equations (on existence and uniqueness of solutions to initial value problems) show that this is only one solution. In conclusion:

22 RODICA D. COSTIN Proposition 6. If the m m constant matrix M has has m independent eigenvectors v,..., v m, corresponding to the eigenvalues λ,..., λ m, then equation (32) has m linearly independent solutions u j (t) = e λ jt v j, j =,..., m and any solution of (32) is a linear combination of them. (37) Example. Solve the initial value problem dx dt = x 2y, = 2x + y, y(0) dy dt x(0) = α = β Denoting u = (x, y) T, problem (37) is [ ] [ ] du 2 α (38) = Mu, where M =, with u(0) = dt 2 β Calculating the eigenvalues of M, we obtain λ =, λ 2 = 3, and corresponding eigenvectors v = (, ) T, v 2 = (, ) T. There are two independent solutions of the differential system: [ ] [ ] u (t) = e t, u 2 (t) = e 3t and a fundamental matrix solution is (39) U(t) = [u (t), u 2 (t)] = [ e t e 3t The general solution is a linear combination of the two independent solutions [ ] [ ] [ ] u(t) = a e t + a 2 e 3t a = U(t) a 2 e t This solution satisfies the initial condition if [ ] [ ] [ ] α a + a 2 = β which is solved for a, a 2 : from [ it follows that [ ] a = a 2 therefore [ ] [ α β ] [ ] [ ] a α = a 2 β ] = (40) u(t) = α+β 2 e t [ so [ 2 2 2 2 e 3t ] [ α β ] + α+β 2 e 3t [ x(t) = α+β 2 e t α+β 2 e 3t y(t) = α+β 2 e t + α+β 2 e 3t ] ] = ] [ α+β 2 α+β 2 ]

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 23 5.2.2. The matrix e Mt. It is often preferable to work with a matrix of independent solutions U(t) rather than with a set of independent solutions. Note that the m m matrix U(t) satisfies d (4) U(t) = M U(t) dt In dimension one this equation reads du dt = λu having its general solution u(t) = Ce λt. Let us check this fact based on the fact that the exponential is the sum of its Taylor series: e x = +! x + 2! x2 +... + n! xn +... = where the series converges for all x C. Then e λt = +! λt + 2! λ2 t 2 +... + n! λn x n +... = n=0 n! xn n=0 n! λn t n and the series can be differentiated term-by-term, giving d dt eλt = d dt n! λn t n = d n! λn dt tn = (n )! λn t n = λe λt n=0 n=0 Perhaps one can define, similarly, the exponential of a matrix and obtain solutions to (4)? For any square matrix M, one can define polynomials, as in (9), and it is natural to define n= (42) e tm = +! tm + 2! t2 M 2 +... + n! tn M n +... = n=0 n! tn M n provided that the series converges. If, furthermore, the series can differentiated term by term, then this matrix is a solution of (4) since (43) d dt etm = d dt n=0 n! tn M n = n=0 n! d dt tn M n = n= n n! tn M n = Me tm Convergence and term-by-term differentiation can be justified by diagonalizing M. Let v,..., v m be independent eigenvectors corresponding to the eigenvalues λ,..., λ m of M, let S = [v,..., v m ]. Then M = SΛS with Λ the diagonal matrix with entries λ,..., λ m. Note that M 2 = ( SΛS ) 2 = SΛS SΛS = SΛ 2 S then M 3 = M 2 M = ( SΛ 2 S ) ( SΛS ) = SΛ 3 S and so on; for any power M n = SΛ n S

24 RODICA D. COSTIN Then the series (42) is (44) e tm = n! tn M n = For n=0 (45) Λ = it is easy to see that (46) Λ n = therefore (47) n= n! tn Λ n = (48) = and (44) becomes n=0 ( n! tn SΛ n S = S n=0 λ 0... 0 0 λ 2... 0 λ n 0... 0 0 λ n 2... 0... 0 0... λ m... 0 0... λ n m for n =, 2, 3... ) n! tn Λ n S n= n! tn λ n 0... 0 0 n= n! tn λ n 2... 0.. 0 0... e tλ 0... 0 0 e tλ 2... 0... 0 0... e tλm (49) e tm = Se tλ S = etλ. n= n! tn λ n m which shows that the series defining the matrix e tm converges and can be differentiated term-by-term (since these are true for each of the series in (47)). Therefore e tm is a solution of the differential equation (4). Multiplying by an arbitrary constant vector b we obtain vector solutions of (32) as (50) u(t) = e tm b, with b an arbitrary constant vector Noting that u(0) = b it follows that the solution of the initial value problem (32), (33) is u(t) = e tm u 0 Note: the fundamental matrix U(t) in (35) is linked to the fundamental matrix e tm by (5) U(t) = Se tλ = e tm S

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 25 Example. For the example (38) we have [ ] [ ] [ ] 0 e t S =, Λ =, e tλ 0 = 0 3 0 e 3t and u (t) = e t [ e tm = Se tλ S = ], u 2 (t) = e 3t [ The fundamental matrix U(t) is given by (39). Using (49) [ 2 e t + 2 e3 t 2 e t 2 e3 t ] 2 e t 2 e3 t 2 e t + 2 e3 t and the solution to the initial value problem is [ ] [ ( e tm α 2 = e t + 2 e3 t) α + ( 2 e t 2 e3 t) ] β ( β 2 e t 2 e3 t) α + ( 2 e t + 2 e3 t) β which, of course, is the same as (40). 5.3. Non-diagonalizable matrix. The exponential e tm is defined similarly, only a Jordan normal form must be used instead of a diagonal form: writing S MS = J where S is a matrix formed of generalized eigenvectors of M, and J is a Jordan normal form, then (52) e tm = Se tj S It only remains to check that the series defining the exponential of a Jordan form converges, and that it can be differentiated term by term. Also to be determined are m linearly independent solutions, since if M is not diagonalizable, then there are fewer than m independent eigenvectors, hence fewer than m independent solutions of pure exponential type. This can be done using the analogue of (5), namely by considering the matrix (53) U(t) = Se tj = e tm S The columns of the matrix (53) are linearly independent solutions, and we will see that among them are the purely exponential ones multipling the eigenvectors of M. Since J is block diagonal (with Jordan blocks along its diagonal), then its exponential will be block diagonal as well, with exponentials of each Jordan block (see 4.7. for multiplication of block matrices). ] 5.3.. Example: 2 2 blocks: for (54) J = [ ] λ 0 λ

26 RODICA D. COSTIN direct calculations give [ ] [ λ 2 (55) J 2 2 λ λ 3 = 0 λ 2, J 3 3 λ 2 = 0 λ 3 ] [ λ k,..., J k k λ k = 0 λ k ],... and then (56) e tj = k=0 [ k! tk J k = e tλ te tλ 0 e tλ For the equation (32) with the matrix M is similar to a 2 2 Jordan block: S MS = J with J as in (54), and S = [x, x 2 ] a fundamental matrix solution is U(t) = Se tj = [e tλ x, e tλ (tx + x 2 )] whose columns are two linearly independent solutions (57) u (t) = e tλ x, u 2 (t) = e tλ (tx + x 2 ) and any linear combination is a solution: (58) u(t) = a e tλ x + a 2 e tλ (tx + x 2 ) (59) Example. Solve the initial value problem dx dt dy dt = ( + a)x y, x(0) = x + (a )y, y(0) ] = α = β Denoting u = (x, y) T, the differential system (59) is du dt = Mu with M given by (28), matrix for which we found that it has a double eigenvalue a and only one independent eigenvector x = (, ) T. Solution. For this matrix we already found an independent generalized eigenvector x 2 = (, 0) T, so we can use formula (58) to write down the general solution of (59). Solution 2. We know one independent solution to the differential system, namely u (t) = e at x. We look for a second independent solution as the same exponential multiplying a polynomial in t, of degree : substituting u(t) = e at (tb + c) in du dt = Mu we obtain that a(tb + c) + b = M(tb + c) holds for all t, therefore Mb = ab and (M ai)c = b which means that b is an eigenvector of M (or b = 0), and c is a generalized eigenvector. We have re-obtained the formula (57). By either method it is found that a fundamental matrix solution is [ ] U(t) = [u (t), u 2 (t)] = e at t + t and the general solution has the form u(t) = U(t)c for an arbitrary constant vector c. We now determine c so that u(0) = (α, β) T, so we solve [ ] [ ] α c = 0 β

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 27 which gives c = [ 0 ] [ α β ] [ 0 = ] [ α β ] [ = β α β ] and the solution to the initial value problem is [ ] [ ] [ u(t) = e at t + β t (α β) + α = e t at α β β + t (α β) or x(t) = e at (t (α β) + α), y(t) = e at (t (α β) + β) ] 5.3.2. Example: 3 3 blocks: for λ 0 (60) J = 0 λ 0 0 λ direct calculations give λ 2 2 λ J 2 = 0 λ 2 2 λ 0 0 λ 2, J 3 = λ 3 3 λ 2 3 λ 0 λ 3 3 λ 2 0 0 λ 3, J 4 = Higher powers can be calculated by induction; it is clear that (6) J k λ k kλ k k(k ) 2 λ k 2 = 0 λ k kλ k Then (62) e tj = k=0 0 0 λ k k! tk J k = e tλ te tλ 2 t2 e tλ 0 e tλ te tλ 0 0 e tλ λ 4 4 λ 3 6 λ 2 0 λ 4 4 λ 3 0 0 λ 4 For M = SJS with J as in (60) and S = [x, x 2, x 3 ], a fundamental matrix solution for (32) is Se tj = [x e λt, (tx + x 2 )e λt, ( 2 t2 x + tx 2 + x 3 )e λt ]

28 RODICA D. COSTIN 5.3.3. In general, if an eigenvalue λ has multiplicity r, but there are only k < r independent eigenvectors v,..., v k then, besides the k independent solutions e λt v,..., e λt v k there are other r k independent solutions in the form e λt p(t) with p(t) polynomials in t of degree at most r k, with vector coefficients (which turn out to be generalized eigenvectors of M). Then the solution of the initial value problem (32), (33) is u(t) = e tm u 0 Combined with the results of uniqueness of the solution of the initial value problem (known from the general theory of ordinary differential equations) it follows that: Theorem 7. Any linear differential equation u = Mu where M is an m m constant matrix, and u is an m-dimensional vector valued function has m linearly independent solutions, and any solution is a linear combination of these. In other words, the solutions of the equation form a linear space of dimension m. 5.4. Fundamental facts on linear differential systems. Theorem 8. Let M be an n n matrix (diagonalizable or not). (i) The matrix differential problem d (63) dt U(t) = M U(t), U(0) = U 0 has a unique solution, namely U(t) = e Mt U 0. (ii) Let W (t) = det U(t). Then (64) W (t) = TrM W (t) therefore (65) W (t) = W (0) e ttrm (iii) If U 0 is an invertible matrix, then the matrix U(t) is invertible for all t, called a fundamental matrix solution; the columns of U(t) form an independent set of solutions of the system du (66) dt = Mu (iv) Let u (t),..., u n (t) be solutions of the system (66). If the vectors u (t),..., u n (t) are linearly independent at some t then they are linearly independent at any t. Proof. (i) Clearly U(t) = e Mt U 0 is a solution, and it is unique by the general theory of differential equations: (63) is a linear system of n 2 differential equation in n 2 unknowns. (ii) Using (52) it follows that W (t) = det U(t) = det(se tj S U 0 ) = det e tj det U 0 = e t n j= λ j det U 0

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 29 = e ttrm det U 0 = e ttrm W (0) which is (65), implying (64). (iii), (iv) are immediate consequences of (65). 5.5. Eigenvalues and eigenvectors of the exponential of a matrix. It is not hard to show that (e M ) = e M, (e M ) k = e km, e M+cI = e c e M More generally, it can be shown that if MN = NM, then e M e N = e M+N. Warning: if the matrices do not commute, this may not be true! Recall that if M is diagonalizable, in other words if M = SΛS where Λ = diag(λ,..., λ n ) is a diagonal matrix, then e M = Se Λ S where e Λ = diag(e λ,..., eλ n). If follows that the eigenvalues of e M are e λ,..., eλ n and the columns of S are eigenvectors of M, and also of e M. If M is not diagonalizable, let J be its Jordan normal form. Recall that if M = SJS then e M = Se J S where e J is an upper triangular matrix, with diagonal elements still being exponentials of the eigenvalues of M. The matrix e J is not a Jordan normal form; however, generalized eigenvectors of M are also of e M. Exercise.. Show that if Mx = 0 then e M x = x. 2. Show that if v is an eigenvector of M corresponding to the eigenvalues λ, then v is also an eigenvector of e M corresponding to the eigenvalues e λ. 3. Show that if Mv = λv then e M v = e λ [v + (M λi)v]. Note that if (M λi) 2 x = 0 then (e M e λ ) 2 x = 0. Indeed, (e M e λ ) 2 x = (e 2M 2e λ e M + e 2λ I) 2 x = e 2λ e 2(M λ) x 2e 2λ e M λ x + e 2λ x = e 2λ [x + 2(M λ)x] 2e 2λ [x + (M λ)x] + e 2λ x = 0. In general, if x is a generalized eigenvector of M corresponding to the eigenvalues λ, then x is also a generalized eigenvector of e M corresponding to the eigenvalues e λ.

30 RODICA D. COSTIN 5.6. Higher order linear differential equations; companion matrix. Consider scalar linear differential equations, with constant coefficients, of order n: (67) y (n) + a n y (n ) +... + a y + a 0 y = 0 where y(t) is a scalar function and a,..., a n are constants. Such equations can be transformed into systems of first order equations: the substitution (68) u = y, u 2 = y,..., u n = y (n ) transforms (67) into the system (69) u = Mu, where M = 0 0... 0 0 0... 0.. 0 0 0... a 0 a a 2... a n The matrix M is called the companion matrix to the differential equation (67). To find its eigenvalues an easy method is to search for λ so that the linear system Mx = λx has a solution x 0: x 2 = λx, x 3 = λx 2,..., x n = λx n, a 0 x a x 2... a n x n = λx n which implies that (70) λ n + a n λ n +... + a λ + a 0 = 0 which is the characteristic equation of M. Note that the characteristic equation (70) can also be obtained by searching for solutions of (67) which are purely exponential: y(t) = e λt. 5.6.. Linearly independent sets of functions. We are familiar with the notion of linear dependence or independence of functions belonging to a given linear space. In practice, functions arise from particular problems, or classes of problems, for example as solutions of equations and only a posteriori we find a linear space to accommodates them. A natural definition of linear dependence or independence which can be used in most usual linear space of functions is: Definition 9. A set of function f,..., f n are called linearly dependent on an interval I if there are constants c,..., c n, not all zero, so that (7) c f (t) +... + c n f n (t) = 0 for all t I A set of functions which are not linearly dependent on I are called linearly independent on I. This means that if, for some constants c,..., c n relation (7) holds, then necessarily all c,..., c n are zero. If all functions f,..., f n are enough many times differentiable then there is a simple way to check linear dependence or independence:

EIGENVALUES AND EIGENVECTORS. APPLICATIONS 3 Theorem 20. Assume functions f,..., f n are n times differentiable on the interval I. Consider their Wronskian f (t)... f n (t) f (t)... f n(t) W [f,..., f n ](t) =.. f (n ) (t)... f n (n ) (t) (i) If the functions f,..., f n are linearly dependent then W [f,..., f n ](t) = 0 for all t I (ii) If there is some point t 0 I so that W [f,..., f n ](t 0 ) 0 then the functions are linearly independent on I. Indeed, to show (i), assume that (7) holds for some constants c,..., c n, not all zero; then by differentiation, we see that the columns of W (t) are linearly dependent for each t, hence W (t) = 0 for all t. Part (ii) is just the negation of (i). Example. To check if the functions, t 2, e t are linearly dependent we calculate their Wronskian W [, t 2, e t t 2 e t ] = 0 2t e t 0 2 e t = 2 et (t ) is not identically 0 so they are linearly independent (even if the Wronskian happens to be zero for t = ). Example 2. If the numbers λ,..., λ n are all distinct then the functions e tλ,..., e tλn are linearly independent. Indeed, their Wronskian equals the product e tλ... e tλn multiplied by a Vandermonde determinant which equals i<j (λ j λ i ) which is never zero if λ,..., λ n are all distinct, or identically zero if two of λs are equal. I what follows we will see that if the functions f,..., f n happen to be solutions of the same linear differential equation, then their Wronskian is either identically zero, or never zero. 5.6.2. Linearly independent solutions of nth order linear differential equations. Using the results obtained for first order linear systems, and looking just at the first component u (t) of the vector u(t) (since y(t) = u (t)) we find: (i) if the characteristic equation (70) has n distinct solutions λ,..., λ n then the general solution is a linear combination of purely exponential solutions y(t) = a e λ t +... + a n e λnt (ii) if λ j is a repeated eigenvalue of multiplicity r j then there are r j independent solutions of the type e λ jt q(t) where q(t) are polynomials in t of degree at most r j, therefore they can be taken to be e λ jt, te λ jt,..., t r j e λ jt.