MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

Similar documents
Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Conceptual Questions for Review

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

6 Inner Product Spaces

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

1 Last time: least-squares problems

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

235 Final exam review questions

Quantum Computing Lecture 2. Review of Linear Algebra

MATHEMATICS 217 NOTES

Symmetric and anti symmetric matrices

Linear Algebra Highlights

1. General Vector Spaces

Spectral Theorem for Self-adjoint Linear Operators

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

MATH 511 ADVANCED LINEAR ALGEBRA SPRING 2006

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Matrices and Linear Algebra

Eigenvalues and Eigenvectors A =

Math Matrix Algebra

MATH 235. Final ANSWERS May 5, 2015

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Recall the convention that, for us, all vectors are column vectors.

Dimension. Eigenvalue and eigenvector

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

Diagonalizing Matrices

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Lecture 8 : Eigenvalues and Eigenvectors

(v, w) = arccos( < v, w >

October 25, 2013 INNER PRODUCT SPACES

MTH 2032 SemesterII

Linear Algebra 2 Spectral Notes

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

The Spectral Theorem for normal linear maps

Announcements Wednesday, November 01

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

Linear Algebra Lecture Notes-II

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

PRACTICE PROBLEMS FOR THE FINAL

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45

A SHORT SUMMARY OF VECTOR SPACES AND MATRICES

I. Multiple Choice Questions (Answer any eight)

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

MA 265 FINAL EXAM Fall 2012

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Linear Algebra. Workbook

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

CS 246 Review of Linear Algebra 01/17/19

Singular Value Decomposition (SVD) and Polar Form

Solutions Problem Set 8 Math 240, Fall

Review of Linear Algebra

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

Designing Information Devices and Systems II

12x + 18y = 30? ax + by = m

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

5.6. PSEUDOINVERSES 101. A H w.

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

q n. Q T Q = I. Projections Least Squares best fit solution to Ax = b. Gram-Schmidt process for getting an orthonormal basis from any basis.

Review problems for MA 54, Fall 2004.

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

Typical Problem: Compute.

Mathematical Methods wk 2: Linear Operators

Elementary linear algebra

Linear algebra I Homework #1 due Thursday, Oct Show that the diagonals of a square are orthogonal to one another.

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

Math 102 Final Exam - Dec 14 - PCYNH pm Fall Name Student No. Section A0

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

Math 291-2: Lecture Notes Northwestern University, Winter 2016

18.06SC Final Exam Solutions

Exercise Sheet 1.

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Lecture 23: Trace and determinants! (1) (Final lecture)

MATH Spring 2011 Sample problems for Test 2: Solutions

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

GQE ALGEBRA PROBLEMS

No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question.

Linear Algebra Review

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Math Linear Algebra II. 1. Inner Products and Norms

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

Math 113 Final Exam: Solutions

First we introduce the sets that are going to serve as the generalizations of the scalars.

Lecture notes: Applied linear algebra Part 1. Version 2

A PRIMER ON SESQUILINEAR FORMS

Transcription:

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES 1 Complex matrices Recall that the complex numbers are given by a + ib where a and b are real and i is the imaginary unity, ie, i 2 = 1 In what we describe below, A is an n n matrix, not necessarily real The set of complex numbers is denoted by C The addition and multiplication of two complex numbers are given by (a + ib) + (c + id) = (a + c) + i(b + d) (a + ib)(c + id) = (ac bd) + i(ad + bc) and the usual rules apply In particular complex multiplication is commutative, ie, z 1 z 2 = z 2 z 1 Notable is the inverse, ie, the complex number z that solves the equation (a + ib)z = 1 The solution is given by z = (a ib) a 2 + b 2 If a complex number z = a + ib, then the complex conjugate z = a ib It has the property that zz = z 2 = a 2 + b 2 If an n m matrix A has complex entries, we call it a complex matrix Thus if A is a given complex matrix and b is a given complex vector we may try to solve the equation A x = b Clearly x will be a complex vector The row reduction algorithm works also in this situation and we can talk about the null space of a matrix the column space These are now complex vector spaces One has to think a little about the other two subspaces and this has to do with what we mean by a dot product for complex vectors Now one is tempted to define the dot product between two complex vectors z and w in C n by z 1 w 1 + z 2 w 2 + z n w n Here is the problem: consider the vector in C 2 1 z = i then z z = 1 2 + i 2 = 1 1 = 0 Recall that for real vectors we interpreted the dot of a vector with itself as the square of the length In the complex domain this does not seem to work If we use the complex conjugate, however, the story looks more promising We define the inner product of two vectors w, z C n by w, z = w 1 z 1 + w n z n Note that we have that z, z = z 1 2 + + z n 2 1

2 MATRICES ARE SIMILAR TO TRIANGULAR MATRICES which is strictly positive unless z is the zero vector Hence we may define z = z, z The name inner product is used to distinguish it from the dot product Note that for real vectors the inner product reduces to the dot product There is one thing one has to be careful about We have z, w = w, z, which is different from the corresponding relation for the dot product There are a number of simple consequences One is that Schwarz s inequality still holds, Further the triangle inequality is also true z, w z w z + w z + w These things are easy to proof Do these as an exercise Now we define two vectors z and w to be orthogonal if z, w = 0 Note that in this definition it is irrelevant whether we consider z, w = 0 or w, z = 0, it amounts to the same We define the adjoint of a matrix A the matrix A T and denote it by A Thus the adjoint is found by taking the the complex conjugate of all the matrix elements and then take the transpose or, what amounts to the same, the transpose and then taking the complex conjugate of all its matrix elements As before one easily finds that (AB) = B A and that (A ) 1 = (A 1 ) It is useful to observe that for any complex vectors z and w we have that We have z, A w = A z, w (1) Theorem 11 Let A be an n m matrix The the null space N(A) is a subspace of C m, C(A) a subspace of C n Similarly N(A ) is a subspace of C n and C(A ) a subspace of C m Moreover, C(A) N(A ) = C n and C(A ) N(A) = C m where denotes the orthogonal sum In other words the orthogonal complement of C(A) in C n is N(A ) and the orthogonal complement of C(A ) in C m is N(A) Proof The orthogonal complement of C(A) consists of vectors in C n that are orthogonal to all the column vectors of the matrix A which is the same as all the vectors that are perpendicular to each row of the matrix A Hence C(A) = N(A ) That C(A ) = N(A) follows in the same fashion The next problem is whether there is a notion of least squares approximation for complex matrices As before we would like to find x C m such that A x b has the smallest length In other words we want to minimize A x b 2

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES 3 We claim that the optimizing x is such that A x b is perpendicular to the column space of A Assuming this, we have that A x b must be in the null space of A, ie, A A x = A b This are once again the normal equations What about projections? Assume that V is a complex subspace of C n whose complex dimension is m We choose a basis of in general complex vectors in V and create a matrix with these vectors as column vectors Given a vector b C n we can try to write this vector as a vector in V, ie, the column space of A and a vector perpendicular to V Ie, we have to find x so that As before A x b N(A ) and hence A x b V A A x = A b The matrix A A is m m and has rank m and hence is invertible Hence and the projection of b onto V is given by x = (A A) 1 A b A x = A(A A)A b The matrix P = A(A A)A is easily seen to be a projection In fact we also have that P = P 2 Unitary Matrices The unitary matrices are the analog of complex matrices but in C n Imagine you are given n orthonormal vectors u 1,, u n The we can form It is now easy to check that and hence u k, u l = 0, k l u k, u k = 1, k = 1,, n U = u 1 u n U U = I n U U = I n = UU, because the matrix U has full rank and hence is invertible Such matrices are called unitary matrices It is straightforward to imitate the Gram-Schmidt procedure and hence any complex matrix can be written as A = UR where U has column vectors that are orthonormal (maybe not n of them) and R is an upper triangular matrix Of course, R is complex As a consequence, the column vectors of U form an orthonormal basis for the column space of A The projection onto this column space is given by UU (note that U U = I) The least square solution is then given by R x = Q b

4 MATRICES ARE SIMILAR TO TRIANGULAR MATRICES 3 Eigenvalues and Eigenvectors Recall that a nonzero vector v is an eigenvector of the n n matrix A, if A v = λ v The number λ is called the eigenvalue We know that if v is an eigenvector then A λi is a singular matrix and hence det(a λi) = 0 The characteristic polynomial p A (λ) = det(a λi) is a polynomial with complex coefficients and hence we can factor it into linear factors p A (λ) = ( 1) n (λ λ 1 ) (λ λ n ) where the roots are the eigenvalues The following theorem is a precursor to what is known as Schur factorization, but somewhat simpler Theorem 31 Let A be any complex n n matrix There exists an invertible n n matrix V and an upper triangular matrix T such that A = V T V 1 The diagonal elements of T are precisely the eigenvalues Proof The matrix A has an eigenvalue λ 1 and hence there exists a vector v 1 0 such that A v 1 = λ 1 v 1 Now extend v 1 to a basis by choosing vectors w 2,, w n such that v 1, w 2,, w n form a basis for C n Now Consider the matrix V 1 = v 1, w 2,, w n and note that We are going to write the right side as AV 1 = λ 1 v 1, A w 2,, A w n V 1 V 1 1 λ 1 v 1, A w 2,, A w n = V 1 λ1 V 1 v 1, A w 2,, A w n The matrix V 1 is invertible (why?) The vector λ 1 V 1 v 1 can be written as c 1 c 2 c n and hence the first column of the vector c 1 c 2 V 1 = c 1 v 1 + c 2 w 2 + + c n w n c n must be equal to λ 1 v 1 Since the vectors v 1, w 2,, w n are linearly independent we must have that c 1 = λ 1 and all the other c s equal to 0 Hence λ1 V 1 v 1, A w 2,, A w n = λ1 e 1, A w 2,, A w n

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES 5 where e 1 is the first element of the canonical basis Thus, we have shown that where T 1 is of the form We write AV 1 = V 1 T 1 λ 1 0 0 0 λ1 T T 1 = 0 A 1 where A 1 is an (n 1) (n 1) matrix and denotes some vector whose precise form is irrelevant for our purposes Now we apply the same procedure again to the matrix A 1 and we find an invertible (n 1) (n 1) matrix W 1 such that where A 1 W 2 = W 2 S 2 µ2 T S 2 =, 0 A 2 where A 2 is an (n 2) (n 2) matrix and µ 2 is an eigenvalue of A 1 Define 1 V 2 = 0 W 2 and note that Thus, where λ1 T λ1 T T 1 V 2 = = 0 A 1 W 2 0 W 2 S 2 AV 1 V 2 = V 1 T 1 V 2 = V 1 V 2 T 2 λ1 T T 2 = = 0 S 2 Continuing in this fashion we arrive at λ 1 T 0 µ 2 T 0 0 A 2 AV 1 V n = V 1 V n T = V 2 λ1 T 0 S 2 where T is upper triangular The matrix V = V 1 V n is invertible The fact that the diagonal elements of T are the eigenvalues follows from the relation det(a λi) = det(v T V 1 λi) = detv (T λi)v 1 = det(t λi) and the fact that the determinant of an upper triangular matrix is the product of the diagonal elements Here are two consequences

6 MATRICES ARE SIMILAR TO TRIANGULAR MATRICES Theorem 32 Let A be an n n matrix and denotes its eigenvalues by λ 1,, λ n Then and Proof The first relation follows from deta = λ 1 λ 2 λ n TrA := n a ii = j=1 n λ j j=1 deta = detv T V 1 = dett = λ 1 λ 2 λ n The second relation follows from the fact that for any two n n matrices A, B, which is very easy to see Now This proves the theorem TrAB = TrBA TrA = TrV T V 1 = TrV 1 V T = TrT We continue discussing Cayley s theorem If p(λ) is a polynomial of degree n we may consider p(a) where A is an n n matrix The polynomial can be written as n p(λ) = c j λ j = c 0 + c 1 λ + c 2 λ 2 + + c n λ n j=0 and then we define p(a) = c 0 I n + c 1 A + c 2 A 2 + + c n A n Over the complex numbers we may factor p(λ) and write p(λ) = a(λ λ 1 )(λ λ 2 ) (λ λ n ) where λ 1, λ n are the roots of the polynomial p(λ) and a is some constant It is easy to see that p(a) = a(a λ 1 I n )(A λ 2 I n ) (A λ n I n ) Theorem 33 Let A be an n n matrix and p(λ) := det(a λi n ) its characteristic polynomial Then p(a) = 0 ie, A is a root of its characteristic polynomial False proof The following is nonsensical (why?) p(a) = det(a AI n ) = det(a A) = det0 = 0 Proof We write the characteristic polynomial in the form and note that p(λ) = c 0 + c 1 λ + + c n λ n c 0 I n + c 1 A + + c n A n = V (c 0 I n + c 1 T + + c n T n )V 1 and hence it suffices to prove the theorem for the upper triangular matrix T which amounts to showing that (T λ 1 I n )(T λ 2 I n ) (T λ n I n ) = 0

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES 7 The first column of the upper triangular matrix (T λ 1 I n ) consists only of zeros and the matrix (T λ 2 I n ) is upper triangular but the second entry of the second column is zero The first two columns of the product (T λ 1 I n )(T λ 2 I n ) must therefore be zero The same reasoning shows that (T λ 1 I n )(T λ 2 I n )(T λ 3 I n ) has the first three columns vanishing and so forth This proves Cayley s theorem An interesting aspect of Cayley s theorem is that the first n powers of an n n matrix are linearly dependent Thus, any power of A can be written as a linear combination of I n, A,, A n 1, which is somewhat surprising Example: Consider the Fibonacci matrix 1 1 A = 1 0 whose characteristic polynomial is λ 2 λ 1 Hence we have that So From which we glean the structure Indeed, A 2 = I + A A 3 = A + A 2 = A + I + A = 2A + I A 4 = 2A 2 + A = 2(A + I) + A = 3A + 2I A n = a n A + a n 1 I A n+1 = a n A 2 + a n 1 A = (a n + a n 1 )A + a n I and we get the recursion a n+1 = a n + a n 1 which is the Fibonacci sequence Of course, while this is interesting from a theoretical point of view, it is less useful in practice 4 Hermitean matrices A matrix A is hermitean if A = A This is one the important classes of matrices chiefly because the ubiquitous in applications, like quantum mechanics and the can be diagonalized Recall from (1) that for a hermitean matrix z, A w = A z, w Lemma 41 The eigenvalues of a hermitean matrix are real Proof Suppose λ is an eigenvalue of A with eigenvector w Then A similar computation shows that λ w, w = w, λ w = w, A w λ w, w = Aw, w = w, A w and since A = A it follows that λ w, w = λ w, w and the eigenvalue is real The notion of invariant subspace is a useful one Definition 42 A subspace V C n is invariant under A, if for every w V, A w V Another key fact about hermitean matrices is the following Lemma 43 Let V C n be a subspace invariant under A The its orthogonal complement V is also invariant under A

8 MATRICES ARE SIMILAR TO TRIANGULAR MATRICES Proof Pick any z V and any w V Then w, A z = A w, z = 0 since A w V Since w V is arbitrary, A z V Here is a statement that shows why the invariant subspace notion is useful Lemma 44 Let A be any complex n n matrix and assume that V is invariant under A Set k = dimv There exists a unitary matrix U such that B C U AU = 0 D where B is a k k matrix, C a k n k matrix and D a (n k) (n k) matrix The zero matrix at the bottom is a (n k) k matrix Note that V is not invariant under A, because we do not assume that A = A Proof Pick an orthonormal basis u 1,, u k in V and an orthonormal basis u k+1,, u n in V Since V is invariant under A we have for 1 j k A u j = l=1 k B jl u l for some coefficients B jl If k + 1 j n we have that k n A u j = C jl u l + D jl u l l=1 l=k+1 Form the matrix U = u 1,, u n, which is a unitary matrix, and note that B C AU = A u 1,, A u k, A u k+1 u n = u 1,, u n = U 0 D which proves the claim As a corollary we have that B C 0 D Theorem 45 Let A be a hermitean n n matrix and V C n a k-dimensional subspace invariant under A Then there exists a unitary matrix U such that U AU has the form B 0 U AU = 0 D The k k matrix B and the (n k) (n k) matrix D are both hermitean Proof Both subspaces V and V are invariant under A and hence the result follows from the previous lemma,