Matrix Theory, Math6304 Lecture Notes from Sept 11, 2012 taken by Tristan Whalen

Similar documents
Math 108b: Notes on the Spectral Theorem

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Symmetric and self-adjoint matrices

A PRIMER ON SESQUILINEAR FORMS

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

Math Linear Algebra II. 1. Inner Products and Norms

NOTES ON BILINEAR FORMS

Spectral Theorem for Self-adjoint Linear Operators

Math Linear Algebra Final Exam Review Sheet

Linear Algebra 2 Spectral Notes

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

A = 3 B = A 1 1 matrix is the same as a number or scalar, 3 = [3].

Final A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017

A VERY BRIEF LINEAR ALGEBRA REVIEW for MAP 5485 Introduction to Mathematical Biophysics Fall 2010

Recall : Eigenvalues and Eigenvectors

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

Matrix Theory, Math6304 Lecture Notes from September 27, 2012 taken by Tasadduk Chowdhury

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Quantum Computing Lecture 2. Review of Linear Algebra

Math 489AB Exercises for Chapter 2 Fall Section 2.3

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Math 291-2: Lecture Notes Northwestern University, Winter 2016

Symmetric and anti symmetric matrices

Notes on the matrix exponential

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Linear Algebra: Matrix Eigenvalue Problems

First we introduce the sets that are going to serve as the generalizations of the scalars.

Chapter 6 Inner product spaces

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

Announcements Wednesday, November 01

Math 489AB Exercises for Chapter 1 Fall Section 1.0

Linear algebra 2. Yoav Zemel. March 1, 2012

6 Inner Product Spaces

A linear algebra proof of the fundamental theorem of algebra

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

Linear Algebra Lecture Notes-II

A linear algebra proof of the fundamental theorem of algebra

Homework 11 Solutions. Math 110, Fall 2013.

1. General Vector Spaces

LINEAR ALGEBRA REVIEW

INTRODUCTION TO LIE ALGEBRAS. LECTURE 2.

MIT Final Exam Solutions, Spring 2017

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

Schur s Triangularization Theorem. Math 422

Algebra I Fall 2007

A Brief Outline of Math 355

LINEAR ALGEBRA REVIEW

E2 212: Matrix Theory (Fall 2010) Solutions to Test - 1

Linear Algebra Highlights

Real symmetric matrices/1. 1 Eigenvalues and eigenvectors

Diagonalization of Matrix

Math Spring 2011 Final Exam

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

Linear algebra and applications to graphs Part 1

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

The converse is clear, since

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.

GQE ALGEBRA PROBLEMS

3 Matrix Algebra. 3.1 Operations on matrices

Linear Algebra. Workbook

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Numerical Linear Algebra Homework Assignment - Week 2

WOMP 2001: LINEAR ALGEBRA. 1. Vector spaces

LINEAR ALGEBRA MICHAEL PENKAVA

Eigenvalues and Eigenvectors

Topics in linear algebra

Lecture Summaries for Linear Algebra M51A

Math 110 Linear Algebra Midterm 2 Review October 28, 2017

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Problem 1: Solving a linear equation

1 Matrices and Systems of Linear Equations

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Definition (T -invariant subspace) Example. Example

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Chapter 3. Determinants and Eigenvalues

Assignment 11 (C + C ) = (C + C ) = (C + C) i(c C ) ] = i(c C) (AB) = (AB) = B A = BA 0 = [A, B] = [A, B] = (AB BA) = (AB) AB

Massachusetts Institute of Technology Department of Economics Statistics. Lecture Notes on Matrix Algebra

Math Final December 2006 C. Robinson

LECTURE VII: THE JORDAN CANONICAL FORM MAT FALL 2006 PRINCETON UNIVERSITY. [See also Appendix B in the book]

Calculating determinants for larger matrices

Chapter 4 Euclid Space

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

1 Linear Algebra Problems

MATH JORDAN FORM

Matrix Representation

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.

Jim Lambers MAT 610 Summer Session Lecture 1 Notes

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS

MATHEMATICS 217 NOTES

Eigenvalues and Eigenvectors

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Recitation 9: Probability Matrices and Real Symmetric Matrices. 3 Probability Matrices: Definitions and Examples

Transcription:

Matrix Theory, Math6304 Lecture Notes from Sept 11, 2012 taken by Tristan Whalen 1 Further Review continued Warm-up Let A, B M n and suppose det(a) = 0.Defineamatrixvaluedfunctionasfollows: F (t) =(A + tb) 1 The determinant of F (t) is just a polynomial involving the matrix entries. So in particular the determinant is continuous, so if it is not 0 at t =0,thenitalsoisnot0 for for t sufficiently small. Thus, A + tb is invertible for sufficiently small t so our function F at least makes sense for such t values. Using an explicit formula for the inverse of a matrix in terms of its determinant and the cofactors, we see that F (t) is actually differentiable for sufficiently small t. Now, we wish to compute F (0). InspiredbytheblissfuldaysofCalculus,webeginwith (F (t))(f (t)) 1 =(A + tb) 1 (A + tb) =I and then take the derivative of both sides with respect to t: The right hand side is 0. Thelefthandsideis: d (F (t))(f (t)) 1 = d dt dt (I) d (F (t))(f (t)) 1 dt = F (t)(f (t)) 1 + F (t) d (F (t)) 1 dt = F (t)(a + tb)+f (t) d (A + tb) dt = F (t)(a + tb)+f (t)b Together, then, we have that F (t)(a + tb) +F (t)b = 0, and evaluating at t = 0 gives F (0)(A)+F (0)(B) =F (0)(A)+(A 1 )(B) =0. Right multiplication by A 1 gives F (0) + A 1 BA 1 =0,andwehavetheanswer: F (0) = A 1 BA 1. 1

1.5 Simultaneous Diagonalization (continued) 1.5.34 Definition. AfamilyF M n is a commuting family if for any A, B F we have AB = BA. 1.5.35 Definition. Let A M n. A subspace W C n is A-invariant if Aw W for each w W. If F M n is any family of matrices, then W is called F -invariant if it is A-invariant for every A F. So, if W C n is A-invariant, A maps W into itself and does not take any vectors out of W. 1.5.36 Remark. If W C n is A-invariant for some A M n and dim(w ) 1, thenthereexists an x W \{0} such that Ax = λx for some λ C (i.e., x is an eigenvector for A). Why? If we restrict A to W,thenA W : W W is a linear map on a vector space, so it has an eigenvalue (since our field is C of course). The notion of invariant subspace gives, in some sense, a weaker quality to observe than a basis of eigenvectors (i.e., diagonalizability). Even if we are not able to find a basis of eigenvectors, we (may?) be able to break things down into proper invariant subspaces: 0 1 0 1.5.37 Example. Consider the matrix A = 0 0 0 as a map A : C 3 C 3. Then 0 0 1 W = {[x, y, 0] x, y C} is an invariant subspace. Two eigenvectors of A are v 1 =[1, 0, 0] and v 2 =[0, 0, 1] with corresponding eigenvalues λ 1 =0and λ 2 =1. Note that there is only aone-dimensionalspaceofeigenvectorsinw,butw is two-dimensional. We also have that V = {[0, 0,z] z C} = C is an invariant subspace. So we can write C 3 as a direct sum of proper A-invariant subspaces W V but not as the direct sum of eigenspaces, because the eigenvectors of A do not span C 3. We are considering matrices over C, sothereadermayreplacer with C and the claims still hold. 1.5.38 Lemma. If F M n is a commuting family, then there is an x C n such that Ax = λx for each A F. By the previous remark, each A F has an eigenvector. This lemma says all A F share a common eigenvector. Proof. Choose an F -invariant subspace W C n. (There is such a space, since C n itself is F -invariant.) Furthermore, choose W so that it has a minimal non-zero dimension. (We are able to do this by the Well-Ordering Principle for the natural numbers. Observe that {k : k = dim(w ) and W is an F -invariant subspace} is a non-empty subset of N, sinceitcontains n, soithasaminimum.) Next, we show that any x W \{0} is an eigenvector for each A F by means of contradiction. So, for the purpose of contradiction, suppose this is not the case. That is, there is an A F and a y W, y = 0,suchthatAy / Cy (this is just saying y is not an eigenvector of A). But, from the remark above, we know A has an eigenvector in W,sochoosex W \{0} such that Ax = λx. 2

Now consider the set W 0 := {z W : Az = λz} Note that W 0 W since y W but y/ W 0. Also note that W 0 is a subspace of W (which is straightforward to check; note that λ is fixed). However, if B F,thenforz W 0 we at least have that Bz W by F -invariance of W. Then we have that for z W 0, A(Bz) =ABz = BAz = B(λz) =λbz where we have used the fact that A and B commute by hypothesis and the fact that Az = λz by definition of W 0.ButthisshowsthatBz W 0 by definition of W 0.SoB maps W 0 into W 0. Since B F was arbitrary, this implies W 0 is F -invariant. But now we have a contradiction. We chose W to be F -invariant of minimum dimension and have produced a proper F -invariant subspace of W. Thus, it must be that for each y W, Ay Cy for all A F. 1.5.39 Remark. Again, this lemma says that when you have a family of commuting matrices, this family has at least one common eigenvector. Now we up the ante and define the following: 1.5.40 Definition. A simultaneously diagonalizable family F M n is a family such that there exists an invertible S M n and S 1 AS is diagonal for each A F. Notice that in the definition the same S works for every member of the family. 1.5.41 Theorem. Let F M n be a family of diagonalizable matrices. Then it is a commuting family if and only if it is simultaneously diagonalizable. Proof. First suppose that F is simultaneously diagonalizable (via S). Then for A, B M n, AB = S 1 D 1 SS 1 D 2 S = S 1 D 1 D 2 S = S 1 D 2 D 1 S = S 1 D 2 SS 1 D 1 S = BA So F is a commuting family. Conversely suppose that F is a commuting family of diagonalizable matrices. We proceed by induction on n via a common trick in matrix theory known as deflation. The base case, n =1, has nothing to prove (M 1 = C and everything s awesome). Now we ll suppose the theorem holds for matrices of size n 1 or less. We ll take on the size n case by deflating it to some n 1 size cases. Inductive hypothesis: given a family of diagonalizable matrices F M k for k n 1, ifit is commuting then it is simultaneously diagonalizable. Now let F M n.ifeacha M n is of the form λi, thenthereisnoworktodo(takes = I). So assume A is diagonalizable with eigenvalues {λ 1,...,λ r } where r 2 and AB = BA for each B F. A is similar to a diagonal matrix, so without loss of generality assume A is diagonal. Since each B commutes with the diagonal matrix A, eachb F is a block diagonal matrix (see the theorem about commuting and simultaneously diagonalizable). Since A had at least two distinct entries, each block of each B has size n 1 or less. 3

By the block-wise commutation property (if we take all the blocks of size k n 1 we have a commuting family), together with the inductive hypothesis, all the blocks (of size k) in all B F are simultaneously diagonalizable. Thus there exist fixed matrices T 1,T 2,...,T r such T 1 T 2 that conjugating each B F with T =... gives a block matrix in which the T r blocks are diagonal, which is a diagonal matrix. That is, for any B F, T1 1 B 1 T 1 T 1 T2 1 B 2 T 2 BT =... Tr 1 B r T r D 1 D 2 =... D r 1.6 Hermitian, Normal, Unitary 1.6.42 Definition. Let A M n,m. The adjoint of A, denoteda,isanm n matrix such that < Ax, y >=< x,a y>for all x C m and y C n.ifa M n and A = A,wesayA is Hermitian or self-adjoint. IfA M n satisfies A = A, thenitisskew-hermitian. Unlike the determinant of a matrix in C n,theadjointofamatrixinc n can be computed easily: 1.6.43 Proposition (The Adjoint is the Conjugate Transpose). If A C n where A =[a j,k ], then A =[a k,j ].Thatis,A is the matrix obtained by taking the transpose of A and replacing each entry with its complex conjugate. Proof. Recall that for complex vectors x =[x 1,...,x n ] and y =[y 1,...,y n ],thenthedotproduct is defined as x, y =Σ n k=1 x ky k. Now, let A =[a j,k ] and A =[b j,k ]. By definition of matrix multiplication and of the dot product, Ax, y = [Σ n k=1a j,k x k ] n j=1, [y]n j=1 =Σ n j=1σ n k=1a j,k x k y j On the other hand, x, A y = [x j ] n j=1, [Σ n k=1b j,k y k ] n j=1 =Σ n j=1 x j (Σ n k=1 b j,ky k ) xj Σ n k=1b j,k y k =Σ n j=1 =Σ n j=1σ n k=1b j,k x j y k 4

where we have used basic facts about taking the conjugate (i.e., ab + c = ab + c). Substituting b j,k = a k,j,theaboveequals Σ n j=1σ n k=1a k,j x j y k and we have shown that A =[a k,j ] (simply swap the roles of the indices j and k and compare to the first equation in this proof). That this is the only such A follows since this must be true for all x, y C n,orbyappealingtoadvancedlinearalgebraandtherieszrepresentation Theorem. If we think of the adjoint as analogous to the complex conjugate, then Hermitian matrices correspond to real numbers and skew-hermitian matrices correspond to imaginary numbers. 1.6.44 Facts. A = A if and only if (ia) = (ia). For any A M n,thereexistunique self-adjoint B,C M n with A = B + ic. Notethat B = A + A 2 and C = A A 2i B is called the Hermitian part of A, andic is called the skew-hermitian part of A. An upcoming proposition gives more credence to the notion of thinking of Hermitian matrices as analogous to real numbers. But to prove it we make use of the Polarization Identity. This identity works for general sesquilinear forms, but we only need it for the inner product on C n. 1.6.45 Lemma (The Polarization Identity). Given x, y C n and a matrix A M n,wehave that Ax, y = 1 [A(x + y),x+ y A(x y),x y + ia(x + iy),x+ iy ia(x iy),x iy] 4 Proof. This can be proven directly by simplifying the right-hand side, using the properties of the inner product (linearity in first coordinate, conjugate-linearity in second coordinate). 1.6.46 Remark. Taking A = I above we have the usual presentation of the polarization identity. Notice that on the right-hand side all the terms are of the form Az, z, sowecanconclude something about Ax, y when we only know Ax, x for all x. 1.6.47 Proposition. AmatrixA M n is Hermitian if and only if for all x C n we have that Ax, x R. Proof. If A is Hermitian, then Ax, x = x, A x = x, Ax = Ax, x which implies Ax, x is a real number. Conversely, suppose A M N satisfies, for all x C n, Ax, x R. We write A in terms of its Hermitian and skew-hermitian parts: A = B + ic, where B = B and C = C.Then(usinglinearityoftheinnerproductinthefirstslot),wehave Ax, x = Bx,x + icx,x 5

Since the left hand side is real, the right hand side must be real as well. So it must be that Cx,x =0for all x C n. We show that C =0. By the Polarization Identity (see lemma above), we have that for any x, y C n, Cx,y = 1 [C(x + y),x+ y 4 C(x y),x y + ic(x + iy),x+ iy ic(x iy),x iy] =0 since all the terms on the right equal 0, sincecx,x =0for all x C n.sincecx,y =0for all x, y C n,weconcludethatc =0.ThisimpliesthatA = B, soa = B = B = A, soa is Hermitian. 6