Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Similar documents
Math Linear Algebra II. 1. Inner Products and Norms

Review problems for MA 54, Fall 2004.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Math 408 Advanced Linear Algebra

The Singular Value Decomposition and Least Squares Problems

Spectral Theorem for Self-adjoint Linear Operators

NORMS ON SPACE OF MATRICES

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

The following definition is fundamental.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Linear Algebra Lecture Notes-II

Elementary linear algebra

Chapter 6: Orthogonality

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

Quantum Computing Lecture 2. Review of Linear Algebra

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam June 8, 2012

Numerical Linear Algebra

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

1 Last time: least-squares problems

Lecture notes: Applied linear algebra Part 1. Version 2

MATHEMATICS 217 NOTES

Math 108b: Notes on the Spectral Theorem

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Linear Algebra Review. Vectors

Symmetric and self-adjoint matrices

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

A PRIMER ON SESQUILINEAR FORMS

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

MATH 115A: SAMPLE FINAL SOLUTIONS

Eigenvalues and Eigenvectors

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Diagonalizing Matrices

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

Numerical Linear Algebra Homework Assignment - Week 2

MATH36001 Generalized Inverses and the SVD 2015

The Spectral Theorem for normal linear maps

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

MATH 583A REVIEW SESSION #1

Homework 11 Solutions. Math 110, Fall 2013.

Functional Analysis Review

Review of some mathematical tools

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Lecture 2: Linear operators

October 25, 2013 INNER PRODUCT SPACES

Singular Value Decomposition (SVD) and Polar Form

1 Invariant subspaces

Chapter 7: Symmetric Matrices and Quadratic Forms

Notes on Linear Algebra

Summary of Week 9 B = then A A =

GQE ALGEBRA PROBLEMS

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Linear Algebra. Session 12

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

Linear Algebra Massoud Malek

1. The Polar Decomposition

Homework set 4 - Solutions

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

MAT Linear Algebra Collection of sample exams

Lecture 8 : Eigenvalues and Eigenvectors

Math Solutions to homework 5

Matrix Theory. A.Holst, V.Ufnarovski

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Matrix Theory, Math6304 Lecture Notes from October 25, 2012

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

I. Multiple Choice Questions (Answer any eight)

be a Householder matrix. Then prove the followings H = I 2 uut Hu = (I 2 uu u T u )u = u 2 uut u

Final A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017

Lecture 1: Review of linear algebra

Determining Unitary Equivalence to a 3 3 Complex Symmetric Matrix from the Upper Triangular Form. Jay Daigle Advised by Stephan Garcia

4. Linear transformations as a vector space 17

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

NOTES ON BILINEAR FORMS

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Algebra C Numerical Linear Algebra Sample Exam Problems

Lecture 2: Linear Algebra Review

Knowledge Discovery and Data Mining 1 (VO) ( )

Mathematical Methods wk 2: Linear Operators

MATH Linear Algebra

EXAM. Exam 1. Math 5316, Fall December 2, 2012

Chapter 6 Inner product spaces

Diagonalization by a unitary similarity transformation

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Lecture 7: Positive Semidefinite Matrices

Solution to Homework 1

Linear Algebra 2 Spectral Notes

Linear Algebra in Actuarial Science: Slides to the lecture

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Lecture 3: Review of Linear Algebra

14 Singular Value Decomposition

Transcription:

Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal basis B = {u 1, u 2,, u n } of V such that [T ] B is upper triangular In particular, if A M nn (C) then there exists an upper triangular matrix and a unitary matrix U such that A = U U Proof : Let T L(V ) We prove this proposition by induction on dim V The base case, is trivial because the matrix of T with respect to any basis is upper triangular if V is 1-dimensional Let n > 1 be an integer Assume that if dim W = n 1 then for any S L(W ) there exists an orthonormal basis C of W such that [S] C is upper triangular Let n = dim V Let λ be an eigenvalue of T and v be a corresponding eigenvector such that v = 1 Let W = span(v) Then dim W = n 1 Consider that for any orthonormal basis C of W, B = {v} C is an orthonormal basis of V In such a basis B we have [T ] B = λ 0 b 11 b 12 b 1(n 1) 0 b 21 b 22 b 2(n 1) 0 b (n 1)1 b (n 1)2 b (n 1)(n 1) In particular, there exists S L(W ) such that [S] C = b 11 b 12 b 1(n 1) b 21 b 22 b 2(n 1) b (n 1)1 b (n 1)2 b (n 1)(n 1) By induction we can pick an orthonormal basis C of W such that this matrix above is upper triangular Then B is an orthonormal basis of V and [T ] B would be upper triangular

Now let A M nn (C) Then there exists T L(C n ) such that its standard matrix is A Then there exists an orthonormal basis B = {v 1, v 2,, v n } of C n such that [T ] B is upper triangular In particular, using S to denote the standard basis, = [T ] B = [I n ] BS A[I n ] SB = [I n ] 1 SB A[I n] SB is upper triangular Let U = [I n ] SB Clearly, U = [v 1 v 2 v n ] is unitary, hence U 1 = U and A = U U Remark: This argument does not hold for R, since a real matrix may have complex eigenvalues and eigenvectors Indeed we know that the entries on the diagonal of are the eigenvalues! Corollary: Let X be a real inner product space Let T L(X) have only real eigenvalues Then there exists an orthonormal basis B = {u 1, u 2,, u n } of X such that [T ] B is upper triangular In particular, if A M nn (R) only has real eigenvalues then there exists an upper triangular matrix and a unitary (orthogonal) matrix U such that A = U U = U U T Proof : Begin by copying the proof of the previous proposition with R in place of C It only remains to show that the eigenvalues of S L(W ) are real Notice that det([t ] B zi n ) = (λ z)det([s] C zi n 1 ) This means that eigenvalues of S are eigenvalues of T, and hence are real, which completes the proof by induction

2 Spectral Theorem for Self-Adjoint and Normal Operators T L(V ) is called self-adjoint if T = T A matrix A M nn (C) is called self-adjoint or Hermitian if A = A Proposition: Let T L(V ) be self-adjoint Then the eigenvalues of T are real and there exists an orthonormal basis of eigenvectors In particular, if A M nn (C) is Hermitian, then there exists a unitary matrix U and a real diagonal matrix D such that A = UDU Proof : Let T L(V ) be self-adjoint Pick an orthonormal basis B of V such that A = [T ] B is upper triangular A is Hermitian or self-adjoint (as T is) Then the entries above the main diagonal must be zero as the entries below the main diagonal are zero Moreover, the entries on the diagonal are real since they satisfy the equation z = z, which requires that their imaginary parts are 0 Thus A is a real diagonal matrix, so the eigenvalues of T are real Furthermore, the basis B consists of orthonormal eigenvectors! Now let A M nn (C) be Hermitian There exists a unitary matrix U such that A = UDU and D is upper triangular Then D = U AU and D = (U AU) = U A (U ) = U AU = D Thus D is real and diagonal As a consequence of this proposition, eigenvectors of a self-adjoint operator corresponding to different eigenvalues must be orthogonal: Corollary: Let T L(V ) be self-adjoint and λ, µ σ(t ) (eigenvalues) corresponding to eigenvectors u, v respectively If λ µ then u v

An operator T L(V ) (or matrix T M nn (C)) is called normal if T T = T T If T L(V ) is self-adjoint (T = T ) or unitary (T T = I = T T ) then T is normal Proposition: Let T L(V ) be normal Then T has an orthonormal basis of eigenvectors In particular, if A M nn (C) is normal, then there exists a unitary matrix U and a diagonal matrix D such that A = UDU Proof : Let T L(V ) be normal Pick an orthonormal basis B of V such that A = [T ] B is upper triangular A is normal It suffices to show that A is diagonal We will do this by induction on the size of A Clearly, if A is a 1 1 matrix, then it is diagonal and T has an orthonormal basis of eigenvectors Let n > 1 be an integer Assume any (n 1) (n 1) matrix that is normal and uppertriangular must be diagonal Let A M nn be normal and upper triangular Then A = a 11 a 12 a 13 a 1n 0 B 0 where B M (n 1)(n 1) is upper triangular By direct calculation (A A) 11 = a 11 a 11 = a 11 2 and n (A A) 11 = a 1j 2 Since A is normal, this implies a 12 = = a 1n = 0 Hence j=1 A = a 11 0 0 0 0 B 0

A A = a 11 2 0 0 0 0 B B 0 = AA = a 11 2 0 0 0 0 BB 0 implies B B = BB So B is normal and by induction diagonal This means A is diagonal Proposition: T L(V ) is normal iff for all x V, T x = T x Proof : Let T L(V ) Assume T is normal Let x V So T x = T x T x 2 = (T x, T x) = (x, T T x) = (x, T T x) = (T x, T x) = T x 2 Now assume for all x V, T x = T x Let x, y V Using polarization identities, (T T x, y) = (T x, T y) = 1 4 = 1 α T (x + αy) 2 = 1 4 4 α=±1,±i α T x + αt y 2 = 1 4 α=±1,±i α T (x + αy) 2 = α=±1,±i α T x + αt y 2 = (T x, T y) = (T T x, y) α=±1,±i Hence T T = T T, so T is normal Suggested Homework: 21, 23, 24, and 210

3 Polar and Singular Value Decompositions A self-adjoint operator T L(V ) is positive definite if for all non-zero x V, (T x, x) > 0, and positive semidefinite if for all x V, (T x, x) 0 We use the notations T > 0 and T 0 for positive definite and semidefinite respectively Proposition: Let T L(V ) be self-adjoint Then (1) T > 0 iff the eigenvalues of T are positive (2) T 0 iff the eigenvalues of T are non-negative Proof : Let T L(V ) be self-adjoint There exists an orthonormal basis B = {v 1, v 2,, v n } of V of eigenvectors of T such that D = [T ] B = diag{λ 1, λ 2,, λ n } is diagonal, where λ 1, λ 2,, λ n are the corresponding real eigenvalues of T (counting multiplicity) Assume T > 0 Then (T v i, v i ) = λ i > 0 for i = 1, 2,, n Now assume λ i > 0 for i = 1, 2,, n Let v V be non-zero There c 1, c 2, c n C, not all zero, such that v = n i=1 c iv i Then (T v, v) = n i=1 c i 2 λ i > 0 Thus T > 0 This proves (1) We get a proof of (2) by replacing all > with Proposition: Let T L(V ) be a positive semidefinite, self-adjoint operator Then there exists a unique positive semidefinite, self-adjoint operator S L(V ) such that S 2 = T S is denoted T and is called the (positive) square root of T Proof : Let T L(V ) be a positive semidefinite, self-adjoint operator Pick a basis B = {v 1, v 2,, v n } of V of eigenvectors of T corresponding to eigenvalues λ 1, λ 2,, λ n (counting multiplicity) Since T 0, the eigenvalues are non-negative Let S L(V ) be given by [S] B = diag{ λ 1, λ 2,, λ n } The eigenvalues of S are λ i 0 for i = 1, 2,, n So S is a positive semidefinite, self-adjoint operator Clearly S 2 = T since [S] 2 B = [T ] B

It only remains to show that S is unique Suppose R L(V ) is a positive semidefinite, selfadjoint operator satisfying R 2 = T Let µ 1, µ 2,, µ n be the eigenvalues of R corresponding to an orthonormal basis C = {u 1, u 2,, u n } of V consisting of eigenvectors of R Then [T ] C = [R 2 ] C = diag{µ 2 1, µ 2 2,, µ 2 n} So (up to a re-ordering) the eigenvalues λ i of T satisfy λ i = µ 2 i for i = 1, 2,, n Hence, since R is positive semidefinite, µ i = λ i for i = 1, 2,, n So R = S Thus S is unique Consider an operator T L(V, W ) Its Hermitian square is the operator T T L(V ) The Hermitian square is a positive semidefinite, self-adjoint operator: (T T ) = T (T ) = T T and for all x V, (T T x, x) = (T x, T x) = T x 2 0 Let T L(V, W ) Let R = T T L(V ), the square root of the Hermitian square This (positive semidefinite, self-adjoint) operator R is called the modulus of the operator T, and is denoted T Proposition: Let T L(V, W ) For all x V, T x = T x Proof : Let T L(V, W ) Let x V Then T x 2 = ( T x, T x) = ( T T x, x) = ( T 2 x, x) = (T T x, x) = (T x, T x) = T x 2 Thus T x = T x As a consequence Corollary: Let T L(V, W ) Ker T = Ker T = (range T ) The first equality follows immediately from the previous proposition The second equality makes use of the fact that Ker S = (range S ) for S L(V ), and T = T

Proposition (Polar Decomposition): Let T L(V ) Then T = U T for some unitary operator U L(V ) Proof : Let T L(V ) Let x range T Then there exists v V such that x = T v Define U 0 : range T V by U 0 x = T v We must make sure U 0 is well-defined (has a unique output) Suppose v 1 V satisfies x = T v 1 But then v v 1 Ker T = Ker T, so T v = T v 1 as required It is easy to check that U 0 is a linear transformation By construction T = U 0 T and U 0 x = T v = T v = x So U 0 is an isometry (range T ) = Ker T = Ker T = Ker T Pick a unitary operator U 1 : Ker T (range T ) = Ker T, which can be done since dim(ker T ) = dim(ker T ) (Rank Theorem for square matrices!) Define U L(V ) by U = U 0 P range T + U 1 P Ker T, where P X denotes orthogonal projection onto the subspace X (ELFY) U is a unitary transformation and T = U T Let T L(V, W ) The eigenvalues of T = T T are called the singular values of T In particular, if λ is an eigenvalue of T T then λ is a singular value of T Let T L(V, W ) Let σ 1, σ 2,, σ n be the singular values of T (counting multiplicity) Assume σ 1, σ 2,, σ r are the non-zero singular values So σ k = 0 for k = r + 1,, n By definition, σ 2 1, σ 2 2,, σ 2 n are the eigenvalues of T T L(V ) Let v 1, v 2,, v n be an orthonormal basis of V consisting of eigenvectors of T T corresponding to the eigenvalues σ 2 1, σ 2 2,, σ 2 n Such a basis can be selected since T T is self-adjoint So T T v i = σ 2 i v i for i = 1, 2,, n

Let w i = 1 σ i T v i for i = 1, 2,, r Then the vectors w 1, w 2,, w r are orthonormal: Let i, j {1, 2,, r} Then (w i, w j ) = ( 1 σ i T v i, 1 σ j T v j ) = 1 σ i σ j (T T v i, v j ) = 1 σ i σ j (σ 2 i v i, v j ) = σ i σ j (v i, v j ) = δ ij, where δ ij is 1 when i = j and 0 otherwise (Kronecker delta function) Claim: The operator T can be represented by T = σ k w k v k To justify the claim, it suffices to check that these operators coincide on a basis of V For i = 1, 2,, r, σ k w k v k v i = T v k (v i, v k ) = T v i, For i = r + 1,, n, σ i = 0 both σ k w k v k v i = 0 and T v i = 0 since σ i = 0 implies T v i 2 = (T v i, T v i ) = (T T v i, v i ) = (0, v i ) = 0 So since the operators agree on a (orthonormal) basis of V they must be equal The representation T = σ k w k v k is called a singular value decomposition of the operator T

Proposition: Let T L(V, W ) Suppose T can be represented by T = σ k w k v k, where σ i > 0 for i = 1, 2,, r, v 1, v 2, v r are orthonormal in V, and w 1, w 2, w r orthonormal in W Then this representation is a singular value decomposition of T Proof : Assume T L(V, W ) has the representation described in the proposition It suffices to show that T T v i = σ 2 i v i for i = 1, 2,, r and that this accounts for all non-zero eigenvalues of T T Let i {1, 2,, r} First notice that T = σ k v k w k and T v i = σ k w k v k v i = σ i w i Thus T T v i = σ i σ k v k w k w i = σi 2 v i Extend v 1, v 2, v r to an orthonormal basis v 1, v 2, v n of eigenvectors of T T Let λ be an eigenvalue of T T not in {σ 2 i i = 1, 2,, r} Then for some i = r + 1,, n, T T v i = λv i However, T v i = 0 since σ k w k v k v i = 0 So T T v i = 0 and thus λ = 0

Matrix representation of a singular value decomposition: By fixing orthonormal bases for V and W and working in these coordinates, we may assume T L(F n, F m ) Let T = be a singular value decomposition of T σ k w k v k This can be rewritten in matrix form as T = B DÃ, where D = diag{σ 1, σ 2,, σ r } M rr, B M mr has columns w k (k = 1, 2,, r) and à M nr has columns v k (k = 1, 2,, r) In particular, à and B are isometries as their columns are orthonormal In the case when m = n = r, à and B are unitary By extending v 1, v 2,, v r and w 1, w 2,, w r to orthonormal bases v 1, v 2,, v n and w 1, w 2,, w m of V, W respectively, we get unitary matrices A = [v 1 v 2 v n ] M nn and B = [w 1 w 2 w m ] M mm such that T = BDA, where D M mn has D in the upper left r r block and zeroes elsewhere This matrix form is also often called a singular value decomposition of T (ELFY) Write down a random 3 2 matrix and perform the singular value decomposition (in matrix form)

From a singular value decomposition to a polar decomposition: Let T = BDA be a singular value decomposition of T L(V ) in matrix form Then set U = BA U is unitary since U 1 = (BA ) 1 = (A ) 1 B 1 = AB = (BA ) = U Notice that T T = AD B BDA = AD 2 A = (ADA ) 2 T = ADA Then U T = (BA )(ADA ) = BDA = T is a polar decomposition Suggested Homework: 31, 33 (first and third matrices only), 35(a), 36(a)(b), and 39

4 What do Singular Values Describe? Essentially a singular value decomposition is a diagonalization of T L(V, W ) with respect to a pair of orthonormal bases (one for V, one for W ) So what does it tell us? To gain the appropriate intuition, consider the case where T L(R n, R m ) Let B = {x R n : x 1} (the closed unit ball) Let σ 1, σ 2,, σ n be the singular values (counting multiplicity, 0 s at the end of the list) of T Let B, C be the corresponding orthonormal bases from the singular value decomposition of V, W respectively Then the matrix of T with respect to these bases, [T ] CB, has the diagonal matrix D = diag{σ 1, σ 2,, σ r } (corresponding to the non-zero singular values σ 1, σ 2,, σ r ) as its upper left r r block and zeros everywhere else Let v B Set x = [v] B = [I V ] BS v, where S denotes the standard basis Since the columns of [I V ] SB = [I V ] 1 BS are the orthonormal basis vectors of B, this matrix unitary (an isometry), and hence x B Then x = (x 1 x 2 x n ) T satisfies n i=1 x2 i 1 Let y = [T ] CB x = (y 1 y 2 y n ) T Clearly y i = σ i x i for i = 1, 2,, r, and y j = 0 for j = r + 1,, n Thus y satisfies y 2 i σ 2 i=1 i 1 So in the first r orthonormal basis directions in basis C of R m (corresponding to non-zero singular values) y is on or within the ellipsoid given by yi 2 = 1 The last n r orthonormal basis directions in y are killed (sent to 0) i=1 σ 2 i Furthermore, when x is on the unit sphere (boundary of the unit ball), then y will be on the ellipsoid above, and T (B) is the closed ellipsoidal ball with axes of length 2σ 1, 2σ 2,, 2σ r with respect to the first r coordinates given by C

(ELFY) Let B = {v i i = 1, 2,, n} be an orthonormal basis of V Then x = [x] B Now let s discuss the notion of operator norm Let T L(V, W ) The operator norm of T is given by T = max{ T x : x V, x 1} First we justify that the largest singular value is the operator norm: Let x V and x 1 Suppose T = σ k w k v k is a singular value decomposition with respect to two orthonormal bases B = {v i i = 1, 2,, n} and C = {w i i = 1, 2,, m} of V and W respectively Without loss of generality assume that the singular values are arranged in a non-increasing order Let [x] B = (c 1 c 2 c n ) (ELFY) n i=1 c2 i 1 Then [T x] C = [T ] CB [x] B = (σ 1 c 1 σ 2 c 2 σ n c n ) Then T x 2 = [T ] CB [x] B 2 = n σi 2 c 2 i σ1 2 i=1 n c 2 i σ1 2 T x σ 1 i=1 Moreover T v 1 = [T ] CB [v 1 ] B = σ 1 Therefore T = σ 1 (ELFY) Show that the operator norm satisfies the four fundamental properties of a norm: (1) Homogeneity: For all T L(V, W ) and for all c F, ct = c T (2) Triangle Inequality: For all T, S L(V, W ), T + S T + S (3) Positivity (aka non-negativity): For all T L(V, W ), T 0 (4) Definiteness (aka non-degeneracy): Let T L(V, W ) Then T = 0 iff T = 0 L(V, W )

An important and immediate consequence of the homogeneity of the operator norm: Let x V be non-zero and c = x Note c R and c > 0 Set y = 1 x Then T y T c by the definition of the operator norm That implies that T x = T cy = c T y c T = T x The so-called Frobenius norm on L(V, W ) is given by T Frob = trace(a A) where A = [T ] CB for some orthonormal bases B, C of V, W respectively This does not depend on the choice of orthonormal bases (so it s well-defined) because a change of basis matrix between orthonormal bases is unitary, and for an m n matrix D, (UDV ) (UDV ) = V D DV has the same characteristic polynomial (and thus the same trace) as D D for all unitary matrices U M mm and for all unitary matrices V M nn How does this one compare to the operator norm? Again, suppose T = σ k w k v k is a singular value decomposition with respect to two orthonormal bases B = {v i i = 1, 2,, n} and C = {w i i = 1, 2,, m} of V and W respectively Without loss of generality assume that the singular values are arranged in a non-increasing order Then [T ] CB has zeros everywhere except that its upper left r r block is the diagonal matrix D = diag{σ 1, σ 2,, σ r } So T 2 Frob = trace([t ] CB [T ] CB) = r i=1 σ2 i = T 2 + r i=1 σ2 i T T Frob [We could also just recall that the trace of A A is the sum of the eigenvalues]

Suppose we have an invertible matrix A M nn and we want to solve the equation Ax = b, but we only have an estimate of b Let b denote a small perturbation accounting for the error, so instead of solving Ax = b for an exact solution x (because b is not known exactly) we solve Ay = b + b for an approximate solution y Let x = y x So A(x + x) = b + b A x = b x = A 1 b Then ( ) x x = A 1 b x = A 1 b b b x = A 1 b b Ax x A 1 b b A x x = A 1 A b b Note that A and A 1 denote operator norms [A, A 1 are matrices of operators in L(R n ) with respect to the standard (orthonormal) basis of R n ] So, x, which is the relative error in the solution, is bounded by the product of the (real) x number C = A 1 A and b, which is the relative error in the estimate of b b This motivates the following definition: Let A M nn be invertible The condition number of A is C = A 1 A The condition number tells you by how much relative error in the estimate of b is amplified into the relative error in the approximate solution How does the condition number relate to singular values? Let A M nn be invertible Let σ 1, σ 2,, σ n be the singular values of A in non-increasing order, so σ 1 is the largest and σ n is the smallest We already know that A = σ 1 In particular, σ 2 1, σ 2 2,, σ 2 n are the eigenvalues of A A Let B = {v 1, v 2,, v n } be an orthonormal basis of F n consisting of eigenvectors of A A (ELFY) A A is invertible, (A ) 1 = (A 1 ), and (A A) 1 = A 1 (A 1 ) Since A A is invertible, none of the singular values can be 0 (recall: a square matrix is invertible iff 0 is not an eigenvalue)

Now let C = {w i = 1 σ i Av i i = 1, 2,, n} Of course these bases B, C are the orthonormal bases in a singular value decomposition of A Then for i = 1, 2,, n, we have A 1 σ i w i = v i and thus A Av i = σ 2 i v i 1 σ i Av i = (A 1 ) σ i v i w i = (A 1 ) A 1 σ 2 i w i Thus for i = 1, 2,, n (A 1 ) A 1 w i = 1 σ 2 i w i, and thereby the singular values of A 1 are 1 σ 1, 1 σ 2,, 1 σ n So A 1 = 1 σ n the largest singular value of A 1 as that would be In conclusion, the condition number of an invertible matrix A M nn is C = A 1 A = s 1 s n, where s 1 is the largest singular value of A and s n is the smallest singular value of A In other words, its the ratio of the largest to smallest singular value (ELFY) If we pick b = w 1 and b = cw n for c F then the inequality ( ) on the last page becomes an equality Suggested Homework: 41 We will not cover the last two sections of this chapter