Singular Value Decomposition (SVD) and Polar Form

Similar documents
Singular Value Decomposition and Polar Form

Singular Value Decomposition and Polar Form

4.2. ORTHOGONALITY 161

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Summary of Week 9 B = then A A =

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

The Cartan Dieudonné Theorem

Spectral Theorems in Euclidean and Hermitian Spaces

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

1. The Polar Decomposition

Math 408 Advanced Linear Algebra

Spectral Theorem for Self-adjoint Linear Operators

9.1 Eigenvectors and Eigenvalues of a Linear Map

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

The Singular Value Decomposition and Least Squares Problems

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Math 115A: Homework 5

I. Multiple Choice Questions (Answer any eight)

Review problems for MA 54, Fall 2004.

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Review of Some Concepts from Linear Algebra: Part 2

Spectral Theorems in Euclidean and Hermitian Spaces

THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION

Linear Algebra Review. Vectors

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Diagonalizing Matrices

The Singular Value Decomposition

Dimension. Eigenvalue and eigenvector

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

Review of Linear Algebra

UNIT 6: The singular value decomposition.

Math 113 Final Exam: Solutions

The Singular Value Decomposition

Warm-up. True or false? Baby proof. 2. The system of normal equations for A x = y has solutions iff A x = y has solutions

Exercise Set 7.2. Skills

(VII.E) The Singular Value Decomposition (SVD)

Math Linear Algebra II. 1. Inner Products and Norms

Numerical Linear Algebra

Recall : Eigenvalues and Eigenvectors

Linear Algebra (Review) Volker Tresp 2017

Lecture 7: Positive Semidefinite Matrices

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Review of some mathematical tools

Lecture notes: Applied linear algebra Part 1. Version 2

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

Linear Algebra using Dirac Notation: Pt. 2

Linear Algebra 2 Spectral Notes

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Analysis Preliminary Exam Workshop: Hilbert Spaces

Quantum Computing Lecture 2. Review of Linear Algebra

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

LECTURE 7. k=1 (, v k)u k. Moreover r

First of all, the notion of linearity does not depend on which coordinates are used. Recall that a map T : R n R m is linear if

Foundations of Matrix Analysis

Conceptual Questions for Review

Matrix Factorization and Analysis

Linear Algebra Lecture Notes-II

Chapter 5. Basics of Euclidean Geometry

1 Linearity and Linear Systems

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Homework sheet 4: EIGENVALUES AND EIGENVECTORS. DIAGONALIZATION (with solutions) Year ? Why or why not? 6 9

Definition of adjoint

Maths for Signals and Systems Linear Algebra in Engineering

Matrix Theory, Math6304 Lecture Notes from October 25, 2012

LinGloss. A glossary of linear algebra

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

Proposition 42. Let M be an m n matrix. Then (32) N (M M)=N (M) (33) R(MM )=R(M)

PRACTICE FINAL EXAM. why. If they are dependent, exhibit a linear dependence relation among them.

1. Select the unique answer (choice) for each problem. Write only the answer.

Final Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015

Elementary linear algebra

Non-square diagonal matrices

(VI.D) Generalized Eigenspaces

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

MATH 583A REVIEW SESSION #1

Problem # Max points possible Actual score Total 120

235 Final exam review questions

Linear algebra 2. Yoav Zemel. March 1, 2012

Basics of Hermitian Geometry

Jordan Normal Form and Singular Decomposition

Solutions to Final Exam

Eigenvalues and Eigenvectors A =

Singular Value Decompsition

MATHEMATICS 217 NOTES

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

A proof of the Jordan normal form theorem

Applied Linear Algebra in Geoscience Using MATLAB

6 Inner Product Spaces

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

Lecture 2: Linear operators

Transcription:

Chapter 2 Singular Value Decomposition (SVD) and Polar Form 2.1 Polar Form In this chapter, we assume that we are dealing with a real Euclidean space E. Let f: E E be any linear map. In general, it may not be possible to diagonalize f. However, note that f f is self-adjoint, since (f f)(u), v = f(u), f(v) = u, (f f)(v). Similarly, f f is self-adjoint. 29

30 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM The fact that f f and f f are self-adjoint is very important, because it implies that f f and f f can be diagonalized and that they have real eigenvalues. In fact, these eigenvalues are all 0. Thus, the eigenvalues of f f are of the form µ 2 1,..., µ 2 r or 0, where µ i > 0, and similarly for f f. The situation is even better, since we will show shortly that f f and f f have the same eigenvalues.

2.1. POLAR FORM 31 Remark: If f: E F and g: F E are linear maps, then g f and f g always have the same non-zero eigenvalues! Furthermore, if E = F, then 0 is an eigenvalue for f g iff it is an eigenvalue for g f. The square roots µ i > 0 of the positive eigenvalues of f f (and f f ) are called the singular values of f. A self-adjoint linear map f: E E whose eigenvalues are all 0 is called positive semi-definite, for short, positive, and if f is also invertible, positive definite. In the latter case, every eigenvalue is strictly positive. We just showed that f f and f f are positive selfadjoint linear maps. The wonderful thing about the singular value decomposition is that there exist two orthonormal bases (u 1,..., u n ) and (v 1,..., v n ) such that with respect to these bases, f is a diagonal matrix consisting of the singular values of f, or 0.

32 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM Given two Euclidean spaces E and F, where the inner product on E is denoted as, 1 and the inner product on F is denoted as, 2, given any linear map f: E F, there is a unique linear map f : F E such that for all u E and all v F. f(u), v 2 = u, f (v) 1 The linear map f is also called the adjoint of f. This more general situation will be encountered when we deal with the singular value decomposition of rectangular matrices.

2.1. POLAR FORM 33 Recall that if f: E F is a linear map, the image Im f of f is the subspace f(e) of F, and the rank of f is the dimension dim(im f) of its image. Also recall that dim (Ker f) + dim (Im f) = dim (E), and that for every subspace W of E dim (W ) + dim (W ) = dim (E). Lemma 2.1.1 Given any two Euclidean spaces E and F, where E has dimension n and F has dimension m, for any linear map f: E F, we have Ker f = Ker (f f), Ker f = Ker (f f ), Ker f = (Im f ), Ker f = (Im f), dim(im f) = dim(im f ), dim(ker f) = dim(ker f ), and f, f, f f, and f f, have the same rank.

34 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM The next Lemma shows a very useful property of positive self-adjoint linear maps. Lemma 2.1.2 Given a Euclidean space E of dimension n, for any positive self-adjoint linear map f: E E, there is a unique positive self-adjoint linear map h: E E such that f = h 2 = h h. Furthermore, Ker f = Ker h, and if µ 1,..., µ p are the distinct eigenvalues of h and E i is the eigenspace associated with µ i, then µ 2 1,..., µ 2 p are the distinct eigenvalues of f, and E i is the the eigenspace associated with µ 2 i, To prove Lemma 2.1.2, we use the following facts: If g is self-adjoint, positive, then (1) Ker g = Ker g 2 ; (2) If µ 1,..., µ p are the distinct nonzero eigenvalues of g, then µ 2 1,..., µ 2 p are the distinct nonzero eigenvalues of g 2 (= f); (3) Ker (g µ i id) = Ker (g 2 µ 2 i id). By the way, one easily checks that g 2 = g g is positive and self-adjoint.

2.1. POLAR FORM 35 In (3), if (g 2 µ 2 i id)(u) = 0 with u 0, then (g + µ i id) (g µ i id)(u) = 0. But, (g + µ i id) must be invertible since, otherwise, µ i < 0 would be an eigenvalue of g, which is absurd. Therefore, (g µ i id)(u) = 0. Actually, it is possible to prove that if f is self-adjoint, positive, then for every n 2, there is a unique selfadjoint, positive, g, so that f = g n. There are now two ways to proceed. We can prove directly the singular value decomposition, as Strang does [?,?], or prove the so-called polar decomposition theorem. The proofs are roughly of the same difficulty. We choose the second approach since it is less common in textbook presentations, and since it also yields a little more, namely uniqueness when f is invertible.

36 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM It is somewhat disconcerting that the next two theorems are only given as an exercise in Bourbaki [?] (Algèbre, Chapter 9, problem 14, page 127). Yet, the SVD decomposition is of great practical importance. This is probably typical of the attitude of pure mathematicians. However, the proof hinted at in Bourbaki is quite elegant. Theorem 2.1.3 Given a Euclidean space E of dimension n, for every linear map f: E E, there are two positive self-adjoint linear maps h 1 : E E and h 2 : E E and an orthogonal linear map g: E E such that f = g h 1 = h 2 g. Furthermore, if f has rank r, the maps h 1 and h 2 have the same positive eigenvalues µ 1,..., µ r, which are the singular values of f, i.e., the positive square roots of the nonnull eigenvalues of both f f and f f. Finally, g, h 1, h 2 are unique if f is invertible, and h 1 = h 2 if f is normal.

2.1. POLAR FORM 37 In matrix form, Theorem 2.2.1 can be stated as follows. For every real n n matrix A, there is some orthogonal matrix R and some positive symmetric matrix S such that A = RS. Furthermore, R, S are unique if A is invertible. A pair (R, S) such that A = RS is called a polar decomposition of A. Remark: If E is a Hermitian space, Theorem 2.1.3 also holds, but the orthogonal linear map g becomes a unitary map. In terms of matrices, the polar decomposition states that for every complex n n matrix A, there is some unitary matrix U and some positive Hermitian matrix H such that A = UH.

38 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM 2.2 Singular Value Decomposition (SVD) The proof of Theorem 2.1.3 shows that there are two orthonormal bases (u 1,..., u n ) and (v 1,..., v n ), where are eigenvectors of h 1 and (u 1,..., u n ) (v 1,..., v n ) are eigenvectors of h 2. Furthermore, (u 1,..., u r ) is an orthonormal basis of Im f, (u r+1,..., u n ) is an orthonormal basis of Ker f, (v 1,..., v r ) is an orthonormal basis of Im f, and (v r+1,..., v n ) is an orthonormal basis of Ker f. Using this, we immediately obtain the singular value decomposition theorem.

2.2. SINGULAR VALUE DECOMPOSITION (SVD) 39 Theorem 2.2.1 Given a Euclidean space E of dimension n, for every linear map f: E E, there are two orthonormal bases (u 1,..., u n ) and (v 1,..., v n ) such that if r is the rank of f, the matrix of f w.r.t. these two bases is a diagonal matrix of the form µ 1.... µ 2........... µ n where µ 1,..., µ r are the singular values of f, i.e. the (positive) square roots of the nonnull eigenvalues of f f and f f, and µ r+1 =... = µ n = 0. Furthermore, (u 1,..., u n ) are eigenvectors of f f, (v 1,..., v n ) are eigenvectors of f f, and f(u i ) = µ i v i when 1 i n. Note that µ i > 0 for all i (1 i n) iff f is invertible.

40 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM Theorem 2.2.1 can be restated in terms of (real) matrices as follows. Theorem 2.2.2 For every real n n matrix A, there are two orthogonal matrices U and V and a diagonal matrix D such that A = V D U, where D is of the form D = µ 1.... µ 2........... µ n where µ 1,..., µ r are the singular values of f, i.e. the (positive) square roots of the nonnull eigenvalues of A A and A A, and µ r+1 =... = µ n = 0. The columns of U are eigenvectors of A A, and the columns of V are eigenvectors of A A. Furthermore, if det(a) 0, it is possible to choose U and V so that det(u) = det(v ) = +1, i.e., U and V are rotation matrices. A triple (U, D, V ) such that A = V D U is called a singular value decomposition (SVD) of A.

2.2. SINGULAR VALUE DECOMPOSITION (SVD) 41 Remarks. In Strang [?], the matrices U, V, D are denoted as U = Q 2, V = Q 1, and D = Σ, and a SVD decomposition is written as A = Q 1 ΣQ 2. This has the advantage that Q 1 comes before Q 2 in A = Q 1 ΣQ 2. This has the disadvantage that A maps the columns of Q 2 (eigenvectors of A A) to multiples of the columns of Q 1 (eigenvectors of A A ). The SVD also applies to complex matrices. In this case, for every complex n n matrix A, there are two unitary matrices U and V and a diagonal matrix D such that A = V D U, where D is a diagonal matrix consisting of real entries µ 1,..., µ n, where µ 1,..., µ r are the singular values of f, i.e. the positive square roots of the nonnull eigenvalues of A A and A A, and µ r+1 =... = µ n = 0.

42 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM It is easy to go from the polar form to the SVD, and backward. Indeed, given a polar decomposition A = R 1 S, where R 1 is orthogonal and S is positive symmetric, there is an orthogonal matrix R 2 and a positive diagonal matrix D such that S = R 2 D R 2, and thus A = R 1 R 2 D R 2 = V D U, where V = R 1 R 2 and U = R 2 are orthogonal.

2.2. SINGULAR VALUE DECOMPOSITION (SVD) 43 Going the other way, given an SVD decomposition A = V D U, let R = V U and S = UD U. It is clear that R is orthogonal and that S is positive symmetric, and RS = V U UD U = V D U = A. Note that it is possible to require that det(r) = +1 when det(a) 0. Theorem 2.2.2 can be easily extended to rectangular m n matrices (see Strang [?]). As a matter of fact, both Theorems 2.1.3 and 2.2.1 can be generalized to linear maps f: E F between two Euclidean spaces E and F.

44 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM In order to do so, we need to define the analog of the notion of an orthogonal linear map for linear maps f: E F. By definition, the adjoint f : F E of a linear map f: E F is the unique linear map such that f(u), v 2 = u, f (v) 1 for all u E and all v F. Then, we have f(u), f(v) 2 = u, (f f)(v) 1 for all u, v E.

2.2. SINGULAR VALUE DECOMPOSITION (SVD) 45 Letting n = dim(e), m = dim(f ), and p = min(m, n), if f has rank p and if for every p orthonormal vectors (u 1,..., u p ) in (Ker f), the vectors (f(u 1 ),..., f(u p )) are also orthonormal in F, then on (Ker f). f f = id The converse is immediately proved. Thus, we will say that a linear map f: E F is weakly orthogonal if is has rank p = min(m, n) and if on (Ker f). f f = id Of course, f f = 0 on Ker f. In terms of matrices, we will say that a real m n matrix A is weakly orthogonal iff its first p = min(m, n) columns are orthonormal, the remaining ones (if any) being null columns.

46 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM This is equivalent to saying that A A = I n if m n, and that A A = ( ) Im 0 m,n m 0 n m,m 0 n m,n m if n > m. In this latter case (n > m), it is immediately shown that A A = I m, and A is also weakly orthogonal. Weakly unitary linear maps are defined analogously.

2.2. SINGULAR VALUE DECOMPOSITION (SVD) 47 Theorem 2.2.3 Given any two Euclidean spaces E and F, where E has dimension n and F has dimension m, for every linear map f: E F, there are two positive self-adjoint linear maps h 1 : E E and h 2 : F F and a weakly orthogonal linear map g: E F such that f = g h 1 = h 2 g. Furthermore, if f has rank r, the maps h 1 and h 2 have the same positive eigenvalues µ 1,..., µ r, which are the singular values of f, i.e., the positive square roots of the nonnull eigenvalues of both f f and f f. Finally, g, h 1, h 2 are unique if f is invertible, and h 1 = h 2 if f is normal. In matrix form, Theorem 2.2.3 can be stated as follows. For every real m n matrix A, there is some weakly orthogonal m n matrix R and some positive symmetric n n matrix S such that A = RS.

48 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM Remark: If E is a Hermitian space, Theorem 2.2.3 also holds, but the weakly orthogonal linear map g becomes a weakly unitary map. In terms of matrices, the polar decomposition states that for every complex m n matrix A, there is some weakly unitary m n matrix U and some positive Hermitian n n matrix H such that A = UH.

2.2. SINGULAR VALUE DECOMPOSITION (SVD) 49 The proof of Theorem 2.2.3 shows that there are two orthonormal bases of E and of F, where are eigenvectors of h 1 and are eigenvectors of h 2. (u 1,..., u n ) (v 1,..., v m ) (u 1,..., u n ) (v 1,..., v m )

50 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM Furthermore, (u 1,..., u r ) is an orthonormal basis of Im f, (u r+1,..., u n ) is an orthonormal basis of Ker f, (v 1,..., v r ) is an orthonormal basis of Im f, and (v r+1,..., v m ) is an orthonormal basis of Ker f. Using this, we immediately obtain the singular value decomposition theorem for linear maps f: E F, where E and F can have different dimensions.

2.2. SINGULAR VALUE DECOMPOSITION (SVD) 51 Theorem 2.2.4 Given any two Euclidean spaces E and F, where E has dimension n and F has dimension m, for every linear map f: E F, there are two orthonormal bases (u 1,..., u n ) and (v 1,..., v m ) such that if r is the rank of f, the matrix of f w.r.t. these two bases is a m n matrix D of the form D = D = µ 1.... µ 2........... µ n 0.... 0...... 0.... 0 or µ 1... 0... 0 µ 2... 0... 0...... 0. 0... µ m 0... 0 where µ 1,..., µ r are the singular values of f, i.e. the (positive) square roots of the nonnull eigenvalues of f f and f f, and µ r+1 =... = µ p = 0, where p = min(m, n).

52 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM Furthermore, (u 1,..., u n ) are eigenvectors of f f, (v 1,..., v m ) are eigenvectors of f f, and f(u i ) = µ i v i when 1 i p = min(m, n). Even though the matrix D is an m n rectangular matrix, since its only nonzero entries are on the descending diagonal, we still say that D is a diagonal matrix. Theorem 2.2.4 can be restated in terms of (real) matrices as follows.

2.2. SINGULAR VALUE DECOMPOSITION (SVD) 53 Theorem 2.2.5 For every real m n matrix A, there are two orthogonal matrices U (n n) and V (m m) and a diagonal m n matrix D such that A = V D U, where D is of the form D = D = µ 1.... µ 2........... µ n 0.... 0...... 0.... 0 or µ 1... 0... 0 µ 2... 0... 0...... 0. 0... µ m 0... 0 where µ 1,..., µ r are the singular values of f, i.e. the (positive) square roots of the nonnull eigenvalues of A A and A A, and µ r+1 =... = µ p = 0, where p = min(m, n). The columns of U are eigenvectors of A A, and the columns of V are eigenvectors of A A.

54 CHAPTER 2. SINGULAR VALUE DECOMPOSITION (SVD) AND POLAR FORM Given a (complex) n n matrix, A, is there an interesting relationship between the eigenvalues of A and the singular values of A? The following remarkable theorem due to Hermann Weyl shows the answer is yes!: Theorem 2.2.6 (Weyl s inequalities, 1949) For any (complex) n n matrix, A, if λ 1,..., λ n C are the eigenvalues of A and σ 1,..., σ n R + are the singular values of A, listed so that λ 1 λ n and σ 1 σ n 0, then λ 1 λ n = σ 1 σ n and λ 1 λ k σ 1 σ k, for k = 1,..., n 1. Actually, Theorem 2.2.6 has a converse due to A. Horn (1954): Given two sequences λ 1,..., λ n C and σ 1,..., σ n R +, if they satisfy the Weyl inequalities, then there is some matrix, A, having λ 1,..., λ n as eigenvalues and σ 1,..., σ n as singular values. The SVD decomposition of matrices can be used to define the pseudo-inverse of a rectangular matrix.