Singular Value Decomposition and Polar Form

Similar documents
Singular Value Decomposition and Polar Form

Singular Value Decomposition (SVD) and Polar Form

Proposition 42. Let M be an m n matrix. Then (32) N (M M)=N (M) (33) R(MM )=R(M)

The Singular Value Decomposition

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

The Singular Value Decomposition and Least Squares Problems

UNIT 6: The singular value decomposition.

1. The Polar Decomposition

Properties of Matrices and Operations on Matrices

Singular Value Decomposition

Chapter 3 Transformations

Review problems for MA 54, Fall 2004.

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

7. Symmetric Matrices and Quadratic Forms

The Singular Value Decomposition

Math 408 Advanced Linear Algebra

1 Singular Value Decomposition and Principal Component

Spectral Theorems in Euclidean and Hermitian Spaces

Singular Value Decomposition

Linear Algebra Review. Vectors

Summary of Week 9 B = then A A =

Spectral Theorems in Euclidean and Hermitian Spaces

Singular Value Decompsition

Main matrix factorizations

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

4.2. ORTHOGONALITY 161

Eigenvalues and diagonalization

Diagonalizing Matrices

Linear Algebra and Dirac Notation, Pt. 2

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

1 Linearity and Linear Systems

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Lecture notes: Applied linear algebra Part 1. Version 2

THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION

I. Multiple Choice Questions (Answer any eight)

Review of similarity transformation and Singular Value Decomposition

18.06 Problem Set 10 - Solutions Due Thursday, 29 November 2007 at 4 pm in

Notes on Eigenvalues, Singular Values and QR

Linear Algebra, part 3 QR and SVD

Linear Algebra (Review) Volker Tresp 2017

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Moore-Penrose Conditions & SVD

Singular Value Decomposition

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

(VII.E) The Singular Value Decomposition (SVD)

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Math 113 Final Exam: Solutions

Recall the convention that, for us, all vectors are column vectors.

AMS526: Numerical Analysis I (Numerical Linear Algebra)

9.1 Eigenvectors and Eigenvalues of a Linear Map

Linear Algebra. and

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Pseudoinverse & Orthogonal Projection Operators

Manning & Schuetze, FSNLP (c) 1999,2000

Pseudoinverse & Moore-Penrose Conditions

Math Linear Algebra II. 1. Inner Products and Norms

Large Scale Data Analysis Using Deep Learning

3.5 SINGULAR VALUE DECOMPOSITION (OR SVD)

First of all, the notion of linearity does not depend on which coordinates are used. Recall that a map T : R n R m is linear if

Problem # Max points possible Actual score Total 120

Matrix Decomposition in Privacy-Preserving Data Mining JUN ZHANG DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF KENTUCKY

MATH36001 Generalized Inverses and the SVD 2015

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Computational Methods. Eigenvalues and Singular Values

Numerical Linear Algebra

1 Linear Algebra Problems

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Maths for Signals and Systems Linear Algebra in Engineering

Lecture 7: Positive Semidefinite Matrices

Elementary linear algebra

Review of some mathematical tools

Homework 2 Foundations of Computational Math 2 Spring 2019

Conceptual Questions for Review

The Cartan Dieudonné Theorem

Linear Least Squares. Using SVD Decomposition.

15.1 Least Squares Problems and the Pseudo-inverse

GRE Subject test preparation Spring 2016 Topic: Abstract Algebra, Linear Algebra, Number Theory.

Jordan Normal Form and Singular Decomposition

COMP 558 lecture 18 Nov. 15, 2010

The Jordan Normal Form and its Applications

PRACTICE FINAL EXAM. why. If they are dependent, exhibit a linear dependence relation among them.

Computational Methods CMSC/AMSC/MAPL 460. EigenValue decomposition Singular Value Decomposition. Ramani Duraiswami, Dept. of Computer Science

Linear Algebra (Review) Volker Tresp 2018

ECE 275A Homework #3 Solutions

Lecture 6, Sci. Comp. for DPhil Students

MATH 583A REVIEW SESSION #1

Math 291-2: Lecture Notes Northwestern University, Winter 2016

Lecture 8 : Eigenvalues and Eigenvectors

(VI.D) Generalized Eigenspaces

ON THE SINGULAR DECOMPOSITION OF MATRICES

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

Applications of Singular-Value Decomposition (SVD)

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Linear algebra 2. Yoav Zemel. March 1, 2012

Transcription:

Chapter 12 Singular Value Decomposition and Polar Form 12.1 Singular Value Decomposition for Square Matrices Let f : E! E be any linear map, where E is a Euclidean space. In general, it may not be possible to diagonalize f. We show that every linear map can be diagonalized if we are willing to use two orthonormal bases. This is the celebrated singular value decomposition (SVD). 617

618 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM AclosecousinoftheSVDisthepolar form of a linear map, which shows how a linear map can be decomposed into its purely rotational component (perhaps with a flip) and its purely stretching part. The key observation is that f f is self-adjoint, since h(f f)(u),vi = hf(u),f(v)i = hu, (f f)(v)i. Similarly, f f is self-adjoint. The fact that f f and f f are self-adjoint is very important, because it implies that f f and f f can be diagonalized and that they have real eigenvalues. In fact, these eigenvalues are all nonnegative. Thus, the eigenvalues of f f are of the form 2 1,..., 2 r or 0, where i > 0, and similarly for f f.

12.1. SINGULAR VALUE DECOMPOSITION FOR SQUARE MATRICES 619 The situation is even better, since we will show shortly that f f and f f have the same eigenvalues. Remark: Given any two linear maps f : E! F and g : F! E, wheredim(e) =n and dim(f )=m, itcan be shown that ( ) m det(g f I n )=( ) n det(f g I m ), and thus g f and f g always have the same nonzero eigenvalues! Definition 12.1. The square roots i > 0ofthepositive eigenvalues of f f (and f f )arecalledthe singular values of f.

620 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM Definition 12.2. Aself-adjointlinearmap f : E! E whose eigenvalues are nonnegative is called positive semidefinite (or positive), and if f is also invertible, f is said to be positive definite. In the latter case, every eigenvalue of f is strictly positive. We just showed that f f and f f are positive semidefinite self-adjoint linear maps. This fact has the remarkable consequence that every linear map has two important decompositions: 1. The polar form. 2. The singular value decomposition (SVD). The wonderful thing about the singular value decomposition is that there exist two orthonormal bases (u 1,...,u n ) and (v 1,...,v n )suchthat,withrespecttothesebases,f is a diagonal matrix consisting of the singular values of f, or0. Thus, in some sense, f can always be diagonalized with respect to two orthonormal bases.

12.1. SINGULAR VALUE DECOMPOSITION FOR SQUARE MATRICES 621 The SVD is also a useful tool for solving overdetermined linear systems in the least squares sense and for data analysis, as we show later on. Recall that if f : E! F is a linear map, the image Im f of f is the subspace f(e) off,andtherank of f is the dimension dim(im f) ofitsimage. Also recall that dim (Ker f)+dim(imf) =dim(e), and that for every subspace W of E, dim (W )+dim(w? )=dim(e).

622 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM Proposition 12.1. Given any two Euclidean spaces E and F, where E has dimension n and F has dimension m, for any linear map f : E! F, we have Ker f =Ker(f f), Ker f =Ker(f f ), Ker f =(Imf )?, Ker f =(Imf)?, dim(im f) =dim(imf ), and f, f, f f, and f f have the same rank. We will now prove that every square matrix has an SVD. Stronger results can be obtained if we first consider the polar form and then derive the SVD from it (there are uniqueness properties of the polar decomposition). For our purposes, uniqueness results are not as important so we content ourselves with existence results, whose proofs are simpler.

12.1. SINGULAR VALUE DECOMPOSITION FOR SQUARE MATRICES 623 The early history of the singular value decomposition is described in a fascinating paper by Stewart [28]. The SVD is due to Beltrami and Camille Jordan independently (1873, 1874). Gauss is the grandfather of all this, for his work on least squares (1809, 1823) (but Legendre also published a paper on least squares!). Then come Sylvester, Schmidt, and Hermann Weyl. Sylvester s work was apparently opaque. He gave a computational method to find an SVD. Schmidt s work really has to do with integral equations and symmetric and asymmetric kernels (1907). Weyl s work has to do with perturbation theory (1912).

624 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM Autonne came up with the polar decomposition (1902, 1915). Eckart and Young extended SVD to rectangular matrices (1936, 1939). Theorem 12.2. For every real n n matrix A there are two orthogonal matrices U and V and a diagonal matrix D such that A = VDU >, where D is of the form D = 0 B @ 1.... 2........... n 1 C A, where 1,..., r are the singular values of f, i.e., the (positive) square roots of the nonzero eigenvalues of A > A and AA >, and r+1 = = n = 0. The columns of U are eigenvectors of A > A, and the columns of V are eigenvectors of AA >.

12.1. SINGULAR VALUE DECOMPOSITION FOR SQUARE MATRICES 625 Theorem 12.2 suggests the following defintion. Definition 12.3. Atriple(U, D, V )suchthat A = VDU >,whereu and V are orthogonal and D is a diagonal matrix whose entries are nonnegative (it is positive semidefinite) is called a singular value decomposition (SVD) of A. The proof of Theorem 12.2 shows that there are two orthonormal bases (u 1,...,u n )and(v 1,...,v n ), where (u 1,...,u n )areeigenvectorsofa > A and (v 1,...,v n )are eigenvectors of AA >. Furthermore, (u 1,...,u r )isanorthonormalbasisofima >, (u r+1,...,u n )isanorthonormalbasisofkera,(v 1,...,v r ) is an orthonormal basis of Im A, and(v r+1,...,v n )isan orthonormal basis of Ker A >.

626 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM Using a remark made in Chapter 1, if we denote the columns of U by u 1,...,u n and the columns of V by v 1,...,v n,thenwecanwrite A = VDU > = 1 v 1 u > 1 + + r v r u > r. As a consequence, if r is a lot smaller than n (we write r n), we see that A can be reconstructed from U and V using a much smaller number of elements. This idea will be used to provide low-rank approximations of a matrix. The idea is to keep only the k top singular values for some suitable k r for which k+1,... r are very small.

12.1. SINGULAR VALUE DECOMPOSITION FOR SQUARE MATRICES 627 Remarks: (1) In Strang [30] the matrices U, V, D are denoted by U = Q 2, V = Q 1,andD =, and an SVD is written as A = Q 1 Q > 2. This has the advantage that Q 1 comes before Q 2 in A = Q 1 Q > 2. This has the disadvantage that A maps the columns of Q 2 (eigenvectors of A > A)tomultiplesofthecolumns of Q 1 (eigenvectors of AA > ). (2) Algorithms for actually computing the SVD of a matrix are presented in Golub and Van Loan [16], Demmel [11], and Trefethen and Bau [32], where the SVD and its applications are also discussed quite extensively.

628 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM (3) The SVD also applies to complex matrices. In this case, for every complex n n matrix A, therearetwo unitary matrices U and V and a diagonal matrix D such that A = VDU, where D is a diagonal matrix consisting of real entries 1,..., n, where 1,..., r are the singular values of A, i.e.,thepositivesquarerootsofthenonzero eigenvalues of A A and AA,and r+1 =...= n = 0. AnotioncloselyrelatedtotheSVDisthepolarformof amatrix. Definition 12.4. Apair(R, S)suchthatA = RS with R orthogonal and S symmetric positive semidefinite is called a polar decomposition of A.

12.1. SINGULAR VALUE DECOMPOSITION FOR SQUARE MATRICES 629 Theorem 12.2 implies that for every real n n matrix A, thereissomeorthogonalmatrixr and some positive semidefinite symmetric matrix S such that A = RS. Furthermore, R, S are unique if A is invertible, but this is harder to prove. For example, the matrix 0 1 1 1 1 1 A = 1 B1 1 1 1 C 2 @ 1 1 1 1A 1 1 1 1 is both orthogonal and symmetric, and A = RS with R = A and S = I, whichimpliesthatsomeoftheeigenvalues of A are negative.

630 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM Remark: In the complex case, the polar decomposition states that for every complex n n matrix A, thereis some unitary matrix U and some positive semidefinite Hermitian matrix H such that A = UH. It is easy to go from the polar form to the SVD, and conversely. Given an SVD decomposition A = VDU >,let R = VU > and S = UDU >. It is clear that R is orthogonal and that S is positive semidefinite symmetric, and RS = VU > UDU > = VDU > = A.

12.1. SINGULAR VALUE DECOMPOSITION FOR SQUARE MATRICES 631 Going the other way, given a polar decomposition A = R 1 S, where R 1 is orthogonal and S is positive semidefinite symmetric, there is an orthogonal matrix R 2 and a positive semidefinite diagonal matrix D such that S = R 2 DR > 2,andthus A = R 1 R 2 DR > 2 = VDU >, where V = R 1 R 2 and U = R 2 are orthogonal. Theorem 12.2 can be easily extended to rectangular m n matrices (see Strang [30] or Golub and Van Loan [16], Demmel [11], Trefethen and Bau [32]).

632 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM 12.2 Singular Value Decomposition for Rectangular Matrices Theorem 12.3. For every real m n matrix A, there are two orthogonal matrices U (n n) and V (m m) and a diagonal m n matrix D such that A = VDU >, where D is of the form D = 0 B @ 1.... 2........... n 0.... 0...... 0.... 0 1 or C A 0 B @ 1 1... 0... 0 2... 0... 0 C...... 0. 0A,... m 0... 0 where 1,..., r are the singular values of f, i.e. the (positive) square roots of the nonzero eigenvalues of A > A and AA >, and r+1 =... = p = 0, where p =min(m, n). The columns of U are eigenvectors of A > A, and the columns of V are eigenvectors of AA >.

12.2. SINGULAR VALUE DECOMPOSITION FOR RECTANGULAR MATRICES 633 Atriple(U, D, V )suchthata = VDU > is called a singular value decomposition (SVD) of A. Even though the matrix D is an m n rectangular matrix, since its only nonzero entries are on the descending diagonal, we still say that D is a diagonal matrix. If we view A as the representation of a linear map f : E! F,wheredim(E) =n and dim(f )=m, the proof of Theorem 12.3 shows that there are two orthonormal bases (u 1,..., u n )and(v 1,...,v m )fore and F,respectively, where (u 1,...,u n )areeigenvectorsoff f and (v 1,...,v m )areeigenvectorsoff f.

634 CHAPTER 12. SINGULAR VALUE DECOMPOSITION AND POLAR FORM Furthermore, (u 1,...,u r )isanorthonormalbasisofimf, (u r+1,...,u n )isanorthonormalbasisofkerf,(v 1,...,v r ) is an orthonormal basis of Im f, and(v r+1,...,v m )isan orthonormal basis of Ker f. The SVD of matrices can be used to define the pseudoinverse of a rectangular matrix. Computing the SVD of a matrix A is quite involved. Most methods begin by finding orthogonal matrices U and V and a bidiagonal matrix B such that A = VBU >.