Review of similarity transformation and Singular Value Decomposition

Similar documents
COMP 558 lecture 18 Nov. 15, 2010

Linear Algebra: Matrix Eigenvalue Problems

Math Homework 8 (selected problems)

CS 143 Linear Algebra Review

The SVD-Fundamental Theorem of Linear Algebra

Linear Algebra Review. Vectors

Linear Algebra. Min Yan

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Math 108b: Notes on the Spectral Theorem

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

LINEAR ALGEBRA KNOWLEDGE SURVEY

Conceptual Questions for Review

I. Multiple Choice Questions (Answer any eight)

Mathematical foundations - linear algebra

Review problems for MA 54, Fall 2004.

Definition (T -invariant subspace) Example. Example

EECS 275 Matrix Computation

Linear Algebra Review. Fei-Fei Li

Cheat Sheet for MATH461

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

1 Last time: least-squares problems

1 Singular Value Decomposition and Principal Component

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

A Brief Outline of Math 355

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Foundations of Computer Vision

18.06 Quiz 2 April 7, 2010 Professor Strang

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

18.06 Problem Set 10 - Solutions Due Thursday, 29 November 2007 at 4 pm in

MATH 583A REVIEW SESSION #1

Properties of Linear Transformations from R n to R m

Singular Value Decomposition

The Singular Value Decomposition

Singular Value Decomposition. 1 Singular Value Decomposition and the Four Fundamental Subspaces

Designing Information Devices and Systems II

Computational math: Assignment 1

Lecture 2: Linear Algebra

Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007

Review of Some Concepts from Linear Algebra: Part 2

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

14 Singular Value Decomposition

Diagonalizing Matrices

Singular Value Decomposition (SVD)

Singular Value Decomposition and Digital Image Compression

Positive Definite Matrix

Linear Algebra Primer

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

Linear Algebra Massoud Malek

Maths for Signals and Systems Linear Algebra in Engineering

EE731 Lecture Notes: Matrix Computations for Signal Processing

Announcements Wednesday, November 01

7. Symmetric Matrices and Quadratic Forms

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Solution of Linear Equations

Linear Algebra, part 3. Going back to least squares. Mathematical Models, Analysis and Simulation = 0. a T 1 e. a T n e. Anna-Karin Tornberg

Lecture 6: Lies, Inner Product Spaces, and Symmetric Matrices

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

Eigenvalues and Eigenvectors

5. Orthogonal matrices

Mathematical foundations - linear algebra

MATH 612 Computational methods for equation solving and function minimization Week # 2

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Chapter 7: Symmetric Matrices and Quadratic Forms

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

Chapter 3 Transformations

Elementary Linear Algebra

Large Scale Data Analysis Using Deep Learning

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes

Main matrix factorizations

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

The Jordan Normal Form and its Applications

Stat 159/259: Linear Algebra Notes

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

Linear Algebra, part 3 QR and SVD

Notes on Eigenvalues, Singular Values and QR

Math 307 Learning Goals. March 23, 2010

We use the overhead arrow to denote a column vector, i.e., a number with a direction. For example, in three-space, we write

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

The Singular Value Decomposition and Least Squares Problems

MATH36001 Generalized Inverses and the SVD 2015

The QR Factorization

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

Linear Algebra and Dirac Notation, Pt. 3

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

Detailed Assessment Report MATH Outcomes, with Any Associations and Related Measures, Targets, Findings, and Action Plans

Solutions to Review Problems for Chapter 6 ( ), 7.1

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

HONORS LINEAR ALGEBRA (MATH V 2020) SPRING 2013

Vectors and Matrices Statistics with Vectors and Matrices

Lecture 8: Linear Algebra Background

1. The Polar Decomposition

Quantum Computing Lecture 2. Review of Linear Algebra

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

Transcription:

Review of similarity transformation and Singular Value Decomposition Nasser M Abbasi Applied Mathematics Department, California State University, Fullerton July 8 7 page compiled on June 9, 5 at 9:5pm Contents Similarity transformation Derivation of similarity transformation based on algebraic method Derivation of similarity transformation based on geometric method 3 Finding matrix representation of linear transformation T 5 Finding matrix representation of change of basis 7 3 Examples of similarity transformation 9 4 How to find the best basis to simplify the representation of T? 4 3 Summary of similarity transformation 4 Singular value decomposition (SVD) 5 What is right and left eigenvectors? 5 SVD 5 3 Conclusion 7 4 References 8

Similarity transformation A similarity transformation is B M AM Where B, A, M are square matrices The goal of similarity transformation is to find a B matrix which has a simpler form than A so that we can use B in place of A to ease some computational work Lets set our goal in having B be a diagonal matrix (a general diagonal form is called block diagonal or Jordan form, but here we are just looking at the case of B being a diagonal matrix) The question becomes: Given A, can we find M such that M AM is diagonal? The standard method to show the above is via an algebraic method, which we show first Next we show a geometric method that explains similarity transformation geometrically Derivation of similarity transformation based on algebraic method Starting with B M AM Our goal is to find a real matrix B such that it is diagonal M we obtain From the above, by pre multiplying each side by AM MB Now, since our goal is to make B diagonal, let us select the eigenvalues of A to be the diagonal of B Now we write the above in expanded form as follows a a n m m n a n a nn m n m nn m m n λ m n m nn λ n The above can be written as n separate equations by looking at the column view of matrix multiplication and all the way to the n th column of M a a n a n a nn a a n a n a nn a a n a n a nn Each of the above n equations is in the form m m n m m n m n m nn Ax λx λ λ λ n m m n m m n m n m nn

and this is the key idea The above shows that if we take the diagonal elements of B to be the eigenvalues of A then the M matrix is the (right) eigenvectors of A Hence, if we want the B matrix to be diagonal and real, this means the A matrix itself must have some conditions put on it Specifically, A eigenvalues must be all distinct and real This happens when A is real and symmetric Therefore, we have the general result that given a matrix A which has distinct and real eigenvalues, only then can we find a similar matrix to it which is real and diagonal, and these diagonal values are the eigenvalues of A Derivation of similarity transformation based on geometric method It turns out that this derivation is more complicated than the algebraic approach It involves a change of basis matrix (which will be our M matrix) and the representation of linear transformation T as a matrix under some basis (which will be our A matrix) Let us start from the beginning Given the vector x in R n, it can have many different representations (or coordinates) depending on which basis are used There are an infinite number of basis that span R n, hence there { are infinite ]} representations for the same vector Consider for example R The standard basis vectors are,, but another basis are { } { },, and yet another are,, and so on In R n any n linearly independent vectors can be used as basis Let x v be the vector representation in wrt basis v, and let x v be the vector representation wrt basis v An important problem in linear algebra is to obtain x v given x v and the change of basis matrix M] v v This requires finding a matrix representation for the change of basis Once the M] v v matrix is found, we can write x v M] v v x v () 3

Where M] v v is the change of basis matrix which when applied to x v wrt basis v From () we see that given x v, then to obtain x v we write x v M] v v x v returns the coordinates of the vector Another important problem is that of linear transformation T, where now we apply some transformation on the whole space and we want to find what happens to the space coordinates (all the position vectors) due to this transformation Consider for example T which is a rotation in R by some angle θ Here, every position vector in the space is affected by this rotation but the basis remain fixed, and we want to find the new coordinates of each point in space Let x v be the position vector before applying the linear transformation, and let y v be the new position vector Notice that both vectors are written with respect to the same basis v Similar to change basis, we want a way to find y v from x v, hence we need a way to represent this linear transformation by a matrix, say T ] v with respect to the basis v and so we could write the following (A) y v T ] v x v () Assume the basis is changed to v Let the representation of the same linear transformation T wrt basis v be called T ] v, so we write y v T ] v x v (A) Hence when the basis changes from v to v, T representation changes from T ] v to T ] v Now assume we are given some linear transformation T and its representation T ] v wrt basis v, how could we find T representation T ] v wrt to new basis v? From () we have 4

y v T ] v x v But from () x v M] v v x v, hence the above becomes y v T ] v (M] v v x v ) Pre-multiplying from left the above equation by M] v v we obtain But since M] v v y v y v, then above reduces to M] v v y v M] v v T ] vm] v vx v y v M] v v T ] vm] v vx v But from (A) y v T ] v x v, hence the above becomes Therefore T ] v x v M] v v T ] vm] v vx v T ] v M] v v T ] vm] v v (3) Notice the difference between change of basis and linear transformation In change of basis, the vector x remained fixed in space, but the basis changed, and we want to find the coordinates of the vector wrt the new basis With linear transformation, the vector itself is changed from x v to y v, but both vectors are expressed wrt the same basis Equation (3) allows us to obtain a new matrix representation of the linear transformation T by expressing the same T wrt to different basis Hence if we are given a representation of T which is not the most optimal representation, we can, by change of basis, obtain a different representation of the same T by using (3) The most optimal matrix representation for linear transformation is a diagonal matrix Equation (3) is called a similarity transformation To make it easier to compare (3) above with what we wrote in the previous section when we derived the above using an algebraic approach, we let T ] v B, T ] v A, hence (3) is B M AM The matrix T ] v is similar to T ] v (ie B is similar to A) Both matrices represent the same linear transformation applied on the space We will show below some examples of how to find T ] v given T ] v and M] But first we show how to obtain matrix representation of T Finding matrix representation of linear transformation T Given a vector x R n and some linear transformation (or a linear operator) T that acts on the vector x transforming this vector into another vector y R n according to some prescribed manner T : x y Examples of such linear transformation are rotation, elongation, compression, and reflection, and any combination of these operations, but it can not include the translation of the vector, since translation is not linear The question to answer here is how to write down a representation of this linear transformation? It turns out that T can be represented by a matrix of size n n, and the actual numbers that go into the matrix, or the shape of the matrix, will depend on which basis in R n we choose to represent x We would like to pick some basis so that the final shape of the matrix is the most simple shape Let us pick a set of basis v {e, e,, e n } that span R n hence x v a e + a e + + a n e n Now T x v T (a e + a e + + a n e n ) 5

And since T is linear we can rewrite the above as T x v a T e + a T e + + a n T e n () We see from above that the new transformed vector T x v has the same coordinates as x v if we view T e i as the new basis Now T e i itself is an application of the linear transformation but now it is being done on each basis vector e i Hence it will cause each specific basis vector to transform in some manner to a new vector Whatever this new vector is, we must be able to represent it in terms of the same basis vectors {e, e,, e n }, therefore, we write Where ξ ij is the i th coordinate of the vector T e j And we do the same for T e T e ξ e + ξ e + + ξ n e n and so on until T e ξ e + ξ e + + ξ n e n Or T e n ξ n e + ξ n e + + ξ nn e n Now plug the above into () and obtain T e j ξ ij e i T x v a (ξ e + ξ e + + ξ n e n ) + a (ξ e + ξ e + + ξ n e n ) + + a n (ξ n e + ξ n e + + ξ nn e n ) Hence, if we take the e i as common factors, we have T x e (a ξ + a ξ + + a n ξ n ) + e (a ξ + a ξ + + a n ξ n ) + + e n (a ξ n + a ξ n + + a n ξ nn ) Hence, since T x v a T e + a T e + + a n T e n from (), then by comparing coefficients of each basis vector e i we obtain a T a ξ + a ξ + + a n ξ n a T a ξ + a ξ + + a n ξ n a n T a ξ n + a ξ n + + a n ξ nn Or in matrix form 6

Hence we finally obtain that a ξ ξ ξ n a a ξ ξ ξ n a T a n ξ n ξ n ξ nn a n ξ ξ ξ n ξ ξ ξ n T T ] v ξ n ξ n ξ nn Let us call the matrix that represents T under basis v as T ] v We see from the above that the j th column of T ] v contain the coordinates of the vector T e j This gives us a quick way to construct T ] v : Apply T to each basis vector e j, and take the resulting vector and place it in the j th column of T ] v We now see that T ] v will have a different numerical values if the basis used to span R n are different from the ones used to obtain the above T ] v Let use pick some new basis, say v {e, e,, e n} Let the new representation of T now be the matrix T ] v, then the j th column of T ] v is found from applying T on the new basis e j T e j ζ ij e i Where now ζ ij is the i th coordinates of the the vector T e j which will be different from ξ ij since in one case we used the basis set v {e i } and in the other case we used the basis set v {e i} Hence we see that T ] v will numerically be different depending on the basis used to span the space even though the linear transformation T itself did not change Finding matrix representation of change of basis Now we show how to determine M] v v, the matrix which when applied to a vector x v will result in the vector x v Given a vector x, it is represented wrt basis v {e, e,, e n } as and wrt basis v {e, e,, e n} as x v a e + a e + + a n e n x v b e + b e + + b n e n 7

But the vector itself is invariant under any change of basis, hence x v x v a e + a e + + a n e n b e + b e + + b n e n () Now write each basis e i in terms of the basis e j, hence we have e c e + c e + + c n e n e c e + c e + + c n e n e n c n e + c n e + + c nn e n () Substitute () into RHS of () we obtain a e + a e + + a n e n b (c e + c e + + c n e n ) + b (c e + c e + + c n e n ) + + b n (c n e + c n e + + c nn e n ) Factor out the basis e i from the RHS, we obtain a e + a e + + a n e n (b c e + b c e + + b c n e n ) + (b c e + b c e + + b c n e n ) + + (b n c n e + b n c n e + + b n c nn e n ) hence a e + a e + + a n e n e (b c + b c + + b n c n ) + e (b c + b c + + b n c n ) + + e n (b c n + b c n + + b n c nn ) Now comparing coefficients of each of the basis e i we obtain the following result a b c + b c + + b n c n a b c + b c + + b n c n a n b c n + b c n + + b n c nn or in Matrix form, we write a c c c 3 c n b a c c c 3 c n b a n c n c n c 3n c nn b n (3) 8

The above gives a relation between the coordinates of the vector x wrt basis v (these are the a i coordinates) to the coordinates of the same vector wrt to basis v (these are the coordinates b j ) The mapping between these coordinates is the matrix shown above which we call the M] matrix Since the above matrix returns the coordinates of the vector wrt v when it is multiplied by the coordinates of the vector wrt basis v, we write it as M] v v to make it clear from which basis to which basis is the conversion taking place Looking at () and (3) above, we see that column j in M] v v is the coordinates of the e j wrt basis v Hence the above gives a method to generate the change of basis matrix M] v v Simply represent each basis in v wrt to basis v Take the result of each such representation and write it as a column in M] v v We now show few examples to illustrate this Example showing how to find { change ] of ]} basis matrix Given the R basis v {e, e },, { } and new basis v {e, e },, find the change of basis matrix M] v v Column of M] v v is the coordinates of e wrt basis v ie e c e + c e (4) and column of M] v v is the coordinates of e wrt basis v ie But (4) is e c e + c e (5) c + c Hence c c and c + c, solving, we obtain c, c and (5) is c + c Hence c c and c + c, solving we obtain c, c, hence M] v v Now we can use the above change of basis matrix ] to find coordinates of the vector under v given its coordinates under v For example, given x v, then 5 x v 3 Examples of similarity transformation ] ] 75 5 5 Example This example from Strang book (Linear Algebra and its applications, 4th edition), but shown here with little more details 9

{ } Consider R, and let T be the projection onto a line at angle θ 35 Let the first basis v, { } and let the second basis v, The first step is to find M] v v The first column of M] v v is found by writing e wrt basis v, and the second column of M] v v by writing e wrt basis v, Hence is found e c e + c e ] ] c + c or c and c And e c e + c e ] ] c + c Hence c and c Hence M] v v Now we need to find T ] v the representation of T wrt basis v

The first column of T ] v is the new coordinates of the basis v after applying T onto it but T e T 5 5 and the second column of T ] v is the new coordinates of the basis v after applying T onto it but T e T 5 5 Hence Therefore 5 5 T ] v 5 5 T ] v M] v v T ] vm] v v 5 5 5 5 T ] v Hence we see that the linear transformation T is represented as wrt basis v, while it is represented 5 5 as wrt basis v Therefore, it will be better to perform all calculations involving this linear 5 5 transformation under basis v instead of basis v Example

{ } Consider R, and let T be a rotation of space by θ π 4 Let the first basis v, and let the { } second basis be v, The first step is to find M] v v The first column of M] v v is found by writing e wrt basis v, and the second column of M] v v by writing e wrt basis v, Hence is found e c e + c e ] ] c + c or c and c and e c e + c e ] ] c + c Hence c and c

Hence M] v v Now we need to find T ] v the representation of T wrt basis v The first column of T ] v is the new coordinates of the basis v after applying T onto it but T e T cos θ sin θ ] and the second column of T ] v is the new coordinates of the basis v after applying T onto it but T e T sin θ T e cos θ Hence Therefore cos θ sin θ T ] v sin θ cos θ cos π 4 sin π 4 sin π 4 cos π 4 T ] v T ] v M] v v T ] vm] v v ] Hence we see that the linear transformation T is represented as represented as wrt basis v, while it is wrt basis v Therefore the change of basis selected above did not result in making the representation of T any simpler (in fact there was no change in the representation) This means we need to find a more direct method of finding the basis under which T has the simplest representation Clearly we can t just keep trying different basis to find if T has simpler representation under the new basis 3

4 How to find the best basis to simplify the representation of T? Our goal of simpler representation of T is that of a diagonal matrix or as close as possible to being diagonal matrix (ie Jordan form) Given T ] v and an infinite number of basis v that we could select to represent T under in the hope we can find a new representation T ] v such that it is simpler than T ] v, we now ask, how does one go about finding such basis? It turns out the if we select the eigenvectors of T ] v as the columns of M] v v, this will result in T ] v being diagonal or as close to being diagonal as possible (block diagonal or Jordan form) Let us apply this to the second example in the previous section In that example, we had T ] v i i i i The eigenvectors of T ] v are and, M] v v Therefore T ] v M] v v T ] vm] v v i i i i ( T ] v i) ] ( + i) Which is diagonal representation ( Hence we see that the linear transformation T is represented as i) ] ( + i) wrt basis v 3 Summary of similarity transformation To obtain B M AM such that B is real and diagonal requires that A be real and symmetric The eigenvalues of A goes into the diagonal of B and the eigenvectors of A go into the columns of M This is an algebraic view Geometrically, A is viewed as the matrix representation under some basis v of a linear transformation T And B is the matrix representation of the same linear transformation T but now under a new basis v and M is the matrix that represents the change of basis from v to v The question then immediately arise: If A must be real and symmetric (for B to be real and diagonal), what does this mean in terms of the linear transformation T and change of basis matrix M? This clearly mean that, under some basis v, not every linear transformation represented by A can have a similar matrix B which is real and diagonal Only those linear transformations which result in real and symmetric representation can have a similar matrix which is real and diagonal This is shown in the previous examples, where we saw that T defined { } as rotation by 45 under the standard basis v, resulted in A and since this is not symmetric, hence we will not be able to factor this matrix using similarity transformation to a diagonal matrix, no matter which change of basis we try to represent T under The question then, under a given basis v what is the class of linear transformation which leads to a matrix representation that is symmetric? One such example was shown above which is the projection into the line at 35 This question needs more time to look into We now write Λ to represent a diagonal matrix, hence the similarity transformation above can be written as Λ M AM 4

or A MΛM Singular value decomposition (SVD) Using similarity transformation, we found that for a real and symmetric matrix A we are able to decompose it as A MΛM where Λ is diagonal and contains the eigenvalues of A on the diagonal, and M contains the right eigenvectors of A in its columns What is right and left eigenvectors? Right eigenvectors are the standard eigenvectors we have been talking about all the time When λ is an eigenvalues of A then x is a right eigenvector when we write However, x is a left eigenvector of A when we have where x H is the Hermitian of x SVD Ax λx x H A λx H Given any arbitrary matrix A m n it can be factored into 3 matrices as follows A m n P m m D m n Q n n where P is a unitary matrix (P H P or P H P I), and Q is also unitary matrix These are the steps to do SVD Find the rank of A, say r Let B A H n ma m n, hence B n n is a square matrix, and it is semi positive definite, hence its eigenvalues will all be Find the eigenvalues of B, call these σi There will be n such eigenvalues since B is of order n n But only r of these will be positive, and n r will be zero Arrange these eigenvalues such n eigenvalues {}}{ that the first r non-zero eigenvalues come first, followed by the zero eigenvalues: σ, σ,, σr,,,, 3 Initialize matrix D m n to be all zeros Take the the first r eigenvalues from above (non-zero ones), and r singular values {}}{ take the square root of each, hence we get σ, σ,, σ r,and write these down the diagonal of D starting 5

at D (, ), ie D (, ) σ, D (, ) σ,, D (r, r) σ r Notice that the matrix D need not square matrix Hence we can obtain an arrangement such as the following for r σ D σ where the matrix A was 3 4 for example 4 Now find each eigenvalue of A H A For each eigenvalue, find the corresponding eigenvector Call these eigenvectors u, u,, u n 5 Normalize the eigenvectors found in the above step Now u, u,, u n eigenvector will be an orthonormal set of vectors Take the Hermitian of each eigenvector u H, uh,, uh n and make one of these vectors (now in row format instead of column format) go into a row in the matrix Qie the first row of Q will be u H, the second row of Q will be uh, etc ( A H A ) n n u, u,, u r, ortho basis (n-r) for N(A) {}}{ u r+,, u n Q n n u T u T u T r u T r+ u T n find eigenvectors and normalize 6 Now we need to find a set of m orthonormal vectors, these will be the columns of the matrix P : find a set of r orthonormal eigenvector v i σ i Au i, for i r Notice that here we only use the first r vectors found in step 5 Take each one of these v i vectors and make them into columns of P But since we need m columns in P not just r, we need to come up with m r more basis vectors such that all the m vectors form an orthonormal set of basis vectors for the row space of A, ie C m If doing this by hand, it is easy to find the this m r by inspection In a program, we could use the same process we used with Gram-Schmidt, where we learned how find a new vector which is orthonormal to a an existing set of other vectors Another way to find the matrix P is to construct the matrix AA H and find its eigenvectors, those go into the columns of the matrix P as follows ( AA H) m m orthonormal basis for range A {}}{ ] v, v,, v r, v r+,, v m P m m v v v r v r+ v m 7 This completes the factorization, now we have A P D Q The following diagram help illustrate the SVD process described above find eigenvectors and normalize 6

3 Conclusion As a conclusion, the following diagram shows the difference between similarity transformation factorization of a matrix and SVD 7

Trying to understand SVD from geometrical point view requires more time, and this is left for future research into this subject 4 References Linear Algebra and its applications 4th edition by Gilbert Strang Course notes, Linear Algebra, Math 37, CSUF mathematics department, spring 7 by Dr Angel Pineda 8

3 Principles and techniques of applied Mathematics by Bernard Friedman, Spectral theory of operators chapter 4 Matrix computation, 3rd edition by Gene Golub and Charles Van Loan 5 Numerical analysis: Mathematics of scientific computing, by David Kincaid and Ward Cheney 9