Pseudoinverse & Moore-Penrose Conditions

Similar documents
Pseudoinverse & Orthogonal Projection Operators

Pseudoinverse and Adjoint Operators

Moore-Penrose Conditions & SVD

ECE 275A Homework #3 Solutions

Normed & Inner Product Vector Spaces

ECE 275A Homework # 3 Due Thursday 10/27/2016

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Solutions to Review Problems for Chapter 6 ( ), 7.1

Lecture 13: Orthogonal projections and least squares (Section ) Thang Huynh, UC San Diego 2/9/2018

UNIT 6: The singular value decomposition.

Math 407: Linear Optimization

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016

Lecture notes: Applied linear algebra Part 1. Version 2

Properties of Matrices and Operations on Matrices

Worksheet for Lecture 25 Section 6.4 Gram-Schmidt Process

Vector Space Concepts

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

Generalized Gradient Descent Algorithms

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

2. Review of Linear Algebra

Miderm II Solutions To find the inverse we row-reduce the augumented matrix [I A]. In our case, we row reduce

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

Problem # Max points possible Actual score Total 120

A Review of Linear Algebra

Review problems for MA 54, Fall 2004.

5. Orthogonal matrices

Stat 159/259: Linear Algebra Notes

LINEAR ALGEBRA SUMMARY SHEET.

. = V c = V [x]v (5.1) c 1. c k

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Chapter 6: Orthogonality

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

Typical Problem: Compute.

Linear Algebra (Review) Volker Tresp 2017

Sample ECE275A Midterm Exam Questions

MATH 323 Linear Algebra Lecture 12: Basis of a vector space (continued). Rank and nullity of a matrix.

EE731 Lecture Notes: Matrix Computations for Signal Processing

Elementary Linear Algebra Review for Exam 2 Exam is Monday, November 16th.

Review of some mathematical tools

I. Multiple Choice Questions (Answer any eight)

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Linear Algebra (Review) Volker Tresp 2018

Conceptual Questions for Review

Linear Algebra- Final Exam Review

The QR Factorization

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45

Linear Algebra, part 3 QR and SVD

Linear Algebra in Actuarial Science: Slides to the lecture

Chapter 6 - Orthogonality

Summary of Week 9 B = then A A =

Notes on Eigenvalues, Singular Values and QR

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra

MATH 22A: LINEAR ALGEBRA Chapter 4

7. Symmetric Matrices and Quadratic Forms

MATH 2360 REVIEW PROBLEMS

The Singular Value Decomposition and Least Squares Problems

ANSWERS. E k E 2 E 1 A = B

COMP 558 lecture 18 Nov. 15, 2010

Linear Algebra, part 3. Going back to least squares. Mathematical Models, Analysis and Simulation = 0. a T 1 e. a T n e. Anna-Karin Tornberg

MA 265 FINAL EXAM Fall 2012

MAT Linear Algebra Collection of sample exams

18.06 Problem Set 10 - Solutions Due Thursday, 29 November 2007 at 4 pm in

Final Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

MATH. 20F SAMPLE FINAL (WINTER 2010)

Deep Learning Book Notes Chapter 2: Linear Algebra

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

EE731 Lecture Notes: Matrix Computations for Signal Processing

Linear Algebra Fundamentals

Math 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections

Linear Algebra Review. Vectors

1. Select the unique answer (choice) for each problem. Write only the answer.

Least squares problems Linear Algebra with Computer Science Application

Section 6.4. The Gram Schmidt Process

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.

Linear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions

18.06 Professor Johnson Quiz 1 October 3, 2007

Linear Systems. Carlo Tomasi

Linear Algebra Review. Fei-Fei Li

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

MATH36001 Generalized Inverses and the SVD 2015

235 Final exam review questions

Lecture II: Linear Algebra Revisited

2. Every linear system with the same number of equations as unknowns has a unique solution.

1 Linearity and Linear Systems

Maths for Signals and Systems Linear Algebra in Engineering

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam June 8, 2012

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

NATIONAL UNIVERSITY OF SINGAPORE MA1101R

EECS 275 Matrix Computation

Transcription:

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 1/1 Lecture 7 ECE 275A Pseudoinverse & Moore-Penrose Conditions

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 2/1 R(A) and R(A ) are Linearly Isomorphic Consider the linear mapping A : X Y restricted to be a mapping from R(A ) to R(A), A : R(A ) R(A). The restricted mapping A : R(A ) R(A) is onto. For all ŷ R(A) there exists ˆx = A + y R(A ) such that Aˆx = ŷ. The restricted mapping A : R(A ) R(A) is one-to-one. Let ˆx R(A ) and ˆx R(A ) both map to ŷ R(A), ŷ = Aˆx = Aˆx. Note that ˆx ˆx R(A ) while at the same time A(ˆx ˆx ) = 0. Therefore ˆx ˆx R(A ) N(A), yielding ˆx ˆx = 0. Thus ˆx = ˆx. Since all of the elements of R(A ) and R(A) are in one-to-one correspondence, these subspaces must be isomorphic as sets (and therefore have the same cardinality).

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 3/1 R(A) and R(A ) are Linearly Isomorphic Cont. The restricted mapping A : R(A ) R(A) is a linear isomorphism. Note that the restricted mapping A : R(A ) R(A) is linear, and therefore it preserves linear combinations in the sense that A(α 1ˆx 1 + + α lˆx l ) = α 1 Aˆx 1 + + α l Aˆx l R(A) Furthermore A maps bases in R(A ) to bases in R(A). This can be proved by showing that any spanning set of vectors in R(A ) maps to a spanning set of vectors in R(A) and that any linearly independent set of vectors in R(A ) maps to linearly independent set of vectors in R(A). Thus the dimension (number of basis vectors in) R(A ) must be the same as the dimension (number of basis vectors) in R(A). Since the restricted mapping A is an isomorphism that preserves the vector space properties of its domain, span, linear independence, and dimension, we say that it is a linear isomorphism. Summarizing: R(A ) and R(A) are Linearly Isomorphic, R(A ) = R(A) r(a ) = dim (R(A )) = dim (R(A)) = r(a)

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 4/1 R(A) and R(A ) are Linearly Isomorphic Cont. The relationship between ŷ = Aˆx R(A) and ˆx = A + ŷ R(A ) ˆx A A + ŷ is one-to-one in both mapping directions. I.e., every ˆx = Aŷ R(A ) maps to the unique element ŷ = Aˆx R(A), and vice versa. Therefore when A and A + are restricted to be mappings between the subspaces R(A) and R(A ) they are inverses of each other. A : R(A ) R(A) and A + : R(A) R(A )

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 5/1 Pseudoinverse & Orthogonal Projections Let A : X Y, with dim(x) = n and dim(y) = m. For any y Y, compute ˆx = A + y We have ŷ = P R(A) y (ŷ is the least-squares estimate of y) = Aˆx (ˆx is a least-squares solution ) = AA + y or Therefore P R(A) y = AA + y, y Y P R(A) = AA + and P R (A) = I AA+

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 6/1 Pseudoinverse & Orthogonal Projections Cont. For any x X, compute ŷ = Ax (note that ŷ R(A) by construction) Then Now let ŷ = P R(A) ŷ = Ax ˆx = A + ŷ = A + A x Then, since ŷ = Aˆx, 0 = A(x ˆx) x ˆx N(A) = R(A ) x ˆx R(A ) ˆx = P R(A )x (by the orthogonality principle) A + A x = P R(A )x Since this is true for all x X, we have P R(A ) = A + A and P R (A ) = I A+ A

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 7/1 Properties of AA + and A + A Having shown that AA + = P R(A) and A + A = P R(A ), we now know that AA + and A + A must satisfy the properties of being orthogonal projection operators. In particular AA + and A + A must be self-adjoint, I. (AA + ) = AA + II. (A + A) = A + A These are the first two of the four Moore-Penrose (M-P) Pseudoinverse Conditions. AA + and A + A must also be idempotent, yielding AA + AA + = AA + and A + AA + A = A + A where both of these conditions are consequences of either of the remaining two M-P conditions, III. AA + A = A IV. A + AA + = A +

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 8/1 Four M-P P-Inv Conditions M-P THEOREM: (To be proven in the next lecture.) Consider a linear operator A : X Y. A linear operator M : Y X is the unique pseudoinverse of A, M = A +, if and only if it satisfies the Four M-P Conditions: I. (AM) = AM II. (MA) = MA III. AMA = A IV. MAM = M Thus one can test any possible candidate p-inv using the M-P conditions. Example 1: Pseudoinverse of a scalar α α + = { 1 α α 0 0 α = 0 Example 2: For general linear operators A, B, and C for which the composite mapping ABC is well-defined we have (ABC) + C + B + A + as C + B + A + in general does not satisfy the M-P conditions to be a p-inv of ABC.

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 9/1 Four M-P P-Inv Conditions Cont. Example 3: In example 2, suppose the spaces have the standard Cartesian metric. Furthermore, assume that A and C are unitary, A = A 1, C = C 1,. For complex spaces this means that A 1 = A = A H, whereas for real spaces this means that A and B are orthogonal, A 1 = A = A T. Then (ABC) + = C B + A = C 1 B + A 1 = C + B + A + as can be verified by showing that C B + A satisfies the M-P conditions to be a p-inv for ABC. Note that with the additional assumptions on A and C. we can now claim that (ABC) + = C + B + A +. Example 4: Suppose that Σ is a block matrix m n matrix with block entries Σ = ( ) S 0 1 0 2 0 = ( ) S 0 0 0 where S is square and the remaining block matrices have only entries with value 0. Then ( ) ( ) Σ + S + 0 T = 2 S + 0 0 T 1 0 T = 0 0

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 10/1 Four M-P P-Inv Conditions Cont. Example 5: Let A be a complex m n matrix mapping between X = C n and Y = C m where both spaces have the standard Cartesian inner product. We shall see that A can be factored as ( ) ( ) ( ) A = UΣV H S 0 V = 1 H U 1 U 2 0 0 V2 H = U 1 SV H 1 This factorization is known as the Singular Value Decomposition (SVD). The matrices have the following dimensions: U is m m, Σ is m n, V is n n. The matrix S is a real diagonal matrix of dimension r r where r = r(a). Its diagonal entries, σ i, i = 1,, r, are all real, greater than zero, and are called the singular values of A. The singular values are usually ordered in descending value σ 1 σ 2 σ r 1 σ r > 0 Note that S + = S 1 = diag 1 (σ 1 σ r ) = diag( 1 σ 1 1 σ r )

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 11/1 Four M-P P-Inv Conditions Cont. Example 5 Cont. The matrix U is unitary, U 1 = U H, and its columns form an orthonormal basis for the codomain Y = C m wrt to the standard inner product. (Its rows are also orthonormal.) If we denote the columns of U, known as the left singular vectors, by u i, i = 1,, m, the first r lsv s comprise the columns of the m r matrix U 1, while the remaining lsv s comprise the columns of the m µ matrix U 2, where µ = m r is the dimension of the nullspace of A. The lsv u i is in one-to-one correspondence with the singular value σ i for i = 1,, r. The matrix V is unitary, V 1 = U H, and its columns form an orthonormal basis for the domain X = C n wrt to the standard inner product. (Its rows are also orthonormal.) If we denote the columns of V, known as the right singular vectors, by v i, i = 1,, m, the first r rsv s comprise the columns of the n r matrix V 1, while the remaining rsv s comprise the columns of the n ν matrix V 2, where ν = n r is the nullity (dimension of the nullspace) of A. The rsv v i is in one-to-one correspondence with the singular value σ i for i = 1,, r. We have A = U 1 SV H 1 = (u 1 u r ) σ1... v H 1... = σ 1u 1 v H 1 + + σ r u r v H r σ r v H r

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 12/1 Four M-P P-Inv Conditions Cont. Example 5 Cont. Using the results derived in Examples 3 and 4, we have = A + = V Σ + U H (from example 3) ( ) ( ) S + 0 ( V 1 V 2 U H 0 0 1 U H 2 ) (from example 4) = V 1 S 1 U H 1 (from invertibility of S) = (v 1 v r ) 1 σ 1... = 1 σ 1 v 1 u H 1 + + 1 σ r v r u H r 1 σ r u H 1... u H r Note that this construction works regardless of the value of the rank r. This shows that knowledge of the SVD of a matrix A allows us to determine its p-inv even in the rank-deficient case.

ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 13/1 Taking Stock Our remaining goals before we finish the segment on inverse problems: Prove the 4 M-P Conditions Derive and Understand the SVD Relationship to QR-factorization and Gram-Schmidt process Least-Squares Via Completing the Square Nonlinear Least-Squares Gradient Descent Methods