Pseudoinverse and Adjoint Operators

Similar documents
Normed & Inner Product Vector Spaces

Pseudoinverse & Moore-Penrose Conditions

Moore-Penrose Conditions & SVD

Pseudoinverse & Orthogonal Projection Operators

Vector Space Concepts

ECE 275A Homework #3 Solutions

Lecture 13: Orthogonal projections and least squares (Section ) Thang Huynh, UC San Diego 2/9/2018

Sample ECE275A Midterm Exam Questions

ECE 275A Homework # 3 Due Thursday 10/27/2016

Vector Derivatives and the Gradient

Lecture II: Linear Algebra Revisited

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra

DS-GA 1002 Lecture notes 10 November 23, Linear models

2. Review of Linear Algebra

Generalized Gradient Descent Algorithms

Chapter 6 - Orthogonality

. = V c = V [x]v (5.1) c 1. c k

Fourier and Wavelet Signal Processing

Linear Algebra Review. Vectors

MATH 304 Linear Algebra Lecture 18: Orthogonal projection (continued). Least squares problems. Normed vector spaces.

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

COMP 558 lecture 18 Nov. 15, 2010

Lecture notes: Applied linear algebra Part 1. Version 2

The Four Fundamental Subspaces

5. Orthogonal matrices

Optimization Theory. Linear Operators and Adjoints

MATH 304 Linear Algebra Lecture 19: Least squares problems (continued). Norms and inner products.

Singular Value Decomposition

Linear Algebra Fundamentals

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

APPLIED LINEAR ALGEBRA

Mathematical foundations - linear algebra

Linear Algebra in A Nutshell

Linear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions

Math 3191 Applied Linear Algebra

INF-SUP CONDITION FOR OPERATOR EQUATIONS

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra

Linear Systems. Carlo Tomasi

A Review of Linear Algebra

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Least squares problems Linear Algebra with Computer Science Application

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016

EE731 Lecture Notes: Matrix Computations for Signal Processing

Solutions to Review Problems for Chapter 6 ( ), 7.1

EECS 275 Matrix Computation

Designing Information Devices and Systems II

Overview. Motivation for the inner product. Question. Definition

Linear Least Square Problems Dr.-Ing. Sudchai Boonto

Lecture 5 Least-squares

MATH 350: Introduction to Computational Mathematics

Vector Spaces, Orthogonality, and Linear Least Squares

ORTHOGONALITY AND LEAST-SQUARES [CHAP. 6]

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Review of Some Concepts from Linear Algebra: Part 2

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

Analysis Preliminary Exam Workshop: Hilbert Spaces

Lecture 1: Review of linear algebra

Lecture 2: Tikhonov-Regularization

Linear Algebra- Final Exam Review

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Least squares: the big idea

Algorithmen zur digitalen Bildverarbeitung I

Lecture 3: Linear Algebra Review, Part II

Inner products and Norms. Inner product of 2 vectors. Inner product of 2 vectors x and y in R n : x 1 y 1 + x 2 y x n y n in R n

Lecture 02 Linear Algebra Basics

1 Linearity and Linear Systems

October 25, 2013 INNER PRODUCT SPACES

LEAST SQUARES SOLUTION TRICKS

Math 2331 Linear Algebra

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6

Linear Analysis Lecture 5

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing

There are six more problems on the next two pages

Linear Algebra Massoud Malek

Linear Algebra using Dirac Notation: Pt. 2

Math 2331 Linear Algebra

6.241 Dynamic Systems and Control

Review problems for MA 54, Fall 2004.

Example Linear Algebra Competency Test

Linear Algebra and Robot Modeling

Applied Linear Algebra in Geoscience Using MATLAB

Inner product spaces. Layers of structure:

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Least Squares. Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Winter UCSD

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Orthogonality and Least Squares

Elementary linear algebra

A SHORT SUMMARY OF VECTOR SPACES AND MATRICES

Department of Aerospace Engineering AE602 Mathematics for Aerospace Engineers Assignment No. 4

Math Linear Algebra

Chapter 6: Orthogonality

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.

A Spectral Characterization of Closed Range Operators 1

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Math 307 Learning Goals

Transcription:

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 1/1 Lecture 5 ECE 275A Pseudoinverse and Adjoint Operators

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 2/1 Generalized Solns to Ill-Posed Linear Inverse Problems Taking the linear operator A to be a mapping between Hilbert spaces, we can obtain a generalized least-squares solution to an ill-posed linear inverse problem Ax = y by looking for the unique solution to the regularized least-squares problem: min x y Ax 2 + β x 2, β > 0 where the indicated norms are the inner product induced norms on the domain and codomain. The solution to this problem, ˆxβ, is a function of the regularization parameter β. The choice of the precise value of the regularization parameter β is often a nontrivial problem. The unique limiting solution ˆx lim β 0 ˆx β, is a minimum norm least-squares solution, aka pseudoinverse solution The operator A + which maps y to the solution, ˆx = A + y is called the pseudoinverse of A. The pseudoinverse A + is a linear operator. In the special case when A is square and full-rank, it must be the case that A + = A 1 showing that the pseudoinverse is a generalized inverse.

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 3/1 The Pseudoinverse Solution The pseudoinverse solution, ˆx, is the unique least-squares solution to the linear inverse problem having minimum norm among all least-squares solutions to the least squares problem of minimizing e 2 = y Ax 2, { ˆx = arg min x x x arg min x y Ax 2} Thus the pseudoinverse solution is a least-squares solution. Because Ax R(A), x, any particular least-squares solution, x, to the inverse problem y = Ax yields a value ŷ = Ax which is the unique least-squares approximation of y in the subspace R(A) Y, ŷ = P R(A) y = Ax As discussed above, the orthogonality condition determines the least-squares approximation ŷ = Ax from the geometric condition Geometric Condition for a Least-Squares Solution: e = y Ax R(A) The pseudoinverse solution has the smallest norm, x, among all vectors x that satisfy the orthogonality condition.

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 4/1 The Pseudoinverse Solution Cont. Because X = N(A) N(A) we can write any particular least-squares solution, x arg min x y Ax 2 as x = (P N(A) + P N(A) ) x = P N(A) x + P N(A) x = ˆx + x null, Note that ŷ = Ax = A (ˆx + x null) = Aˆx + Ax null = Aˆx. ˆx N(A) is unique. I.e., independent of the particular choice of x. The least squares solution ˆx is the unique minimum norm least-squares solution. This is true because the Pythagorean theorem yields x 2 = ˆx 2 + x null 2 ˆx 2, showing that ˆx is indeed the minimum norm least-squares solution. Thus the geometric condition that a least-squares solution x is also a minimum norm solution is that x N(A), or equivalently that Geometric Condition for a Minimum Norm LS Solution: x N(A)

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 5/1 Geometric Conditions for a Pseudoinverse Solution 1. Geometric Condition for a Least-Squares Solution: e = y Ax R(A) 2. Geometric Condition for a Minimum Norm LS Solution: x N(A) The primary condition (1) ensures that x is a least-squares solution to the inverse problem y = Ax. The secondary condition (2) then ensures that x is the unique minimum norm least squares solution, x = ˆx = A + y. We want to move from the insightful geometric conditions to equivalent algebraic conditions that will allow us to solve for the pseudoinverse solution ˆx. To do this we introduce the concept of the Adjoint operator, A, of a linear operator A.

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 6/1 The Adjoint Operator - Motivation Given a linear operator A : X Y mapping between two Hilbert spaces X and Y, suppose that we can find a companion linear operator M that maps in the reverse direction M : Y X such that 1. N(M) = R(A) 2. R(M) = N(A) Then the two geometric conditions for a least-squares solution become M(y Ax) = 0 MA x = M y and x = Mλ x Mλ = 0 for some λ Y which we can write as ( ) ( ) MA 0 x I M λ = ( ) My 0 which, if the coefficient matrix is full rank, can be jointly solved for the pseudoinverse solution x = ˆx and the nuisance parameter λ. The companion linear operator M having properties (1) and (2) above exists, is unique, allows for the unique determination of ˆx and λ and is known as the Adjoint Operator, A, of A.

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 7/1 Existence of the Adjoint Operator Given A : X Y define A = M by My, x = y, Ax x X y Y Uniqueness: Suppose M and M both satisfy the above condition. Then My, x = M y, x = x, y My M y, x = 0 x, y (M M )y, x = 0 x, y (M M )y, (M M )y = 0 y (M M )y 2 = 0 y (M M )y = 0 y M M = 0 M = M

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 8/1 Existence of the Adjoint Operator Cont. Linearity: For any α 1, α 2, y 1, y 2, and for all x, we have M (α 1 y 1 + α 2 y 2 ), x = α 1 y 1 + α 2 y 2, Ax = ᾱ 1 y 1, Ax + ᾱ 2 y 2, Ax = ᾱ 1 My 1, x + ᾱ 2 My 2, x = α 1 My 1 + α 2 My 2, x M (α 1 y 1 + α 2 y 2 ) (α 1 My 1 + α 2 My 2 ), x = 0 Existence: Typically shown by construction. For example take X = C n with inner product x 1, x 2 = x H 1 Ωx 2 for hermitian, positive-definite Ω and Y = C m with inner product y 1, y 2 = y H 1 Wy 2 for hermitian, positive-definite W. Assume the standard column-vector representation for all vectors. Then My, x = y H M H Ωx and y, Ax = y H WAx = y H (WAΩ 1 ) Ωx }{{} M H showing that M = (WAΩ 1 ) H M = Ω 1 A H W

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 9/1 Existence of the Adjoint Operator Cont. Proof of Property (1): N(M) = R(A). Recall that two sets are equal iff they contain the same elements. We have y R(A) y, Ax = 0, My, x = 0, x x My = 0 (prove this last step) y N(M) showing that R(A) = N(M).

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 10/1 Existence of the Adjoint Operator Cont. Proof of Property (2): R(M) = N(A). Note that R(M) = N(A) iff R(M) = N(A) We have x R(M) My, x = 0, y, Ax = 0, y y Ax = 0 (prove this last step) x N(A) showing that R(M) = N(A).

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 11/1 Algebraic Conditions for the Pseudoinverse Solution We have transformed the geometric conditions (1) and (2) for obtaining the minimum norm least squares solution to the linear inverse problem y = Ax into the corresponding algebraic conditions 1. Algebraic Condition for an LS Solution The Normal Equation: A Ax = A y 2. Algebraic Condition for a Minimum Norm LS Solution: x = A λ where A : Y X, the adjoint of A : X Y, is given by Definition of the Adjoint Operator, A : A y, x = y, Ax x, y When the (finite dimensional) domain has a metric matrix Ω and the (finite dimensional) codomain has metric matrix W, then (assuming the standard column vector coordinate representation) adjoint operator is given by A = Ω 1 A H W which is a type of generalized transpose of A.

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 12/1 Solving for the Pseudoinverse Solution Note that the normal equation is always consistent by construction A e = A (y ŷ) = A (y Ax) = 0 (consistent) A Ax = A y Similarly, one can always enforce the minimum norm condition x = A λ by an appropriate projection of n onto R(A ) = N(A). The combination of these two equations ( ) ( ) A A 0 x I A λ = ( ) A y 0 can always be solved uniquely for ˆx and λ, as we shall come to understand. Recall that the process of solving for ˆx given a measurement y is described as an action, or operation, on y by the so-called pseudoinverse operator A +, ˆx = A + y. Because A is a linear operator (as was shown above) and the product of any two linear operators is also a linear operator (as can be easily proved), the above system of equations is linear. Thus the solution of ˆx depends linearly on y and therefore the pseudoinverse A + is a linear operator.

ECE 275AB Lecture 5 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 13/1 Solving for the Pseudoinverse Solution - Cont. In general, solving for the pseudoinverse operator A + or solution ˆx requires nontrivial numerical machinery (such as the utilization of the singular value decomposition (SVD)). However, there are two special cases for which the pseudoinverse equations given above have a straightforward solution. These are When A is onto When A is one-to-one