ECE 275A Homework #3 Solutions

Similar documents
ECE 275A Homework # 3 Due Thursday 10/27/2016

Pseudoinverse & Moore-Penrose Conditions

Pseudoinverse & Orthogonal Projection Operators

Moore-Penrose Conditions & SVD

Sample ECE275A Midterm Exam Questions

Pseudoinverse and Adjoint Operators

Normed & Inner Product Vector Spaces

Lecture notes: Applied linear algebra Part 1. Version 2

Vector Space Concepts

Review problems for MA 54, Fall 2004.

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

EE731 Lecture Notes: Matrix Computations for Signal Processing

MIT Final Exam Solutions, Spring 2017

2. Review of Linear Algebra

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

A Brief Outline of Math 355

COMP 558 lecture 18 Nov. 15, 2010

MATH36001 Generalized Inverses and the SVD 2015

EE731 Lecture Notes: Matrix Computations for Signal Processing

Linear Algebra Review. Vectors

Stat 159/259: Linear Algebra Notes

Problem # Max points possible Actual score Total 120

. = V c = V [x]v (5.1) c 1. c k

Deep Learning Book Notes Chapter 2: Linear Algebra

MAT Linear Algebra Collection of sample exams

1. Select the unique answer (choice) for each problem. Write only the answer.

Further Mathematical Methods (Linear Algebra) 2002

A Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Maths for Signals and Systems Linear Algebra in Engineering

Lecture II: Linear Algebra Revisited

Proposition 42. Let M be an m n matrix. Then (32) N (M M)=N (M) (33) R(MM )=R(M)

Properties of Matrices and Operations on Matrices

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

UNIT 6: The singular value decomposition.

APPLIED LINEAR ALGEBRA

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra

The Singular Value Decomposition

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

Chapter 6: Orthogonality

Elementary linear algebra

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Math Fall Final Exam

Chapter 3 Transformations

Linear Systems. Carlo Tomasi

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Linear Algebra Highlights

Matrices A brief introduction

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

Solutions to Review Problems for Chapter 6 ( ), 7.1

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

B553 Lecture 5: Matrix Algebra Review

Large Scale Data Analysis Using Deep Learning

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

7. Symmetric Matrices and Quadratic Forms

Math Linear Algebra Final Exam Review Sheet

ECE 275A Homework 6 Solutions

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

Linear Algebra Primer

5. Orthogonal matrices

Chapter 3. Matrices. 3.1 Matrices

LinGloss. A glossary of linear algebra

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Matrices A brief introduction

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

Seminar on Linear Algebra

Linear Algebra Formulas. Ben Lee

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Math Linear Algebra II. 1. Inner Products and Norms

MAT 2037 LINEAR ALGEBRA I web:

Linear Algebra Fundamentals

Introduction to Matrix Algebra

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

MTH 5102 Linear Algebra Practice Final Exam April 26, 2016

Functional Analysis Review

DS-GA 1002 Lecture notes 10 November 23, Linear models

MATH 532: Linear Algebra

Conceptual Questions for Review

MATH 581D FINAL EXAM Autumn December 12, 2016

2. Every linear system with the same number of equations as unknowns has a unique solution.

Chapter 4 Euclid Space

Typical Problem: Compute.

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Linear Models Review

Singular Value Decomposition (SVD)

Lecture Notes in Linear Algebra

Transcription:

ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y = Ax. This yields Ax = 0 Ax = 0. Proof of (b). Let A be m n (i.e., be a mapping from an n-dimensional space to an m-dimensional space) and r(a) = dim R(A). Recall that r(a) = r(a ). We have A is onto r(a) = r(a ) = m dim N (A ) = m r(a ) = 0 A is one-to-one. Applying this result, we have that A is onto if and only if A = A is one-to-one. Proof of (c). We have, y N (A ) A y = 0 A y, x = 0, x (from part (a)) y, Ax = 0, x y R(A). Proof of (d). Let A be m n. Then A is n m, A A is n n, and AA is m m. First note that it is obvious that N (A) N (A A) as x N (A) Ax = 0 A Ax = 0 x N (A A). On the other hand, x N (A A) A Ax = 0 ξ, A Ax = 0, ξ (from part (a)) Aξ, Ax = 0, ξ Ax, Ax = Ax = 0 (in particular take ξ = x) Ax = 0 x N (A) showing that N (A A) N (A). It is therefore the case that indeed N (A A) = N (A) thereby ensuring that ν(a) = ν(a A). Thus A A is invertible (has rank n) if and only

if A is one-to-one (has rank n). (And which is true if and only if A is onto from part (b) above.) If we apply the above result to the n m operator M = A, we obtain N (AA ) = N (A ). Thus AA is invertible if and only if A is one-to-one which is true if and only if A is onto, from part (b) above. Proof of (e). Most generally, let us assume that the space, X, is a vector space (not necessarily a normed or inner product space). Consider two (complete) disjoint subspaces, V W = {0} such that X = V + W. We call two such subspaces V and W complementary subspaces. It is the case that every vector, x, in X can be uniquely represented as x = v + w where v V X and w W X. Furthermore, the component vectors v and w are linearly independent. Note that v and w remain well defined vectors in X. The projection of X onto V along a complementary space W, is defined by the mapping, x = v + w v, for all x X. Because the decomposition of x into components along V and W is unique for every x in X, this mapping is obviously unique. Regardless of what we call this mapping, say P or Q, the mapping always has the same domain, codomain, and value v = P x = Qx for all x in the domain X. By the definition of a function, operator, or mapping, if any two mappings P and Q have the same domain and codomain and P x = Qx for all x then it must be the case that P = Q. Let P be the projection operator defined, as above, by the mapping x = v + w v for arbitrary x X with unique decomposition v and w along a pair of specified complementary subspaces V and W. We have that for all vectors x X, P x = P (P x) = P v = v = P x and therefore P = P, showing that a projection operator P is idempotent. By induction, we have more generally that P n = P for all integers n. Essentially this says that P acts like the identity operator when acting on its range space (i.e., on the subspace, V, that it projects onto.) Thus, the first time P is applied, it annihilates any component, w W, of x that may exist along the complementary space W, and thereafter acts like the identity operator on the remnant component v V. Proof of (f). Let P be an orthogonal projection operator on a Hilbert space X.,3 Recall that P and I P project onto complementary subspaces V = R(P ) and Can you prove this? Recall that this means that the complementary subspaces V and W associated with P are orthogonal complements with respect to the inner product on X. 3 As a projection operator P must be idempotent.

W = R(I P ). Because these complementary spaces are assumed to be orthogonal complements, W = V, we have that for any pair of arbitrary vectors x, ξ X and therefore 4 0 = (I P ) x, P ξ = P (I P ) x, ξ 0 = P (I P ) = P P P P = P P showing that P must be self adjoint as well as idempotent. On the other hand, if P is idempotent and self-adjoint we have P = P = P P = P P and we can reverse the above steps to show that the projections (I P )x and P ξ are orthogonal for all x and ξ so that the complementary subspaces associated with P and I P, V = R(P ) and W = R(I P ), are orthogonal complements. Proof of (g). This is actually geometrically obvious from the fact that the nullspaces of A and A are both trivial. However, we will give a more mundane proof. The matrix A is square iff A is square. Furthermore r(a ) = r(a). Therefore if A is square and full rank (and hence invertible) then A is square and full rank (and hence invertible). Since A (being square) has full column rank and full row rank, we have A + = A (AA ) = (A A) A. With A and A both invertible, it is easily shown using either of the above expressions for A + that A + = A. Proof of (h). From the definition of the adjoint operator it can readily be shown that A = Ω A H W. (a) Suppose A is onto. Then (in the manner used to show invertibility of AA in problem a above) it can be shown that AΩ A H is invertible (or just temporarily pretend it is the case that W = I, so that A = Ω A H, and apply the result of problem a). This results in, A + = A (AA ) = Ω A H W ( AΩ A H W ) = Ω A H W W ( AΩ A H) = Ω A H ( AΩ A H), which is independent of the weighting matrix W. 4 Why? On an exam I could ask you to justify this step. 3

(b) Suppose A is one-to-one. Then it can be shown that A H W A is invertible. This yields, which is independent of Ω. A + = (A A) A = ( Ω A H W A ) Ω A H W = ( A H W A ) ΩΩ A H W = ( A H W A ) A H W, Proof of (i). Part (a). Since x p and x p are both assumed to any two particular least-squares solutions to the inverse problem y = A x, we have ŷ P R(A) y = A x p = A x p so that A x p = A x p Define ˆx = P R(A )x p and ˆx = P R(A )x p and note that ˆx ˆx R(A ) Also note that x p ˆx N (A) = R(A ) and x p ˆx N (A) = R(A ) or A(x p ˆx) = 0 and A(x p ˆx ) = 0 We have A x p = A x p A (ˆx + (x p ˆx)) = A (ˆx + (x p ˆx ) ) A ˆx = A ˆx A(ˆx ˆx ) = 0 ˆx ˆx = 0 where the last step follows because ˆx ˆx R(A ) N (A) = {0}. Thus it must be the case that ˆx = ˆx as claimed. Part (b). By assumption ˆx = A + y. Let ŷ = P R(A) be the least squares estimate of y. Then Aˆx = ŷ Suppose x p {ˆx} + N (A). Then x p must be of the form x p = ˆx + x 0 with x 0 N (A). Obviously then A x p = A(ˆx + x 0 ) = A ˆx = ŷ 4

showing that x p is a particular least-squares solution. This shows that {ˆx} + N (A) is a subset of the set of all least-squares solutions. On the other hand, suppose x p is any particular least-squares solution, ŷ = A x p. Then we can write x p = ˆx + (x p ˆx) = ˆx + (x p P R(A )x p ) (from the ˆx uniqueness result proven above) = ˆx + x 0 where x 0 = x p P R(A )x p N (A) This shows that the set of all least-squares solution must be a subset of {ˆx} + N (A) Therefore we have shown that the set of all particular least-squares solutions must be precisely the set {ˆx} + N (A).. Solution to Part (a). The matrices in V inherit the properties of vector addition and scalar multiplication from the parent space X. It is easily shown from linearity of the transpose operator that for all scalars α, β C and symmetric n n complex matrices A and B, (α A + β B) T = α A T + β B T = α A + β B showing closure. Hence that V is a subspace. Because to uniquely and unambiguously specify any symmetric complex matrix one must specify the n elements on the diagonal and the n n elements above (or, equivalently, below) the diagonal, the dimension of the space V is n + n n = n(n + ) The set of hermitian matrices is not a subspace as (α A) H = α A H = α A α A for arbitrary scalar α and hermitian matrix A. Solution to Part (b). First we show that P is a linear operator, 5 P (α A + β B) = (α A + β B) + (α A + β B)T = α A + AT + β B + BT = α P (A) + β P (B). 5 An operator is a projection operator onto its range (which is necessarily a subspace of its domain) if and only if it is linear and idempotent. Therefore to prove that P is a projection operator onto V we have to show that P is a) linear; b) idempotent; and c) has range equal to V. Two nonlinear symmetrizing operators are P (A) A + AH and P (A) A + (A ) T The first definition is idempotent and produces Hermitian symmetric matrices, but is nonlinear and is therefore not a projection operator. The second definition is an obviously nonlinear symmetrizing operator which is not indempotent. 5.

for all scalars α, β C and A, B C n n. Now we determine the range of P. Obviously (by the definition of P ) On the other hand since P v = v for all v V. Thus, since we have P (X ) = R(P ) V P (V) = V V X P (V) P (X ), V = P (V) P (X ) = R(P ) V showing that the operator P is onto V, R(P ) = V. Now we prove that P is idempotent, showing that P is a projection operator onto R(P ) = V. For any n n complex matrix A, we have ( ) ( ) T ( ) A+A A + A T T A+A + T P (P (A)) = P = = A + AT = P (A) demonstrating that P is indeed idempotent. Finally, to prove that P : X X is an orthogonal projection operator, we show that it is self-adjoint, P = P. 6 To prove that P is self-adjoint we must show that A, P (B) = P (A), B for any square complex matrices A, B X. From the definition of the Frobenius inner product on the domain and range, this condition is equivalent to Identity: tr A H P (B) = tr P (A) H B. () The proof of the useful identity () is straightforward: ( ) B + B tr A H P (B) = tr A H T = ( tr A H B + tr A H B ) T = ( tr A H B + tr B T A H) = ( tr A H B + tr A B ) ( ) A + A T H = tr B = tr P (A) H B. 6 Note that we cannot claim the existence of an adjoint, much less inquire into its properties, until we have first proved that P is a linear operator. 6

Comments on the Solution to Part (b). I. The property that P is self-adjoint, P = P, is equivalent to the orthogonality condition that (I P ) A, P B = 0 () for all matrices A, B X. 7 Therefore one can show that P is self-adjoint by proving the equivalent orthogonality condition (). Note that it is not enough to prove the weaker condition (I P ) A, P A = 0, which is a necessary, but not sufficient, condition for the orthogonality condition () to hold. 8 II. One might be tempted to treat the range space V as the codomain of P, and then determine the adjoint of this codomain-restricted operator. However, to do this is to treat V as a separate Hilbert space which is not a subspace of X (so that the resulting operator is not a projector). Remember, every projection operator has codomain equal to its domain and projects onto its range which is a subspace of the domain/codomain. Solution to Part (c). The orthogonal projector onto the orthogonally complementary subspace V is given by I P. If we call P a symmetrizing operator, then the complementary orthogonal projection operator I P can be considered an skewsymmetrizing operator. This is because it is readily determined that (I P )A = A AT so that the resulting matrix B = (I P )A is skew-symmetric B T = B. Thus the subspace of symmetric matrices and the subspace of skew-symmetric matrices are orthogonal complements with respect to the Frobenius inner product. That symmetric and skew-symmetric matrices are orthogonal to each other with respect to the Frobenious inner product can be directly shown. Let A be symmetric, A T = A, let B be skew symmetric, B T = B, and note that therefore A H = A. We have showing that tr A H B = tr AB = tr BA = tr A H B A, B = tr A H B = 0. 7 Cf. the proof of self-adjointness of an orthogonal projection operator given in lecture. 8 A simple drawing should convince you of this fact. 7

3. Solution to Part (a). It is evident that A( ) is a linear operator. Because the matrix A is onto, so that the matrix AA H is invertible, we have 0 = A(X) 0 = X A 0 = X AA H 0 = X showing that A( ) is one-to-one. Thus a unique least squares solution exists and is the unique solution to the normal equations A (A(X)) = A (Y ). (3) It remains to determine the adjoint A ( ) in order to complete the determination of the normal equations. 9 The adjoint is determined from the condition that Y, A(X) = A (Y ), X for all Y Y and X X. Note that X is necessarily symmetric, so that where X = P (X) P (A) = A + AT was considered in the previous homework problem. We have Y, A(X) = tr Y H A(X) = tr Y H XA = tr A Y H X = tr A Y H P (X) = tr ( Y A H) H P (X) = tr [ P ( Y A H)] H X (from Identity () of the previous problem) = A (Y ), X showing that A (Y ) = P (Y A H ). (4) 9 Recall that the action of the adjoint must produce an output in Y, A ( ) : Y X. Therefore A (Y ) must be a symmetric matrix. If you did not get a symmetric form for A (Y ) in your solution, you should have known that you had done something wrong. 8

The normal equations then become which is the desired solution. Solution to Part (b). A (A(X)) = A (Y ) A (XA) = A (Y ) P (XAA H ) = P (Y A H ) X ( AA H) + ( AA H) T X = Y A H + ( Y A H) T (i) Term-by-term manipulation of the power series expansion e At = I + At +! A t + 3! A3 t 3 + readily shows the validity of the claimed properties. (ii) The theoretical solution to the Lyapunov equation X = 0 e M T t Λ e M t dt (5) XM + M T X = Λ exists because the eigenvalues of M = AA H are all real (since M is hermitian) and strictly greater than zero (because A onto implies that M is positive definite, not just positive semidefinite). Note that because Λ is symmetric, the theoretical solution X is symmetric. 0 To show that the theoretical solution satisfies the Lyapunov equation is straightforward. For X given by (5) we have XM + M T X = 0 = ) (e M T t Λ e M t M + M T e M T t Λ e M t dt 0 ) d (e M T t Λ e M t = Λ lim t e M T t Λ e M t = Λ where the last step follows because the eigenvalues of M (and therefore M T ) are strictly positive. 0 The integral of a matrix function is defined component wise. Thus the integral of a transpose is the transpose of the integral. 9

4. For the two linear operators B and C we are given that R(C) N (B) = {0}. We have BC = 0 (BC) (x) = 0 x B (C(x)) = 0 x C(x) = 0 x [Because C(x) R(C) N (B)] C = 0. 5. Proof of (a). The goal is to show that the candidate matrix M = (A + ) is the pseudoinverse of A, which is accomplished by showing that M satisfies the four Moore- Penrose pseudoinverse conditions: I. (MA ) = AM = AA + = (AA + ) = (A + ) A = MA. II. (A M) = M A = A + A = (A + A) = A (A + ) = A M. III. A MA = A (A + ) A = (AA + A) = (A) = A. IV. MA M = (A + ) A (A + ) = (A + AA + ) = (A + ) = M. Note that in the proof we exploit the fact that A + is assumed a priori to be a pseudoinverse and therefore itself satisfies the four Moore-Penrose conditions. Proof of (b). Let M = A (AA ) + and N = (A A) + A. To proved that these are both pseudoinverses will require some work. Using the identity (A + ) = (A ) + proved in Part (a), it is easy to show that, ( (A A) +) = (A A) + ( and (AA ) +) = (AA ) +, two facts that we will use repeatedly below. In addition, we will also use the extremely useful Theorem A, which was proved in the previous problem. Theorem A Let the nullspace of B and the range of C be disjoint. Then C = 0 iff BC = 0. Finally, note that (AA ) + and (A A) + used in the definition of the candidate pseudoinverses M and M are both pseudoinverses by assumption and therefore both satisfy the four Moore-Penrose pseudoinverse conditions (MPCs). To prove that M is a pseudoinverse, we show that it satisfies the MPC s. (AM) = ( AA (AA ) +) = AA (AA ) + (Property of (AA ) + ) = AM. 0

(MA) = ( A (AA ) + A ) = A ( (AA ) +) A = A (AA ) + A = MA. MAM = A (AA ) + AA (AA ) + = A (AA ) + (Property of (AA ) + ) = M. Because (AA ) + is a pseudoinverse it satisfies, AA (AA ) + AA = AA which we rewrite as, A ( A (AA ) + AA A ) = 0. Applying Theorem A results in, A (AA ) + AA = A, which (by taking the adjoint of both sides) is readily shown to be equivalent to the final MPC, AMA = A. We have now proved that M is a pseudoinverse. To prove that N is a pseudoinverse, we show that it also satisfies the MPC s. (NA) = ( (A A) + A A ) = (A A) + A A (Property of (A ) + A) = NA. (AN) = ( A(A A) + A ) = A ( (A A) +) A = A(A A) + A = AN. NAN = (A A) + A A(A A) + A = (A A) + A (Property of (A A) + ) = N. Because (A A) + is a pseudoinverse it satisfies, A A(AA ) + A A = A A which we rewrite as, A ( A(A A) + A A A ) = 0.

Applying Theorem A results in, A(A A) + A A = A, which is just the final MPC, ANA = N. Proof of (c). One can readily show that, { α + for α 0 = α 0 for α = 0, satisfies all four MPC s. Proof of (d). One can readily show that, satisfies all four MPC s. D + = diag (d +,, d + n ), Proof of (e). One can readily show that, ( ) S Σ + = + 0, 0 0 satisfies all four MPC s, where the shown block zero matrices of Σ + are identical to the block zero matrices of Σ T. Proof of (f). Let M = C B + A be a candidate pseudoinverse. One can readily ascertain that M satisfies all four MPC s and therefore indeed M = (ABC) +. Proof of (g). We have, A + = ( UΣV H) + = V ΣU H (From (f) above) ( ) + S 0 = V U H 0 0 ( ) S = V + 0 U H (From (e) above) 0 0 ( ) ( ) S = (V V ) + 0 U H = V 0 0 S + U H = V S U H U H = σ v u H + + σ r v r u H r.

6. Note that although the SVD is highly constrained, it is not necessarily unique. We exploit the fact that the SVD of any real matrix, M, can be partitioned as, ) S. 0 V M = U Σ V T = (U H. U, 0. 0 V H where the real matrices U and V are orthogonal and the r = rank (M) nonzero elements of the real r r diagonal matrix S are the (unique) nonzero singular values of M. Obtaining any factorization of a real matrix M such that the factorization has these properties is equivalent to providing an SVD of M. For the matrices shown, we can perform this factorization by inspection and geometric reasoning alone. We do not have to perform an eigenvector and eigenvalue analysis of MM T and M T M, although this is often done for general, arbitrary matrices. (Remember, ours are hand crafted for ease of factorization.) Although the SVD itself may not be unique, the four orthogonal projectors onto the fundamental subspaces of A, the pseudoinverse of A, and the singular values of A all are unique (why?). Having the SVD factorization at hand, we know the rank of A, the dimensions of the four fundamental subspaces, and orthonormal bases for each of the four subspaces. And we can readily construct the pseudoinverse and the orthogonal projectors onto each of thee four fundamental subspaces. (See the first problem below for a detailed discussion. The subsequent solutions only give an SVD factorization and assume that you can perform the required matrix multiplications to obtain the remaining requested quantities.) Matrix A is obviously full rank, r = n =, and can be decomposed as, ( ) ( ). 0 A = = (). 0 0. 0 We immediately see that (nonzero) V is nonexistant, showing that the nullspace of A, N (A), is trivial (ν = n r = 0), which is consistent ( with ) the obvious fact that the matrix is full rank, r = n =. The column U = is an orthonormal basis for 0 ( ) 0 the r = dimension range of A, R(A). The column U = is a orthonormal basis for the m r = = dimensional nullspace of A T, N (A T ). The (scalar) column V = () is an orthonormal basis for the r = dimensional range of A T, R(A T ). Nonunqueness of the SVD shows up in the choice of signs used when defining the orthonormalized columns of U and V (the signs just have to be chosen in a consistent manner). We obtain, ( ) ( ) 0 0 0 P R(A) = I P N (A T ) = U U T =, P 0 0 N (A T ) = I P R(A) = U U T = 0 P R(A T ) = I P N (A) = V V T = (), P N (A) = I P R(A T ) = (0). 3

Note that more than one way exists to construct the projection operators. This fact can be used to check for errors. Finally, the pseudoinverse is obtained as, A + = V T Σ + U = V T S U = () ( 0) = ( 0). Matrix B is obviously full rank, r = n =, and can be decomposed as, 0 0. B = 0 = 0. 0 0 0. 0 0 0 0 ( ) 0. 0 Nonuniqueness here shows up in the fact that the sign of the last column of U can be reversed. From this factorization, one can determine the subspace dimensions. Matrix C is obviously rank deficient, r = < min(m, n) = 3. One possible SVD is, 0 0 0. 0. 0 0 0 C = 0 0 = 0. 0 0. 0 0 0 0 0. 0. 0 0 0 0. 0 Matrix D is obviously full rank, r = m =. An SVD is given by, 0 0 D = (0 3 0) = () (3. 0 0) 0 0. 0 0 Nonuniqueness shows up here not just in the choice of sign. For instance, equally good SVD s are, 0 0 D = ( ) (3. 0 0) 0 0, 0 0 and 0 0 D = () (3. 0 0) 0. 0 There are, in fact, an infinity of possibilities. (Can you categorize all of them?) 4

7. Kay 8.3. Note that Kay interchangeably uses x[n] and s[n] to mean the same sampled signal (this is likely a typo). Assume the signal model s[n] = A + Bn, where A and B are real scalars, and determine a least squares fit of this model to empirical data {s[n], n = 0,, N }. We have s[0] 0 s[] s[] =... s[n ] N which we write in vector-matrix form as, s = Hθ. ( ) A, B Note that H is full-rank, so we can determine the least-squares estimates as, θ = (H T H) H T s One determines that (using some standard integer summation formulas), N s[n] N(N ) H T n=0 N s = N, HT H =, N(N ) N(N )(N ) n s[n] 6 n=0 and (using the standard formula for inverting matrices), ( H T H ) (N ) 6 N(N+) N(N+) =. 6 N(N+) N(N ) This yields equation (8.4) of Moon (remember that s[n] and x[n] are the same quantity). 5