THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION
|
|
- Cameron Shields
- 5 years ago
- Views:
Transcription
1 THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION MANTAS MAŽEIKA Abstract. The purpose of this paper is to present a largely self-contained proof of the singular value decomposition (SVD), and to explore its application to the low rank approximation problem. We begin by proving background concepts used throughout the paper. We then develop the SVD by way of the polar decomposition. Finally, we show that the SVD can be used to achieve the best low rank approximation of a matrix with respect to a large family of norms. Contents 1. Background Concepts Self-adjoint Linear Maps The Spectral Theorem Direct Sums & Eigenspaces Singular Values 4 2. The Polar Decomposition The Polar Decomposition for Endomorphisms The General Polar Decomposition 8 3. The Singular Value Decomposition The SVD for Endomorphisms The General SVD 9 4. The Low Rank Approximation Problem Matrix Norms & Their Generalizations The Min-max Theorem & Weyl s Inequality for Singular Values Low Rank Approximation Discussion 16 Acknowledgments 16 Bibliography Guide 16 References 16 Date: September 3. Contact: mantas@uchicago.edu. 1
2 2 MANTAS MAŽEIKA 1. Background Concepts Notation 1.1. Unless otherwise stated, throughout this section U and V refer to finite-dimensional Hermitian inner product spaces. Notation 1.2. Suppose we have bases u = {u 1, u 2,..., u n } and v = {v 1, v 2,..., v m } of U and V respectively. Given a vector x U, we denote the coordinate vector of x with respect to u by x 1 x 2 [x] u :=.. x n For a linear map f : U V, we denote the matrix of f with respect to u and v by [f] u,v := [f(u 1 )] v [f(u 2 )] v... [f(u 2 )] v Self-adjoint Linear Maps. Theorem 1.3. For all linear maps f : U V, there exists a unique g : V U such that x U, y V : f(x), y U = x, g(y) V. Proof. Let f : U V be a linear map, let u = {u 1, u 2,..., u n } be an orthonormal basis for U, and let v = {v 1, v 2,..., v n } be an orthonormal basis for V. Let x U, and let y V. With an asterisk denoting the conjugate transpose of a matrix, we have f(x), y V = ([f] u,v [x] u ) [y] v = [x] u[f] u,v[y] v. We want to find a matrix G, which can be interpreted as a linear map with respect to u and v, such that [x] u(g[y] v ) = [x] u[f] u,v[y] v. Recall that, for matrices G and F, if Gx = F x for all vectors x, then we have G = F. Therefore, G must equal [f] u,v. The unique linear map g is equal to the matrix G interpreted as a linear map with respect to the bases u and v. For readability, we will omit subscripts on inner products in the rest of this paper. Definition 1.4. Let f : U V be a linear map. We define the adjoint of f as the unique linear map f : V U satisfying x U, y V : f(x), y = x, f (y). If f = f, we say f is self-adjoint. Note that in this case, U and V must be the same. Additionally, when f : V V is self-adjoint and V is a Hermitian inner product space, we say f is Hermitian. We will now state some properties of the adjoint operator, which can be proven by converting to matrices and proving the properties for the conjugate transpose. Proposition 1.5. Let f : U V, g : U V, and h: V W be linear maps. Then we have (f + g) = f + g. (f h) = h f. We will also make use of the following property of self-adjoint operators. Proposition 1.6. If f : V V is self-adjoint, then it has real eigenvalues. Proof. Let λ be an eigenvalue of f, and let v be an associated eigenvector. λ v, v = f(v), v = v, f(v) = v, v λ Since v 0, we have v, v > 0, so λ = λ. Therefore, λ R.
3 THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION The Spectral Theorem. Theorem 1.7. If f : V V is Hermitian, then f has an orthonormal eigenbasis. That is, there exists an orthonormal basis of V consisting of eigenvectors of f. We denote this basis u = {u 1, u 2,..., u n }, indexed such that the associated eigenvalues are ordered from greatest to least: λ 1 λ 2 λ n. Proposition 1.8. If f : V V has an orthonormal eigenbasis and all the eigenvalues of f are real, then f is Hermitian. Proof. Let u = {u 1, u 2,..., u n } be the orthonormal eigenbasis. Then [f] u,u is a real diagonal matrix, so its conjugate transpose [f] u,u equals [f] u,u. This is only the case if f = f Direct Sums & Eigenspaces. Definition 1.9. For a vector space A and subsets A 1, A 2,..., A n A, we define the sum of the A i as { n n } A i := a i a 1 A 1, a 2 A 2,..., a n A n. Definition Let A be a vector space over a field F, and let A 1, A 2,..., A n be subspaces of A. We say that A is a direct sum of the A i if all x A can be written uniquely as a sum n a i, where each a i is an element of the corresponding A i. To denote that A is a direct sum of the A i, we write n A = A i. The following are a few well-known properties of the direct sum. For their proofs, see [2]. Propositions (1) A vector space A is a direct sum of A 1, A 2,..., A k A if we have A i A j = {0} and A = A j. j i (2) If A is a direct sum of A 1, A 2,..., A n A, then dim A = n dim A i. (3) If B is a subspace of a Hermitian inner product space A, then B B = A. We will now prove a few small lemmas that will be used in proving a lemma to the polar decomposition. Proposition Let C be a Hermitian inner product space, and let A, B C such that A B = C and A B. Then, A = B. Proof. Let x C such that x, y = 0 for all y B. We know x = a + b for a A and b B. Suppose for contradiction b 0. Then, we have x, b = a, b + b, b = 0 + v, v > 0. This is a contradiction, so b = 0, and we know x A. Therefore, A = B. Proposition Let f : V V be a Hermitian linear map, and let E i denote the eigenspace for the eigenvalue λ i. Then, E i E j for j i. Proof. Let x E i and let y E j with j i. Then we have x, y λ j = x, f(y) = f(x), y = λ i x, y = λ i x, y. This gives us (λ i λ j ) x, y = 0. Since λ i λ j, we know x, y = 0. Therefore x E j, and E i E j. Proposition Let f : V V be a Hermitian linear map, where dim(v ) = n. Then V is a direct sum of the eigenspaces of f.
4 4 MANTAS MAŽEIKA Proof. We know f has k eigenspaces, which are perpendicular to each other, and whose dimensions sum to n. Let i [k]. Any nonzero element of j i E j represents a sum of eigenvectors with coefficient 1. These eigenvectors and any v E i are linearly independent, so their sum cannot equal v. Therefore, E i j i E j = {0}. We can find orthonormal bases for each eigenspace separately. Put together, these bases will form a set of cardinality n of orthonormal vectors in V. Therefore, they must be an orthonormal basis of V, which tells us that V is the sum of the eigenspaces of f. By Proposition we know V is a direct sum of the eigenspaces of f. Though we only require the previous two propositions for Hermitian linear maps, they also apply to normal linear maps, for which f f = f f. Proposition Let f : V V and g : V V be Hermitian linear maps. Suppose {x U f(x) = λx} is an eigenspace of f if and only if {x U g(x) = λx} is an eigenspace of g. Then f = g. Proof. By the spectral theorem, f has an orthonormal eigenbasis u := {u 1, u 2,..., u n } that uniquely defines the action of f on V. By the premises, these are eigenvectors of g to the same eigenvalues. Since these vectors form a basis, they uniquely define the action of g on U as the same of f. Therefore, f = g Singular Values. Proposition For any linear map f : U V, the compositions f f and f f self-adjoint. are Proof. Let f : U V be a linear map. For all u U and v V, we have (f f)(u), v = f (f(u)), v = f(u), f(v) = u, f (f(v)) = u, (f f)(v). Thus, f f is self-adjoint. Showing that f f is self-adjoint follows analogously. Proposition If f : U V is a linear map, then f f and f f have non-negative, real eigenvalues. Proof. Since f f and f f are Hermitian, they have real eigenvalues. Let λ be an eigenvalue of f f, and let v be an associated eigenvector. We have v, v λ = v, λv = v, (f f)(v) = f(v), f(v). Since the Hermitian inner product is positive definite, we know f(v), f(v) is non-negative, and v, v is positive, because v is not equal to zero by definition. Therefore, we can divide by v, v, and we know λ must be non-negative. The proof for f f is analogous. Note that this is equivalent to the statement that f f and f f are positive semi-definite. In other words, the inner product v, (f f)(v) is real and non-negative for all v U, and likewise for f f.
5 THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION 5 Lemma If f : U V is a linear map, then we have ker(f) = ker(f f) ker(f ) = ker(f f ) ker(f) = (im(f )) ker(f ) = (im(f)) rk(f f) = rk(f) = rk(f ) = rk(f f ). Proof. Let f : U V be a linear map, with dim(u) = n and dim(v ) = m. First, we will show ker(f) = ker(f f). Let x ker(f). Then, f (f(x)) = f (x) = 0, so x ker(f f). Therefore, ker(f) ker(f f). Let x ker(f f). We have f(x), f(x) = (f f)(x), x = 0, x = 0. Since the inner product is positive definite, we know f(x) = 0, so ker(f f) ker(f). Therefore, ker(f) = ker(f f). The proof that ker(f ) equals ker(f f ) follows analogously. To see ker(f) = (im(f )), let x ker(f) and let y V. We have x, f (y) = f(x), y = 0, y = 0, so x f (y) for all y V. Therefore, x (im(f )). Now let x (im(f )). Since f (f(x)) im(f ), we have 0 = x, (f f)(x) = f(x), f(x), which is only the case when f(x) = 0, so x ker(f). Therefore, ker(f) = (im(f )). The proof that ker(f ) equals (im(f)) follows analogously. The nullity of f and f f are equal, and likewise for f and f f. By the rank-nullity theorem, we know rk(f) = rk(f f) and rk(f ) = rk(f f ). All that remains is to show rk(f) = rk(f ). We proved (ker(f)) = im(f ), and by Proposition we have ker(f) (ker(f)) = U. Therefore, by Proposition we have dim(ker(f)) + dim(im(f )) = n. By the rank-nullity theorem, dim(im(f)) = dim(im(f )), or rk(f) = rk(f ). The following theorem is an important prerequisite for the definition of singular values. Theorem If f : U V is a linear map of rank r, then there are exactly r nonzero eigenvalues of f f, and r nonzero eigenvalues of f f. Proof. By Lemma 1.18, we know f and f f have rank r. By the rank nullity theorem, the dimension of ker(f) and ker(f f) is n r. Since the kernel of a function is equal to its eigenspace associated with the eigenvalue 0, we know the algebraic multiplicity of the eigenvalue 0 for f f is n r. Since f f is Hermitian, it is diagonalizable, so its eigenvalues have the same algebraic and geometric multiplicity. Therefore, there are r nonzero, and hence positive, eigenvalues of f f. The same follows by analogous reasoning for f f. Definition The singular values of a linear map f : U V are the strictly positive square roots of the eigenvalues of f f, denoted as σ 1 σ 2 σ r > 0. Remark The eigenvalues of f f and f f are in fact equal, with the same multiplicity. We will see a proof of this later in the paper with Proposition The Polar Decomposition In this section, we will prove the polar decomposition for linear maps between Euclidean spaces of equal dimension. The generalization to arbitrary linear maps is proved analogously, but introduces complexities unnecessary for a conceptual understanding of the topic. It is stated without proof at the end of this section. The structure of the proofs in this section, of Lemma 1.18, and of the proofs in section three, comes from [3].
6 6 MANTAS MAŽEIKA Lemma 2.1. Let V be a Hermitian inner product space of dimension n. Let f : V V be a positive semi-definite, self-adjoint linear map. Then, there exists a unique, positive semi-definite, self-adjoint linear map h: V V such that h h equals f. Furthermore, if µ 1, µ 2,..., µ p are the distinct eigenvalues of h, and E i is the eigenspace associated with µ i, then µ 2 1, µ 2 2,..., µ 2 p are the distinct eigenvalues of f, and E i is the eigenspace associated with µ 2 i. Finally, ker(f) = ker(h). Proof. By the spectral theorem, f has an orthonormal eigenbasis u := {u 1, u 2,..., u n } with associated eigenvalues λ 1, λ 2,..., λ n. Since f is positive semi-definite, its eigenvalues are non-negative, which lets us define σ i as the positive square root λ i. We now define a linear map h: V V as mapping each u i to σ i u i. This gives us h h = f, and that the squared eigenvalues of h equal the eigenvalues of f, so h is positive semi-definite. What s more, h has an orthonormal eigenbasis by its definition, and since the eigenvalues of f are real, the eigenvalues of h are real. By Proposition 1.8, h is self-adjoint. Since u is an orthonormal eigenbasis to both f and h, we know σ i has the same multiplicity as λ i for each i. Therefore, if µ 1, µ 2,..., µ p are the distinct eigenvalues of h, then µ 2 1, µ 2 2,..., µ 2 p are the distinct eigenvalues of f. We now consider the eigenspaces E i := {x V h(x) = µ i v} E i := {x V f(x) = µ2 i v}. Fix k [p]. To see E k E k, let x E k. We have h(x) = µ i v, so (h h)(x) = µ 2 i x. Since f = h h, this gives us x E i. For the other inclusion, since h is Hermitian, we know by the propositions in 1.11 E i E k = V and (E k ) E k = V. i k By Proposition 1.12, we have (E k ) = E i. i k Thus, we want to show E k E i. i k Fix x E k, and let y be in the direct sum of the E i for i k. We have y = i k y i for y i E i. Since we have already proven E i E i, we know y i E i. Therefore, we have x, y = x, i k y i = i k x, y i = i k 0 = 0 (E k E i ) Therefore, x is perpendicular to y, so x is perpendicular to the direct sum of the E i for i k. In other words, x is in the orthogonal complement of this direct sum, which is just E k. Therefore, E k = E k. Clearly, ker(h) ker(f). To see the other inclusion, let x ker(f) such that x 0. Then x must be an eigenvector of f to eigenvalue 0. As we just proved, it is also an eigenvector of h to eigenvalue 0, so x ker(h). Therefore, ker(h) = ker(f). All that remains is to show uniqueness. Suppose there exists a self-adjoint linear map g : V V such that g g = f. By the spectral theorem, g has an orthonormal eigenbasis. By the same argument that was used with h, we know g and f have the same eigenspaces, so g and h must have the same eigenspaces. By Proposition 1.15, g = h.
7 THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION The Polar Decomposition for Endomorphisms. Theorem 2.2 (The Polar Decomposition for Endomorphisms). Let V be an n-dimensional Hermitian inner product space. For any linear map f : V V, there are two positive semi-definite, self-adjoint linear maps h 1 : V V and h 2 : V V, and a unitary linear map g : V V such that f = g h 1 = h 2 g. Furthermore, if f has rank r, the maps h 1 and h 2 have the same positive eigenvalues σ 1, σ 2,..., σ r, which are the singular values of f. Finally, h 1 and h 2 are unique, g is unique if f is invertible, and h 1 = h 2 if f is normal. Proof. The composition f f is a positive semi-definite, self-adjoint linear map, so by Lemma 2.1 there exists a unique positive semi-definite, self-adjoint linear map h 1 : V V satisfying h 1 h 1 = f f. In the lemma, we also specified the existence of an orthonormal eigenbasis u := {u 1, u 2,..., u n } for h 1. Moreover, the indices are arranged such that µ 1 µ 2 µ n, where µ i is the eigenvalue associated with u i. Let r be the rank of f. By Theorem 1.19, we know µ 1, µ 2,..., µ r are the singular values of f, and µ r+1, µ r+2,..., µ n are all zero. We now wish to find a g such that f = g h 1. A hint for where to start comes if we suppose h 1 is invertible. Then, we expect f h 1 1 to equal g. Thinking back to how h 1 acts on the u i, we see that h 1 1 (u i) would equal u i /µ i, for positive µ i. Therefore, we expect g(u i ) to equal f(u i )/µ i for positive µ i. Taking advantage of this knowledge, we can define a new orthonormal basis of V. For i [r], we let v i = f(u i )/µ i. These vectors are orthonormal already. To see this, let i, j [r]. v i, v j = 1 µ i µ j f(u i ), f(u j ) = 1 µ i µ j (f f)(u i ), u j = 1 µ i µ j µ 2 i u i, u j = µ i µ j δ ij = δ ij. Therefore, we can extend the v i to an orthonormal basis of V. Let g be defined as the linear map sending u i to v i. Since g maps one orthonormal basis to another, it is unitary. Defining g in this way also leads to the desired equality f = g h 1. To see this, let i [r]. (g h)(u i ) = g(µ i u i ) = µ i g(u i ) = µ i v i = f(u i ) For r < i n, we have f(u i ) 2 = f(u i ), f(u i ) = (f f)(u i ), u i = 0, u i = 0, so f(u i ) must equal the zero vector. This gives us (g h)(u i ) = g(0) = 0 = f(u i ). Therefore (g h)(u i ) = f(u i ) for all i [n], and therefore for all vectors in V. Now we will turn to h 2. Because g is unitary, it must be invertible. Defining h 2 as f g 1, we obtain f = g h 1 = h 2 g. Therefore, we have g h 1 g 1 = h 2. The adjoint of g h 1 g 1 is equal to itself, so h 2 is self-adjoint. Also, from the definition of v i, we know that for i [r], we have f(u i ) = µ i v i. For r < i n, we have h 2 (v i ) = f(u i ) = 0 = µ i v i. Therefore, h 1 and h 2 have the same positive eigenvalues, which are the singular values of f. To see that h 2 is unique, suppose there is a positive semi-definite, self-adjoint linear map h 3 such that f = h 3 g. We have f g 1 = h 2 and f g 1 = h 3. Therefore, h 2 h 2 = h 2 h 2 = f f, and likewise h 3 h 3 = f f. By Lemma 2.1, we can deduce that h 2 is the unique positive semi-definite, self-adjoint linear map such that h 2 h 2 = f f, and likewise for h 3. Therefore, h 2 = h 3. Finally, if f is invertible, then it has full rank, so its nullity is zero. By Lemma 2.1, we have ker(f) = ker(h 1 ), so the nullity of h 1 is zero, so h 1 has full rank and is invertible. We can then write g = f h 1 1. Since h 1 is unique, we know g is unique. If f is normal, then we have f f = f f, so h 1 h 1 = h 2 h 2. The polar decomposition can be stated in matrix form as follows.
8 8 MANTAS MAŽEIKA Theorem 2.3. For every complex n n matrix A, there is an n n unitary matrix R and an n n positive semi-definite, Hermitian matrix S such that A = RS. Furthermore, R and S are unique if A is invertible The General Polar Decomposition. Let U and V be Hermitian inner product spaces of dimensions n and m respectively. Definition 2.4. A linear map f : U V is weakly unitary if it has rank s = min(m, n), and if f f = id on (ker(f)). Theorem 2.5 (The Polar Decomposition). For any linear map f : U V, there are two positive semi-definite, self-adjoint linear maps h 1 : U U and h 2 : V V, and a weakly unitary linear map g : U V such that f = g h 1 = h 2 g. Furthermore, if f has rank r, the maps h 1 and h 2 have the same positive eigenvalues σ 1, σ 2,..., σ r, which are the singular values of f. Finally, h 1 and h 2 are unique, g is unique if the rank of f is min{m, n}, and h 1 = h 2 if f is normal The SVD for Endomorphisms. 3. The Singular Value Decomposition Lemma 3.1. Let V be an n-dimensional Hermitian inner product space. For any linear map f : V V, there exist two orthonormal bases u = {u 1, u 2,..., u n } and v = {v 1, v 2,..., v n } such that [f] u,v is diagonal, and has the form µ 1 µ 2... µn where the first r of the µ i are equal to σ 1, σ 2,..., σ r, the singular values of f, and the remaining n r of the µ i are zero. Proof. In the proof of the polar decomposition, we specified two orthonormal bases u = {u 1, u 2,..., u n } and v = {v 1, v 2,..., v n }. Let r be the rank of f. Then, for i [r], we have f(u i ) = µ i v i. Expressing this as a coordinate vector with respect to the basis v gives µ i in the i th coordinate, and zero everywhere else. For i greater than r, we have f(u i ) = µ 2 i u i = 0, so the coordinate vector with respect to te basis v is the zero vector. Therefore, we have µ 1 µ 2 [f] u,v =.... µn Theorem 3.2. Let M be an n n complex matrix. There exist two n n unitary matrices U and V, and a positive semi-definite diagonal matrix D such that Furthermore, D is of the form M = UDV. µ 1 µ 2... µn
9 THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION 9 where the first r of the µ i are equal to σ 1, σ 2,..., σ r, the singular values of M, and the remaining n r of the µ i are zero. Proof. Let u and v be defined as before. By considering M as the linear map m with respect to the standard basis e, we see that this becomes a change of bases problem. Namely, we are trying to convert [m] u,v to [m] e,e. This can be obtained by considering the change of basis transformations σ : C n C n and τ : C n C n, defined as σ(u i ) = e i τ(v i ) = e i. The rank of both σ and τ is equal to n, so they are invertible, and we have σ 1 (e i ) = u i τ 1 (e i ) = v i. We see then that [σ 1 ] e,e is the unitary matrix with column vectors, from left to right, [u 1 ] e, [u 2 ] e,..., [u n ] e. Likewise, [τ 1 ] e,e is the unitary matrix with column vectors, from left to right, [v 1 ] e, [v 2 ] e,..., [v n ] e. These matrices are equal to [σ] 1 e,e and [τ] 1 e,e, respectively. Therefore, if we let U := [σ] e,e, and V := [τ] e,e, and D := [m] u,v, we can use change of bases to obtain 3.2. The General SVD. M = [m] e,e = [σ] e,e [m] u,v [τ] 1 e,e = UDV. Lemma 3.3. Let U and V be Hermitian inner product spaces of dimension n and m respectively. Let f : U V be a linear map with rank r, and let s = min(m, n). Then, there exist two orthonormal bases u = {u 1, u 2,..., u n } and v = {v 1, v 2,..., v m } such that [f] u,v is a matrix of the form µ 1 µ 2... µ 1 D 1 = µn µ 2 or D 2 =... 0 µm 0 where the first r of the µ i equal σ 1, σ 2,..., σ r, the singular values of f, and the remaining s r of the µ i are zero. Theorem 3.4. Let M be an m n complex matrix. There exists an m m unitary matrix U, an n n unitary matrix V, and an m n positive semi-definite rectangular diagonal matrix D such that M = UDV. Furthermore, D is of one of the forms specified in 3.3, depending on which of m and n is greater. Before moving on, we will return to proving that f f and f f have the same eigenvalues with the same multiplicity, and hence that f and f have the same singular values with the same multiplicity. Proposition 3.5. Let f : U V be a linear map for Hermitian inner product spaces U and V. Then f f and f f have the same eigenvalues with the same multiplicity. Proof. By Theorem 3.4, we have [f] e,e = UDV, where U is an m m unitary matrix, V is an n n unitary matrix, and D is an m n rectangular diagonal matrix. Let F denote [f] e,e. We have F F = V D U UDV = V D T U 1 UDV 1 = V D T DV 1.
10 10 MANTAS MAŽEIKA Analogously, we have F F = V DD T V 1. In other words, F F is unitarily similar to D T D, and F F is unitarily similar to DD T. Therefore, F F and D T D share the same eigenvalues with the same multiplicity, and likewise with F F and DD T. Since D T D and DD T visibly share the same eigenvalues with the same multiplicity, we can conclude that F F and F F share the same eigenvalues with the same multiplicity. 4. The Low Rank Approximation Problem We will now explore an application of the SVD to low rank approximation of matrices. The goal of low rank approximation is to find a matrix B that is closest to an initial matrix A such that the rank of B is less than or equal to q N. In other words, given a matrix norm and a q N, we wish to minimize A B subject to the constraint rk(b) q. The traditional solution obtained from the SVD, where A = UDV, is to set B equal to B q := UD q V, where the smallest s q entries of D are zeroed out to obtain D q. An interesting question is how exactly B q attains closeness to A. Is it the closest low rank approximation with respect to all matrix norms? If not all matrix norms, then which subset? To start, we can notice that B q is essentially the same as A, but with the smallest s q singular values set to zero. Therefore, we expect low rank approximation to hold for norms for which B A measures closeness of the largest singular values of A and B Matrix Norms & Their Generalizations. Notation 4.1. Unless otherwise stated, throughout this section we let A be an m n complex matrix of rank r, with s := min(m, n). We let σ i (A) denote the i th singular value of A if i [r], and if r < i s, we let σ i (A) equal 0. Additionally, we let λ i (A) denote the i th eigenvalue of A, where λ 1 (A) λ 2 (A) λ s (A). We will use similar notation to write the singular values and eigenvalues of other matrices. Definitions 4.2. For p, k N such that k s, we have the following definitions. Schatten p-norm ( s ) 1 p S A p := σ p i (A) Ky Fan k-norm K A k := σ i (A) Notice that the Schatten p-norm is an L p norm of the singular values, which is a smooth measure of the largest singular values, biasing towards larger singular values as p goes to infinity. On the other hand, the Ky Fan k-norm is a discrete measure of the k largest singular values. In other words, these norms are the norms we would expect to work with B q. The norm for which we will prove the low rank approximation theorem is the natural generalization of the Schatten p-norm and Ky Fan k-norm. 1 Namely, for p, k N such that k s we define A (p,k) := ( σ p i (A) ) 1 p. We will refer to this norm as the (p, k) norm, but we have yet to prove that it is, in fact, a norm. This becomes easier with the help of a few trace identities. Before proving these identities, however, we require some knowledge about sums of eigenvalues. Definition 4.3. Let A be an n n Hermitian matrix, and let k [n] be given. Let V be an m-dimensional subspace of C n. We define the partial trace tr(a V ) as m tr(a V ) := vi Av i, where v 1, v 2,... v m are any orthonormal basis of V. 1 For more information regarding these three norms, see [4] and [8]
11 THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION 11 The partial trace is invariant to the choice of orthonormal basis. This can be shown by writing one orthonormal basis as another orthonormal basis times a unitary matrix, and solving for the partial trace. Therefore, the partial trace is well-defined. Proposition 4.4 (Extremal partial trace). Let A be an n n Hermitian matrix, and let k [n] be given. We have λ i (A) = sup tr(a V ). V C n dim(v )=k A clear proof of the extremal partial trace identity can be found in [7], so we will omit it here. Corollary 4.5 (Schur-Horn inequality). Let A be an n n Hermitian matrix, and let k [n] be given. Let {i 1, i 2,..., i k } [n] and let a i i denote the i th diagonal element of A. We have a i1i 1 + a i2i a ik i k λ 1 (A) + λ 2 (A) + λ k (A). Proof. Let u = {u 1, u 2,..., u n } be an orthonormal eigenbasis of A, and let V C n be defined as the direct sum of k of the spans of the u i. We have tr(a V ) = λ i1 (A) + λ i2 (A) + λ ik (A) for {i 1, i 2,..., i k } [n], which is less than or equal to the supremum of tr(a V ) over all k- dimensional V. Definition 4.6. A projection is a linear transformation ρ: V V such that ρ ρ = ρ. A projection is orthogonal if V has an inner product and the projection s range and null space are orthogonal with respect to this inner product. One can show that a projection is orthogonal if and only if it is self-adjoint. A good proof of this can be found in [6], so we will leave it unproven here. Proposition 4.7. Let A refer to the unique Hermitian matrix such that A 2 = A A. Let P k be the matrix space of rank k orthogonal projections. The following identities hold. S A p = tr( A p ) 1/p (4.8) K A k = max{tr(p A )} P P k (4.9) A (p,k) = max P P k {tr(p A p ) 1/p } (4.10) Proof. First we will prove 4.8. Since A is Hermitian, we have λ i ( A p ) = λ p i ( A ). Hence, n n s tr( A p ) = λ i ( A p ) = λ p i ( A ) = σ p i (A) = S A p p. Now we will prove 4.9. Let ρ be any rank k orthogonal projection. Since ρ is self-adjoint, it has an orthonormal eigenbasis u = {u 1, u 2,..., u n }. Let P := [ρ] e,e and let P := [ρ] u,u. Let A be the matrix obtained by changing the bases of A from e to u. The only eigenvalues of a projection are 0 and 1, where the multiplicity of 1 equals the rank of the projection. Therefore, P is a diagonal matrix with 1 on k diagonals and 0 everywhere else. Therefore, the matrix A P is equal to A, but with n k columns zeroed out. By Corollary 4.5, we can conclude tr( A P ) K A k. By the properties of the trace, we also have tr( A P ) = tr(p A ) = tr(p A ). To see that K A k is obtained by tr(p A ) for some P, let v = {v 1, v 2,..., v n } be an orthonormal eigenbasis of A. Define ρ on this basis by letting ρ(v i ) equal v i for i [k], and 0 otherwise. We can see that ρ is a rank k orthogonal projection. Defining P and A in the same way as before, but on v instead of u, we can see that A P is a diagonal matrix with a trace equal to K A k. The proof of 4.10 is a straightforward combination of the proofs for 4.8 and 4.9. From the previous proposition, one can easily show that the (p, k) norm is positive definite, and that αa (p,k) = α A (p,k) holds for all scalars α C. All that remains is to prove the triangle inequality. First, however, we will introduce a theorem about symmetric gauge functions.
12 12 MANTAS MAŽEIKA In short, symmetric gauge functions are functions on the reals that act like norms. For their full definition, and a proof of the following theorem, see [5]. Notation For x, y R n, if the sum of the components of x is less than or equal to the sum of the components of y, we say x y. Theorem Let x, y R n such that x and y are element-wise non-negative. Then, x y if and only if Φ(x) Φ(y) for all symmetric gauge functions Φ. Proposition For all complex m n matrices A and B, we have A + B (p,k) A (p,k) + B (p,k). Proof. The triangle inequality for the Ky Fan k-norm can be stated as σ 1 (A + B) σ 1 (A) + σ 1 (B) σ 2 (A + B). σ 2 (A) + σ 2 (B).. σ k (A + B) σ k (A) + σ k (B) Since the L p norm is a symmetric gauge function 1, we know by Theorem 4.12 σ 1 (A + B) σ 1 (A) + σ 1 (B) σ 2 (A + B) A + B (p,k) = L p. L σ 2 (A) + σ 2 (B) p.. σ k (A + B) σ k (A) + σ k (B) By Minkowski s inequality, we have σ 1 (A) + σ 1 (B) σ 1 (A) σ 1 (A) σ 2 (A) + σ 2 (B) L p. L σ 2 (A) p. + L σ 2 (A) p. = A (p,k) + B (p,k). σ k (A) + σ k (B) σ k (A) σ k (A) 4.2. The Min-max Theorem & Weyl s Inequality for Singular Values. The proof of the low rank approximation theorem for the (p,k) norm requires a few lemmas. One of them is the singular value version of Weyl s inequality. This, in turn, requires the min-max theorem, also known as the Courant-Fischer min-max theorem. Definition Let A be an n n Hermitian matrix, and let x C n such that x 0. We define the Rayleigh quotient of A as R A (x) := x Ax x x. Note that for a vector x C n and scalar α C, we have R A (αx) = R A (x). Lemma Let V be a finite dimensional vector space, and let X, Y V. We have dim(x + Y ) + dim(x Y ) = dim(x) + dim(y ). For the sake of conciseness, we will not prove this lemma. See [2] for a full proof. Theorem 4.16 (The Min-max Theorem). Let A be an n n Hermitian matrix. Then we have ( (x)) λ i (A) = R A. min U C n dim U=n i+1 max x U x 0 1 This can be seen by comparing the properties of the Lp norm to the definition of symmetric gauge functions provided in [5].
13 THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION 13 Since R A (αx) = R A (x) for any scalar α, an equivalent statement of the theorem is ( Ax) λ i (A) = x. min U C n dim U=n i+1 inf U C n dim U=n i+1 max x U x =1 Proof. First, we will prove the infimum-supremum version of the theorem, starting with ( (x)) λ i (A) R A. sup x U x 0 It suffices to show that for all U C n of dimension n i + 1, there exists an x in U with x 0 such that λ i (A) R A (x). Let U C n have dimension n i + 1. Where E j is the j th eigenspace for A, we have i dim U + dim E j > n. j=1 ( i ) This is because dim E j 1 for all j [n]. Using Lemma 4.15, we can show U j=1 E j is nontrivial. Moreover, the Rayleigh quotient for any x in this intersection is greater than or equal to λ i (A). This completes the first part of the proof. Now we will show λ i (A) inf U C n dim U=n i+1 sup x U x 0 ( R A (x)). It suffices to show the existence of a U C n of dimension n i + 1 such that for all x U with x 0, we have λ i (A) R A (x). Once again, we know n dim n i + 1. j=i E j Let U be a subspace of this direct sum of dimension n i + 1. For all x U with x 0, we have R A (x) λ i (A), completing the infimum-supremum proof. Showing that the supremum is in fact a maximum can be done by noting that the supremum is taken over the unit ball, a compact set, and that the Rayleigh quotient is continuous. By the extreme value theorem, the maximum exists. Showing that the infimum is a minimum can be achieved by noting that the U specified in the proof of the second inequality attains a value for R A (x) with x 0 and x U equal to λ i. By the first inequality, this is less than or equal to to the infimum over all U with dimension n i + 1 of this quantity. Therefore, the U specified in the proof of the second inequality attains a minimum value. The proof of the following theorem is based on the proof of the same theorem in [7]. Theorem 4.17 (Weyl s Inequality for Singular Values). Let A and B be m n complex matrices, and let i and j be in [n] such that i + j 1 is in [n]. Then we have σ i+j 1 (A + B) σ i (A) + σ j (B). Proof. Let A and B be m n complex matrices. σ i (A) is equal to λ i (A A), so by the min-max theorem we have the following for any m n complex matrix C. ( Cx) σi 2 (C) = x C. min U C n dim U=n i+1 max x U x =1
14 14 MANTAS MAŽEIKA Notice that x C Cx equals (Cx) Cx. This is just the inner product Cx, Cx, or Cx 2. Therefore, we have σi 2 (C) = max ( Cx 2). x U x =1 min U C n dim U=n i+1 Since max(x 2 ) = max(x) 2 and min(x 2 ) = min(x) 2 are true for x 0, we can simplify this to ( Cx ) σ i (A) =. (4.18) min U C n dim U=n i+1 max x U x =1 We want to show σ i+j 1 (A+B) σ i (A)+σ j (B), so it suffices to show there exists a W C n of dimension n i j + 2 such that for all unit vectors x W, we have (A + B)x σ i (A) + σ j (B). By 4.18, there exists a U C n of dimension n i + 1 such that the maximum among all unit vectors x U of Ax is equal to σ i (A). Therefore, for all unit vectors x U, we have Ax σ i (A). Likewise, there exists a V C n of dimension n j + 1 such that for all unit vectors x V, we have Bx σ j (B). By Lemma 4.15, we have dim(u + V ) + dim(u V ) = dim(u) + dim(v ). Since U + V C n, we know dim(u + V ) n, so we have n + dim(u V ) 2n i j + 2 dim(u V ) n i j + 2. Furthermore, since i + j 1 n, we have 1 n i j + 2, so we can find W U V, where dim(w ) = n i j + 2. For all unit vectors x W, we have 4.3. Low Rank Approximation. (A + B)x = Ax + Bx Ax + Bx σ i (A) + σ j (B). Lemma Given an m n complex matrix A, we have A (p,k) = A (p,k). Proof. Let A be an m n complex matrix, and let p := min(m, n). From Proposition 3.5, we have σ i (A) = σ i (A ) for i [p]. Since k p, we have ( ) 1 ( p ) 1 p A (p,k) = σ i (A ) p = σ i (A) p = A (p,k). Lemma The norm (p,k) is unitarily invariant. That is, for an m n complex matrix A, an m m unitary matrix S, and a n n unitary matrix T, we have Proof. Given A, S, and T, we have SA p (p,k) = SA (p,k) = A (p,k) = AT (p,k). = σ p i (SA) = λ i (A S SA) p 2 λ i (A A) p 2 = σ i (A) p = A p (p,k). Taking the p th root, we get SA (p,k) = A (p,k).
15 THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION 15 From Lemma 4.19, we know A (p,k) = A (p,k). This lets us deduce AT p (p,k) = T A p (p,k) = σ p i (T A ) = λ p i (AT T A ) = λ p i (AA ) = Taking the p th root, we get AT (p,k) = A (p,k). σ p i (A ) = A p (p,k) = A p (p,k). The strategy of using Weyl s inequality for the following proof comes from [1]. Theorem 4.21 (The Low Rank Approximation Theorem). Let A be an m n complex matrix. By the SVD, we have A = UDV for an m n rectangular diagonal matrix D and unitary matrices U and V. Let D q be the matrix obtained by setting all but the first q diagonal entries of D to zero, and let B q := UD q V. Then, B q is a closest matrix to A of rank at most q. That is, for all m n complex matrices B of rank at most q, we have A B q (p,k) A B (p,k). Proof. Let A be an m n complex matrix, and define B q as in the theorem statement. By Lemma 4.20, we have A B q (p,k) = UDV UD q V (p,k) = U(D D q )V (p,k) = D D q (p,k). The matrix D D q is an m n rectangular diagonal matrix equal to D, except that the first q diagonal entries have been zeroed out. Moreover, since the singular values of D are equal to the singular values of A, we have D D q p (p,k) = σ i (D D q ) p = i=q+1 σ i (D) p = i=q+1 σ i (A) p. By Weyl s inequality for singular values, for compatible X, Y, i, and j we have σ i+j 1 (X + Y ) σ i (X) + σ j (Y ). Let B be an m n complex matrix of rank at most q. We will make the following substitutions. j = q + 1 Y = B X = A B. This gives us σ i+q (A) σ i (A B) + σ q+1 (B). Since rk(b) q, we know σ q+1 (B) = 0, so this simplifies to σ i+q (A) σ i (A B). Now, we complete the proof. k q A B p (p,k) = k q σ i (A B) p σ i (A B) p σ i+q (A) p i=q+1 = σ i (A) p = D D q p (p,k) = A B q p (p,k). Therefore, A B q p (p,k) A B p (p,k).
16 16 MANTAS MAŽEIKA 4.4. Discussion. We have shown that B q is the closest low rank approximation to A with respect to the (p, k) norm. As was mentioned at the start of this section, this result is exactly what we would expect from a norm like the (p, k) norm, which measures largeness of the greatest singular values. Whether or not the (p, k) norm encompasses all norms for which low rank approximation holds with B q is an avenue for further exploration. Acknowledgments. It is a pleasure to thank my mentor, Tung Tho Nguyen, for helping to guide me in writing this paper, and in learning about the SVD and its applications. I am also grateful to Professor László Babai for his lectures in the UChicago Mathematics REU apprentice class, which inspired this paper, and Professor Peter May for the opportunity to participate in this REU. Bibliography Guide. As a computational courtesy to the reader, this section contains some pointers on using the references efficiently. The bullets denote the citation being referred to. [ 1 ] The edit to the original post by Algebraic Pavel is where the Weyl inequality is used. [ 2 ] The Direct Sums section number in this book is 2.3. [ 4 ] The Everything in Between section describes the (p, k) norm, and mentions the norms that it generalizes. [ 5 ] The theorem used in this paper is titled Theorem 1 by Mirsky, and can be found on the second page of the article. [ 6 ] The proof referred to in this paper is in section 2.1 of the Wikipedia article. This citation refers to a specific version of the article, so be sure to view that version. [ 8 ] This reference was useful mainly because it clued me in to using symmetric gauge functions for proving the triangle inequality for the (p, k) norm. It has a lot of other information on Schatten p-norms and Ky Fan k-norms, though, and even proves unitary invariance for the (p, k) norm in section 6.2. References [1] Algebraic Pavel ( Proof of Eckart-Young- Mirsky theorem, URL (version: ): [2] Broida, Joel G., and S. Gill Williamson. Direct Sums. Comprehensive Introduction to Linear Algebra. N.p.: n.p., n.d Print. [3] Gallier, Jean H. Singular Value Decomposition (SVD) and Polar Form. Geometric Methods and Applications: For Computer Science and Engineering. New York: Springer, SpringerLink. Web. 26 Aug [4] Johnston, Nathaniel. Ky Fan Norms, Schatten Norms, and Everything in Between. Nathaniel Johnston. N.p., 21 Aug Web. 25 Aug [5] Mirsky, L. Symmetric Gauge Functions And Unitarily Invariant Norms. The Quarterly Journal of Mathematics Q J Math 11.1 (1960): Web. [6] Projection (linear Algebra). Wikipedia. Wikimedia Foundation, 24 Aug Web. 27 Aug [7] Tao, Terence. 254A, Notes 3a: Eigenvalues and Sums of Hermitian Matrices. Whats New. N.p., 13 Jan Web. 29 Aug [8] Vasuki, Vishvas. SCRIPT OF A TALK ON THE KY FAN AND SCHATTEN P NORMS. Web. 25 Aug
Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationSingular Value Decomposition (SVD) and Polar Form
Chapter 2 Singular Value Decomposition (SVD) and Polar Form 2.1 Polar Form In this chapter, we assume that we are dealing with a real Euclidean space E. Let f: E E be any linear map. In general, it may
More information08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms
(February 24, 2017) 08a. Operators on Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/real/notes 2016-17/08a-ops
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More informationIr O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )
Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O
More informationElementary linear algebra
Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The
More information2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.
Hints for Exercises 1.3. This diagram says that f α = β g. I will prove f injective g injective. You should show g injective f injective. Assume f is injective. Now suppose g(x) = g(y) for some x, y A.
More informationLecture 7: Positive Semidefinite Matrices
Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.
More informationDesigning Information Devices and Systems II
EECS 16B Fall 2016 Designing Information Devices and Systems II Linear Algebra Notes Introduction In this set of notes, we will derive the linear least squares equation, study the properties symmetric
More informationSpectral Theorem for Self-adjoint Linear Operators
Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;
More informationSingular Value Decomposition and Polar Form
Chapter 14 Singular Value Decomposition and Polar Form 14.1 Singular Value Decomposition for Square Matrices Let f : E! E be any linear map, where E is a Euclidean space. In general, it may not be possible
More informationThe University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.
The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment Two Caramanis/Sanghavi Due: Tuesday, Feb. 19, 2013. Computational
More informationMATH 240 Spring, Chapter 1: Linear Equations and Matrices
MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear
More informationA PRIMER ON SESQUILINEAR FORMS
A PRIMER ON SESQUILINEAR FORMS BRIAN OSSERMAN This is an alternative presentation of most of the material from 8., 8.2, 8.3, 8.4, 8.5 and 8.8 of Artin s book. Any terminology (such as sesquilinear form
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationLecture notes: Applied linear algebra Part 1. Version 2
Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationLinear Algebra. Workbook
Linear Algebra Workbook Paul Yiu Department of Mathematics Florida Atlantic University Last Update: November 21 Student: Fall 2011 Checklist Name: A B C D E F F G H I J 1 2 3 4 5 6 7 8 9 10 xxx xxx xxx
More informationThe following definition is fundamental.
1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationMATRICES ARE SIMILAR TO TRIANGULAR MATRICES
MATRICES ARE SIMILAR TO TRIANGULAR MATRICES 1 Complex matrices Recall that the complex numbers are given by a + ib where a and b are real and i is the imaginary unity, ie, i 2 = 1 In what we describe below,
More informationReview of Some Concepts from Linear Algebra: Part 2
Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set
More informationMaths for Signals and Systems Linear Algebra in Engineering
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE
More informationLINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM
LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator
More informationLecture 8: Linear Algebra Background
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationStat 159/259: Linear Algebra Notes
Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationLinear Algebra Lecture Notes-II
Linear Algebra Lecture Notes-II Vikas Bist Department of Mathematics Panjab University, Chandigarh-64 email: bistvikas@gmail.com Last revised on March 5, 8 This text is based on the lectures delivered
More information2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian
FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian
More informationLinear algebra and applications to graphs Part 1
Linear algebra and applications to graphs Part 1 Written up by Mikhail Belkin and Moon Duchin Instructor: Laszlo Babai June 17, 2001 1 Basic Linear Algebra Exercise 1.1 Let V and W be linear subspaces
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationNotes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.
Notes on singular value decomposition for Math 54 Recall that if A is a symmetric n n matrix, then A has real eigenvalues λ 1,, λ n (possibly repeated), and R n has an orthonormal basis v 1,, v n, where
More informationLINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS
LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F has characteristic zero. The following are facts (in
More informationMath 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.
Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More information6 Inner Product Spaces
Lectures 16,17,18 6 Inner Product Spaces 6.1 Basic Definition Parallelogram law, the ability to measure angle between two vectors and in particular, the concept of perpendicularity make the euclidean space
More informationLinGloss. A glossary of linear algebra
LinGloss A glossary of linear algebra Contents: Decompositions Types of Matrices Theorems Other objects? Quasi-triangular A matrix A is quasi-triangular iff it is a triangular matrix except its diagonal
More informationMATH 115A: SAMPLE FINAL SOLUTIONS
MATH A: SAMPLE FINAL SOLUTIONS JOE HUGHES. Let V be the set of all functions f : R R such that f( x) = f(x) for all x R. Show that V is a vector space over R under the usual addition and scalar multiplication
More informationMath 108b: Notes on the Spectral Theorem
Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationMath 408 Advanced Linear Algebra
Math 408 Advanced Linear Algebra Chi-Kwong Li Chapter 4 Hermitian and symmetric matrices Basic properties Theorem Let A M n. The following are equivalent. Remark (a) A is Hermitian, i.e., A = A. (b) x
More informationLinear Algebra Formulas. Ben Lee
Linear Algebra Formulas Ben Lee January 27, 2016 Definitions and Terms Diagonal: Diagonal of matrix A is a collection of entries A ij where i = j. Diagonal Matrix: A matrix (usually square), where entries
More informationLinear Algebra Highlights
Linear Algebra Highlights Chapter 1 A linear equation in n variables is of the form a 1 x 1 + a 2 x 2 + + a n x n. We can have m equations in n variables, a system of linear equations, which we want to
More informationNumerical Linear Algebra
University of Alabama at Birmingham Department of Mathematics Numerical Linear Algebra Lecture Notes for MA 660 (1997 2014) Dr Nikolai Chernov April 2014 Chapter 0 Review of Linear Algebra 0.1 Matrices
More informationMATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.
MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation
More informationReview of some mathematical tools
MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical
More informationMath 4A Notes. Written by Victoria Kala Last updated June 11, 2017
Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...
More informationMATH 431: FIRST MIDTERM. Thursday, October 3, 2013.
MATH 431: FIRST MIDTERM Thursday, October 3, 213. (1) An inner product on the space of matrices. Let V be the vector space of 2 2 real matrices (that is, the algebra Mat 2 (R), but without the mulitiplicative
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationAlgebra II. Paulius Drungilas and Jonas Jankauskas
Algebra II Paulius Drungilas and Jonas Jankauskas Contents 1. Quadratic forms 3 What is quadratic form? 3 Change of variables. 3 Equivalence of quadratic forms. 4 Canonical form. 4 Normal form. 7 Positive
More informationTest 1 Review Problems Spring 2015
Test Review Problems Spring 25 Let T HomV and let S be a subspace of V Define a map τ : V /S V /S by τv + S T v + S Is τ well-defined? If so when is it well-defined? If τ is well-defined is it a homomorphism?
More informationFinal Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2
Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch
More information1. General Vector Spaces
1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA
ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA Kent State University Department of Mathematical Sciences Compiled and Maintained by Donald L. White Version: August 29, 2017 CONTENTS LINEAR ALGEBRA AND
More informationLinear algebra 2. Yoav Zemel. March 1, 2012
Linear algebra 2 Yoav Zemel March 1, 2012 These notes were written by Yoav Zemel. The lecturer, Shmuel Berger, should not be held responsible for any mistake. Any comments are welcome at zamsh7@gmail.com.
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More informationMATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by
MATH 110 - SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER 2009 GSI: SANTIAGO CAÑEZ 1. Given vector spaces V and W, V W is the vector space given by V W = {(v, w) v V and w W }, with addition and scalar
More informationChapter 0 Miscellaneous Preliminaries
EE 520: Topics Compressed Sensing Linear Algebra Review Notes scribed by Kevin Palmowski, Spring 2013, for Namrata Vaswani s course Notes on matrix spark courtesy of Brian Lois More notes added by Namrata
More informationSPECTRAL THEORY EVAN JENKINS
SPECTRAL THEORY EVAN JENKINS Abstract. These are notes from two lectures given in MATH 27200, Basic Functional Analysis, at the University of Chicago in March 2010. The proof of the spectral theorem for
More informationComputational math: Assignment 1
Computational math: Assignment 1 Thanks Ting Gao for her Latex file 11 Let B be a 4 4 matrix to which we apply the following operations: 1double column 1, halve row 3, 3add row 3 to row 1, 4interchange
More informationThen x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r
Practice final solutions. I did not include definitions which you can find in Axler or in the course notes. These solutions are on the terse side, but would be acceptable in the final. However, if you
More informationThen since v is an eigenvector of T, we have (T λi)v = 0. Then
Problem F02.10. Let T be a linear operator on a finite dimensional complex inner product space V such that T T = T T. Show that there is an orthonormal basis of V consisting of eigenvectors of B. Solution.
More informationLecture 2: Linear operators
Lecture 2: Linear operators Rajat Mittal IIT Kanpur The mathematical formulation of Quantum computing requires vector spaces and linear operators So, we need to be comfortable with linear algebra to study
More informationMath 321: Linear Algebra
Math 32: Linear Algebra T. Kapitula Department of Mathematics and Statistics University of New Mexico September 8, 24 Textbook: Linear Algebra,by J. Hefferon E-mail: kapitula@math.unm.edu Prof. Kapitula,
More informationMath Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88
Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant
More informationCategories and Quantum Informatics: Hilbert spaces
Categories and Quantum Informatics: Hilbert spaces Chris Heunen Spring 2018 We introduce our main example category Hilb by recalling in some detail the mathematical formalism that underlies quantum theory:
More informationFinal Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson
Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson Name: TA Name and section: NO CALCULATORS, SHOW ALL WORK, NO OTHER PAPERS ON DESK. There is very little actual work to be done on this exam if
More informationMATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL
MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left
More informationAnalysis Preliminary Exam Workshop: Hilbert Spaces
Analysis Preliminary Exam Workshop: Hilbert Spaces 1. Hilbert spaces A Hilbert space H is a complete real or complex inner product space. Consider complex Hilbert spaces for definiteness. If (, ) : H H
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationLinear Algebra- Final Exam Review
Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.
More informationThe Spectral Theorem for normal linear maps
MAT067 University of California, Davis Winter 2007 The Spectral Theorem for normal linear maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 14, 2007) In this section we come back to the question
More informationDefinitions for Quizzes
Definitions for Quizzes Italicized text (or something close to it) will be given to you. Plain text is (an example of) what you should write as a definition. [Bracketed text will not be given, nor does
More informationVector Spaces and Linear Transformations
Vector Spaces and Linear Transformations Wei Shi, Jinan University 2017.11.1 1 / 18 Definition (Field) A field F = {F, +, } is an algebraic structure formed by a set F, and closed under binary operations
More informationMath113: Linear Algebra. Beifang Chen
Math3: Linear Algebra Beifang Chen Spring 26 Contents Systems of Linear Equations 3 Systems of Linear Equations 3 Linear Systems 3 2 Geometric Interpretation 3 3 Matrices of Linear Systems 4 4 Elementary
More informationLinear Algebra Lecture Notes-I
Linear Algebra Lecture Notes-I Vikas Bist Department of Mathematics Panjab University, Chandigarh-6004 email: bistvikas@gmail.com Last revised on February 9, 208 This text is based on the lectures delivered
More informationFunctional Analysis Review
Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all
More informationGQE ALGEBRA PROBLEMS
GQE ALGEBRA PROBLEMS JAKOB STREIPEL Contents. Eigenthings 2. Norms, Inner Products, Orthogonality, and Such 6 3. Determinants, Inverses, and Linear (In)dependence 4. (Invariant) Subspaces 3 Throughout
More informationChapter 3 Transformations
Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 16: Eigenvalue Problems; Similarity Transformations Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 18 Eigenvalue
More informationOctober 25, 2013 INNER PRODUCT SPACES
October 25, 2013 INNER PRODUCT SPACES RODICA D. COSTIN Contents 1. Inner product 2 1.1. Inner product 2 1.2. Inner product spaces 4 2. Orthogonal bases 5 2.1. Existence of an orthogonal basis 7 2.2. Orthogonal
More informationMath 413/513 Chapter 6 (from Friedberg, Insel, & Spence)
Math 413/513 Chapter 6 (from Friedberg, Insel, & Spence) David Glickenstein December 7, 2015 1 Inner product spaces In this chapter, we will only consider the elds R and C. De nition 1 Let V be a vector
More informationEigenvalues and Eigenvectors
/88 Chia-Ping Chen Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Eigenvalue Problem /88 Eigenvalue Equation By definition, the eigenvalue equation for matrix
More informationNOTES ON BILINEAR FORMS
NOTES ON BILINEAR FORMS PARAMESWARAN SANKARAN These notes are intended as a supplement to the talk given by the author at the IMSc Outreach Programme Enriching Collegiate Education-2015. Symmetric bilinear
More informationBASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x
BASIC ALGORITHMS IN LINEAR ALGEBRA STEVEN DALE CUTKOSKY Matrices and Applications of Gaussian Elimination Systems of Equations Suppose that A is an n n matrix with coefficents in a field F, and x = (x,,
More informationis an isomorphism, and V = U W. Proof. Let u 1,..., u m be a basis of U, and add linearly independent
Lecture 4. G-Modules PCMI Summer 2015 Undergraduate Lectures on Flag Varieties Lecture 4. The categories of G-modules, mostly for finite groups, and a recipe for finding every irreducible G-module of a
More informationMath 113 Winter 2013 Prof. Church Midterm Solutions
Math 113 Winter 2013 Prof. Church Midterm Solutions Name: Student ID: Signature: Question 1 (20 points). Let V be a finite-dimensional vector space, and let T L(V, W ). Assume that v 1,..., v n is a basis
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course
More informationLecture notes on Quantum Computing. Chapter 1 Mathematical Background
Lecture notes on Quantum Computing Chapter 1 Mathematical Background Vector states of a quantum system with n physical states are represented by unique vectors in C n, the set of n 1 column vectors 1 For
More informationContents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2
Contents Preface for the Instructor xi Preface for the Student xv Acknowledgments xvii 1 Vector Spaces 1 1.A R n and C n 2 Complex Numbers 2 Lists 5 F n 6 Digression on Fields 10 Exercises 1.A 11 1.B Definition
More informationRecall the convention that, for us, all vectors are column vectors.
Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists
More informationEigenvalues and eigenvectors
Chapter 6 Eigenvalues and eigenvectors An eigenvalue of a square matrix represents the linear operator as a scaling of the associated eigenvector, and the action of certain matrices on general vectors
More informationReview of Linear Algebra
Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=
More informationAssignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.
Assignment 1 Math 5341 Linear Algebra Review Give complete answers to each of the following questions Show all of your work Note: You might struggle with some of these questions, either because it has
More informationFall TMA4145 Linear Methods. Exercise set Given the matrix 1 2
Norwegian University of Science and Technology Department of Mathematical Sciences TMA445 Linear Methods Fall 07 Exercise set Please justify your answers! The most important part is how you arrive at an
More information