Dimensionality Reduction Using the Sparse Linear Model: Supplementary Material

Similar documents
Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction Using the Sparse Linear Model

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Mathematical foundations - linear algebra

14 Singular Value Decomposition

Linear Subspace Models

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Chapter 3 Transformations

CS 143 Linear Algebra Review

Designing Information Devices and Systems II

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Review of some mathematical tools

Review problems for MA 54, Fall 2004.

Linear Algebra Massoud Malek

Mathematical foundations - linear algebra

CS281 Section 4: Factor Analysis and PCA

When Fisher meets Fukunaga-Koontz: A New Look at Linear Discriminants

Linear Algebra - Part II

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Linear Methods for Regression. Lijun Zhang

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Linear Algebra. Min Yan

B553 Lecture 5: Matrix Algebra Review

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

15 Singular Value Decomposition

The Singular Value Decomposition

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

EECS 275 Matrix Computation

Principal Component Analysis

1 Last time: least-squares problems

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang

CS 231A Section 1: Linear Algebra & Probability Review

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Background Mathematics (2/2) 1. David Barber

1 Singular Value Decomposition and Principal Component

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

LINEAR ALGEBRA REVIEW

I. Multiple Choice Questions (Answer any eight)

Review of Some Concepts from Linear Algebra: Part 2

Principal Component Analysis

Advanced Introduction to Machine Learning CMU-10715

Introduction to Machine Learning

Lecture Notes 1: Vector spaces

Lecture 3: Review of Linear Algebra

The following definition is fundamental.

MA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2

Lecture 3: Review of Linear Algebra

Kernel Methods. Machine Learning A W VO

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis

Lecture Notes 2: Matrices

Large Scale Data Analysis Using Deep Learning

Stat 159/259: Linear Algebra Notes

LECTURE NOTE #11 PROF. ALAN YUILLE

Fantope Regularization in Metric Learning

Maximum variance formulation

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Functional Analysis Review

2. Every linear system with the same number of equations as unknowns has a unique solution.

DS-GA 1002 Lecture notes 10 November 23, Linear models

2.3. Clustering or vector quantization 57

2. Review of Linear Algebra

Lecture 7: Positive Semidefinite Matrices

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Bastian Steder

Machine Learning - MT & 14. PCA and MDS

. = V c = V [x]v (5.1) c 1. c k

Lecture 2: Linear operators

Kernel Method: Data Analysis with Positive Definite Kernels

Singular Value Decomposition and Principal Component Analysis (PCA) I

Unsupervised Learning: Dimensionality Reduction

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Review of Linear Algebra

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Linear Algebra Review. Vectors

Statistical Pattern Recognition

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Basic Calculus Review

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Solution of Linear Equations

Machine learning for pervasive systems Classification in high-dimensional spaces

Principal Component Analysis. Applied Multivariate Statistics Spring 2012

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent

The Matrix Representation of a Three-Dimensional Rotation Revisited

5 Linear Algebra and Inverse Problem

This appendix provides a very basic introduction to linear algebra concepts.

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag

Lecture 7 Spectral methods

MAT Linear Algebra Collection of sample exams

Foundations of Computer Vision

SUMMARY OF MATH 1600

LECTURE NOTE #10 PROF. ALAN YUILLE

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

series. Utilize the methods of calculus to solve applied problems that require computational or algebraic techniques..

Seminar on Linear Algebra

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Linear Algebra. Session 12

Transcription:

Dimensionality Reduction Using the Sparse Linear Model: Supplementary Material Ioannis Gkioulekas arvard SEAS Cambridge, MA 038 igkiou@seas.harvard.edu Todd Zickler arvard SEAS Cambridge, MA 038 zickler@seas.harvard.edu Abstract In this supplementary material, we report results from additional experiments, and provide complete proofs for the mathematical statements made in the paper. Specifically, we provide larger scale versions of the visualization results on the CMU PIE dataset presented in the paper, as well as the corresponding visualization when using PCA. We also perform further experiments report on facial images using the Extended Yale B. Furthermore, we present the derivation of the optimization problems for the general dictionary and sparse linear models and their closed-form solutions, presented in Sections 3. and 3 of the paper, respectively. For the nonlinear extension of Section 3., we discuss an alternative generalization of the sparse linear model than that discussed in the paper, and rigorously derive the optimization problem and representer theorem we used the kernel case. We use the same notation as in the paper. To avoid confusion, when referring to sections, figures, and equations in the main paper, we always explicitly indicate so. Visualization of CMU PIE In Figure of the paper, we show two-dimensional projections of all samples in the CMU PIE dataset, as well as identity-averaged faces across the dataset for various illuminations, poses, and expressions, produced using LPP and the proposed method. In Figures -, we show larger scale versions of these projections, so that details are better visible. Furthermore, in Figure 3, we show the corresponding projections produced by using PCA. Experiments on Extended Yale B Dataset We report additional experimental results on facial images in the linear case, this time using the Extended Yale B dataset (specifically, the subset used in ). We pre-normalize face images to be unit-length vectors, and use the same settings for dictionary learning as for the experiments on CMU PIE. We repeat the recognition experiments on facial images of Section 4 of the paper. Due to the much smaller size of the Extended Yale B dataset when compared to CMU PIE, we only consider the cases of 50 and 40 training samples for each of the 38 individuals in the dataset. Other than that, we use the same experimental setting as with CMU PIE. In Figure 4, we show the average recognition accuracy versus the number of projections achieved by various methods, and for different numbers of training samples. The main conclusions drawn from the experiments on CMU PIE apply here as well, with the proposed method outperforming competing methods. Note, however, that in this case performance

Figure : Two-dimensional projection of CMU PIE dataset, colored by identity, obtained using LPP. Shown at high resolution and at their respective projections are identity-averaged faces across the dataset for various illuminations, poses, and expressions. Insets show projections of samples from only two distinct identities. (Best viewed in color.) for all methods deteriorates faster as the training samples per individual decrease, due to the much smaller dataset. 3 Derivation of optimization problem for the general and sparse linear models We begin by deriving equation (0) of the paper for the more general dictionary model. We denote for convenience E D T SD I, and therefore equation (5) of the paper (which holds for the general dictionary model as well) becomes (a T min E a,a,, Ea + T L M N SDa + T SDa + T ) S, () which, after expanding the square, can be written as (a ) T ( ) min E a,a,, Ea + T ( ) SDa + T ( ) SDa + T S L M N +a T Ea T SDa +a T Ea T SDa +a T Ea T S + T SDa T SDa + T SDa T S + T SDa T S. () Due to the zero-mean and independence assumptions fora,a,,, it is straightforward to show that the expectation of the summands corresponding to the cross-terms is equal to zero. Therefore, () can be reduced to (a ) T ( ) T min E a,a Ea +E a, SDa L M N ( ) T ( ) T +E a, SDa +E, S. (3)

Figure : Two-dimensional projection of CMU PIE dataset, colored by identity, obtained using the proposed method. Shown at high resolution and at their respective projections are identity-averaged faces across the dataset for various illuminations, poses, and expressions. Insets show projections of samples from only two distinct identities. (Best viewed in color.) Figure 3: Two-dimensional projection of CMU PIE dataset, colored by identity, obtained using PCA. Shown at high resolution and at their respective projections are identity-averaged faces across the dataset for various illuminations, poses, and expressions. Insets show projections of samples from only two distinct identities. (Best viewed in color.) 3

00 90 90 Classification accuracy 80 70 random 60 PCA 50 LPP NPE 40 SRLPP 30 SmoothLPP dict 0 0 40 60 80 00 Number of dimensions Classification accuracy 80 70 60 SmoothLPP 40 dict 50 50 0 40 60 80 00 Number of dimensions Figure 4: Results on Extended YaleB. Left: Classification accuracy of various methods, for different values of the number of projections M, using 50 training samples per individual. Right: Classification accuracy of proposed and best alternative method, for different values of the number of training samples per individual (color) and projections M. (a ) T We consider the first term in (3), E a,a Ea. If we perform the vector-matrix-vector multiplication and expand the square, the resulting scalar is equal to a linear combination of terms of the forma i a k a j a l, fori,j,k,l,...,k, wherea (a,...,a K ) and similarly fora, and with constant coefficients that depend on the entries of the matrix E. From the independence and zero-mean assumptions for the components of the vectors a and a, the expectation of all the terms withi k or j l is equal to zero. Therefore, we have that (a ) T E a,a Ea c ij a ia j, (4) i j and by evaluating the vector-matrix-vector product analytically, it is easy to see that c ij (E) ij. (5) Using the fact that, for any matrix M, its Frobenius norm can be written as M F i j (M) ij, along with equations (4) and (5), we get directly that (a ) T E a,a Ea E W, (W ) ij E a ia j. (6) Using derivations analogous to the above, it is easy to also prove that ( ) T ( ) T E a, SDa +E a, SDa (SD) W, (7) F ( ) T E, S S W 3, (8) with Combining the above, we obtain equation (0) of the paper. F (W ) ij E ia j +E i a j, (9) (W 3 ) ij E i j. (0) In the case of the sparse linear model, from the assumption that the components ofaandare i.i.d. Laplace and Gaussian respectively, we have for all i and j, E a ia j E a i E a j 4τ 4, () E ia j E i E a j σ τ, () E a i j E a i E j σ τ, (3) E i j E i E j σ 4. (4) Using these, equation (0) of the paper is simplified into equation (6) of the paper, which is the optimization problem we consider for the case of the sparse linear model. F 4

4 Derivation of solution for the sparse linear model case We introduce some notation. The singular value decomposition of L is L UΣZ T, (5) where U is a M M orthogonal matrix, Σ is an invertible M M diagonal matrix, and Z is a N M matrix with orthonormal columns, that is Z T Z I. Similarly, the eigendecomposition of the positive semidefinite matrixdd T is DD T V ΛV T, (6) where V is an N orthogonal matrix, and Λ is an N diagonal matrix. We re-write below for convenience the objective function of the optimization problem of equation (6) of the paper, F (L) 4τ 4 D T L T LD I +4τ σ L T LD +σ 4 L T L. (7) F F We can obtain the stationary points of (7) by solving F 0. (8) L Using standard matrix differentiation tools, we can write the derivative of F with respect tolas F ( ) L 4L 4τ 4 DD T L T LDD T 4τ 4 DD T +4σ τ L T LDD T +σ 4 L T L. (9) From (8) and (9), we firstly obtain the trivial solution L 0. All other solutions of (8) correspond to the case when all of the columns of the matrix 4τ 4 DD T L T LDD T 4τ 4 DD T +4σ τ L T LDD T +σ 4 L T L (0) belong to in the null space of L. From basic linear algebra, this condition can be equivalently expressed as ( ) Z T 4τ 4 DD T L T LDD T 4τ 4 DD T +4σ τ L T LDD T +σ 4 L T L 0. () ( By multiplying the above from the right with DD T) ZΣ, with some basic matrix manipulation and using the orthogonality properties of the matrices involved, the above becomes 4τ 4 Z T V ΛV T Z +σ 4 Σ Z T V Λ V T ZΣ 4τ 4 Σ 4σ τ I. () We notice that the right-hand part of (8) is always symmetric. For the left-hand part of (8) to be symmetric, it is easy to check that one of the following conditions must be true: Σ must be the identify matrix; or the off-diagonal elements of the matrix Z t V must be zero. The first condition can be rejected, as it is easy to check that then it is not possible for both (8) and the requirement Z T Z I to be true. The second condition, and the fact that V is orthogonal, imply that Z is formed from any M columns of V, that isz V M. Then, () becomes 4τ 4 Λ M +σ 4 Σ Λ M Σ 4τ 4 Σ 4σ τ I, (3) where Λ M is the diagonal matrix formed from the M eigenvalues of DD T corresponding to the M columns of V used to form Z. This is an equation involving only diagonal matrices and can be solved trivially for its diagonal elements. Combining the above, we obtain that the solutions of (8), and therefore the stationary points of (7), are either L 0, or of the form (up to an arbitrary rotation corresponding to the matrixu in (5)) L diag(f (λ M ))V T M. (4) where λ M (λ,...,λ M ) is a M vector composed of any subset of M eigenvalues of the N N matrix DD T, V M is the N M matrix with the corresponding eigenvectors as columns, f ( ) is a function applied element-wise to the vector λ M, equal to (as obtained by solving (3)) 4τ f (λ i ) 4 λ i σ 4 +4τ σ λ i +4τ 4 λ, (5) i F 5

anddiag(f (λ M )) is them M diagonal matrix withf (λ M ) as its diagonal. Substituting the above solution in (7), and using (5), (6) and the orthogonality properties of the matrices involved, it is easy to see that the objective function evaluated at the stationary points is equal to F (L) 4τ 4 Λ M diag(f (λ M)) Λ M I +4τ σ diag(f (λm )) Λ M F F +σ4 diag(f (λm )) F. (6) From the definition of the Frobenius norm, and some algebraic manipulation, we can rewrite the above as M F (L) h(λ i ), (7) where i h(λ i ) 4σ τ 4( σ +4λ i τ ) (σ +λ i τ ). (8) It is easy to see that h(λ) is a strictly decreasing function of λ for all positive values of σ and τ. Consequently, the stationary point where the objective function F (L) has the lowest value is obtained when the M largest eigenvalues are selected in (4), therefore arriving at the solution presented in equation (8) of the paper. We also see that for σ 0, the objective function has the same value for all stationary points, which is the reason for the solution ambiguity we discuss in the paper with regards to the noiseless case (see equation (9) of the paper). 5 Alternative model for the kernel case We briefly discuss here an extension of the sparse linear model to the kernel case different from the one used in Section 3. of the paper. We denote by C the subspace spanned by the atoms of dictionary D, } C span{ di,i,...,k. (9) C has finite dimension d K, and thus is closed. Therefore, it has an orthogonal complement C and we can write C C. We also denote by P C the orthogonal projection toc. Denote by the set {ẽ i,i,...,d} the orthonormal basis of C, and by the set {ẽ i,i d+,...} its extension to the rest of the space. Then, we assume a probabilistic model where samples are generated from equation () of the paper, and under the same assumptions for a as in the paper. We also still assume that is a Gaussian process over R N with sample patchs in. owever, we assume that the covariance operator of is defined as follows, ẽ i,c ẽ j σ, i,j d, (30) ẽ i,c ẽ j 0, i > d or j > d. (3) As the set {ẽ i,i,...} is an orthonormal basis of, the above offers a full characterization of the operator C. It is easy to construct such an operator, and therefore the Gaussian process, even on infinite dimensional spaces. Equivalently, the operator acts as the identity (times a constant) for the projections of signals in the subspace C, and as the zero operator for the residuals. As a consequence, the above model allows for Gaussian noise of zero-mean and variance σ along all of the dimensions of the subspace C, but does not allow for noise along any other dimension of (or alternatively, noise there has zero variance). If we denote by { b i R d,i,...,k } the coordinates of the dictionary elements with respect to the subspace basis{ẽ i,i,...,d}, then all functions generated by the above model will also belong to C, and their coordinates with respect to the above basis will be c j b ij a i +ǫ j, j,...,d, (3) i where ǫ j, j,...,d are Gaussian random variables with mean zero and variance σ. Then, for such functions, it is easy to see that MAP estimation ofareduces to the kernelized lasso of equation (3) of the paper. 6

We note that, for anyf, the optimization problem (3) of the paper can be written equivalently as min a R K σ f P Cf + σ P Cf Da + τ a. (33) As the part σ f P C f does not depend on a, the above is exactly equivalent to min a R K σ P Cf Da + τ a, (34) and as P C f C, the above optimization problem can be rewritten as MAP estimation of the coordinates (3) and be explained probabilistically by the model we introduced above. Therefore, the above model can be used for the projections on C of all functions in, and then kernel lasso becomes MAP estimation of their projection s coordinates in some basis. The term f P C f is not explained by the above model, and in fact any non-zero such component occurs only with zero probability. The fact that σ f P C f cannot be the likelihood of some distribution for noise is exactly the problem we run into when trying to extend the model we used in the paper to the infinite dimensional case. owever, if one is willing to discard the component f P C f for all samples arising in practice, this alternative model can be applied to the infinite dimensional case for all signals. 6 Derivation of solution for the nonlinear case Firstly, we derive equation (4) of the paper. We have ( ) δp (VΦx ) T (VΦx ) a T a (35) ( VΦx,VΦx R M a T a ) (36) ( Φx,V VΦx a T a ) (37) ( Φx,SΦx a T a ), (38) where we have used that S V V. Then, using equation () of the paper, the above becomes δp ( Da +,S(Da + ) a T ) a (39) ( Da,S(Da ) +,S(Da ) + Da,S +,S a T a ) (40) ( K + i j a i a j ( di,s d j δ ij a i di,s i ) + j a j,s d j +,S a T a ). (4) whereδ ij is the Kronecker delta. From the equivalence of Gaussian processes with sample paths on ilbert spaces and Gaussian measures 3, 4, we have that E, f 0, f, (4) and similarly for. Furthermore, from the properties of the covariance operatorc C σ I, we have that E, f, g C f, g σ f, g, f, g, (43) and similarly for. If in (4) we expand the square and take the expectation, then using (4), (43), and an analysis similar to that presented before for deriving equation (0) of the paper, we arrive at E δp K 4τ 4 ( di,s d ) K j δ ij +4τ σ S d i,s d i + S S, (44) i i 7 i

which corresponds to equation (4) in the paper. We now need to prove that the minimizer of (44) over the set of compact, positive semi-definite, self-adjoint, and linear operators S of rank M has the form of equation (5) of the paper. For this purpose, we use and extend the representer theorem presented in 5. The first and third term of the objective function (44) along with the rank constraint correspond to the conditions of Theorem 3 of 5. ere, we have additionally the constraint that S be self-adjoint, and also the second term of the objective function that violates the conditions of that theorem. Extending the theorem to the case when S is also required to be positive semi-definite is straightforward. In order to handle terms of the form S d i,s d i, for i,...,k, note that due to the ilbert-schmidt norm term in the objective function (44), its minimizer S has finite ilbert-schmidt norm and thus is a ilbert- Schmidt operator. Therefore, we can consider its decomposition as S S S +S, where S S is the projection of S onto the linear span of { di d } j, i,j,...,k, S S i j γ ij di d j, (45) and S is orthogonal to each element of the above set, di,s dj 0, i,j,...,k. (46) AsS is compact and self-adjoint,s S ands are also compact and self-adjoint. Therefore, from the spectral theorem, S admits an eigendecomposition S λ k ṽ k ṽ k, (47) k where ṽ k is an orthonormal set and λ k are positive. Combining (47) with (46) for i j, i,...,k, we obtain ṽ k, d 0, i,...,k, k such that λ k 0. (48) For each i,...,k, we have S d i,s d i di,ss d i di, ( S S +S )( S S +S ) di di,s S S S di di +,S S S di + di,s S S di + di,s S di. (5) where we have used linearity and self-adjointness. owever, using (45), (47) and (48), it is easy to see that the terms involvings S S,S S S ands S are equal to zero. Therefore we conclude that S d i,s d i S S di,s S di. (5) From (5) and equation (0) of the proof of Theorem 3 of 5, and using the full rank assumption fork DD, we can deduce that the minimizer of (44) can be written in the form S γ ij di d j. (53) i j The matrix γ formed by the coefficients γ ij is related to the matrix α of Theorem 3 of 5 through the equation (again assuming that K DD has full rank) α K DD γk DD (54) as proved in the proof of Theorem 3 of 5 (see fifth equation in page 84) and applied to our case where the set of elements in the optimizer is { di d j, i,j,...,k}. The eigenvalues of α are the same of those of S, as shown in the proof of Theorem 3 of 5, and therefore α is positive semi-definite and of rank M. Under the assumption that the matrix K DD is full rank, we deduce that γ is also positive-semidefinite and of rank M. Combining this with (53), we conclude that the minimizer of (44) has the form of equation (5) of the paper. (49) (50) 8

References K.C. Lee, J. o, and D.J. Kriegman. Acquiring linear subspaces for face recognition under variable lighting. PAMI, 005. D. Cai, X. e, Y. u, J. an, and T. uang. Learning a spatially smooth subspace for face recognition. CVPR, 007. 3 A. Berlinet and C. Thomas-Agnan. Reproducing kernel ilbert spaces in probability and statistics. Kluwer, 004. 4 V.I. Bogachev. Gaussian measures. AMS, 998. 5 J. Abernethy, F. Bach, T. Evgeniou, and J.P. Vert. A new approach to collaborative filtering: Operator estimation with spectral regularization. JMLR, 009. 9