Matrix Completion from Noisy Entries

Similar documents
Matrix Completion from Noisy Entries

Matrix Completion from Noisy Entries

Matrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University

Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison

Matrix Completion from a Few Entries

Matrix Completion: Fundamental Limits and Efficient Algorithms

Low-rank Matrix Completion from Noisy Entries

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Robust Principal Component Analysis

Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds

Matrix Completion from Fewer Entries

Compressed Sensing and Robust Recovery of Low Rank Matrices

Analysis of Robust PCA via Local Incoherence

Guaranteed Rank Minimization via Singular Value Projection

Recovering any low-rank matrix, provably

Matrix estimation by Universal Singular Value Thresholding

Conditions for Robust Principal Component Analysis

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Matrix Completion from a Few Entries

Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization

Linear dimensionality reduction for data analysis

Random Methods for Linear Algebra

Low Rank Matrix Completion Formulation and Algorithm

An iterative hard thresholding estimator for low rank matrix recovery

Low-Rank Matrix Recovery

EE 381V: Large Scale Learning Spring Lecture 16 March 7

arxiv: v5 [math.na] 16 Nov 2017

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

The convex algebraic geometry of rank minimization

Breaking the Limits of Subspace Inference

Tighter Low-rank Approximation via Sampling the Leveraged Element

Dense Error Correction for Low-Rank Matrices via Principal Component Pursuit

Concentration-Based Guarantees for Low-Rank Matrix Reconstruction

A Randomized Algorithm for the Approximation of Matrices

From Compressed Sensing to Matrix Completion and Beyond. Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison

arxiv: v1 [stat.ml] 1 Mar 2015

Rank Determination for Low-Rank Data Completion

Sensor Network Localization from Local Connectivity : Performance Analysis for the MDS-MAP Algorithm

Solving Corrupted Quadratic Equations, Provably

Structured matrix factorizations. Example: Eigenfaces

A New Estimate of Restricted Isometry Constants for Sparse Solutions

Matrix Completion for Structured Observations

Lecture Notes 10: Matrix Factorization

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

CSC 576: Variants of Sparse Learning

Linear Regression and Its Applications

Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Principal Component Analysis

Lecture Notes 1: Vector spaces

Recent Developments in Compressed Sensing

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR

MATH 581D FINAL EXAM Autumn December 12, 2016

A Unified Approach to Proximal Algorithms using Bregman Distance

Joint Capped Norms Minimization for Robust Matrix Recovery

GoDec: Randomized Low-rank & Sparse Matrix Decomposition in Noisy Case

Analysis of Spectral Kernel Design based Semi-supervised Learning

Linear Algebra Massoud Malek

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

STAT 200C: High-dimensional Statistics

Institute for Computational Mathematics Hong Kong Baptist University

Divide-and-Conquer Matrix Factorization

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

Information-Theoretic Limits of Matrix Completion

Note on the convex hull of the Stiefel manifold

Lecture 9: Matrix approximation continued

arxiv: v1 [math.st] 10 Sep 2015

STUDY of genetic variations is a critical task to disease

Introduction to Machine Learning

Robust PCA by Manifold Optimization

Lecture 5 : Projections

On Optimal Frame Conditioners

1-Bit Matrix Completion

Robust PCA via Outlier Pursuit

An algebraic perspective on integer sparse recovery

A Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem

arxiv: v1 [cs.it] 21 Feb 2013

sublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU)

Technical Report 110. Project: Streaming PCA with Many Missing Entries. Research Supervisor: Constantine Caramanis WNCG

ELE 538B: Mathematics of High-Dimensional Data. Spectral methods. Yuxin Chen Princeton University, Fall 2018

Low rank matrix completion by alternating steepest descent methods

CS598 Machine Learning in Computational Biology (Lecture 5: Matrix - part 2) Professor Jian Peng Teaching Assistant: Rongda Zhu

The Power of Convex Relaxation: Near-Optimal Matrix Completion

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

S 1/2 Regularization Methods and Fixed Point Algorithms for Affine Rank Minimization Problems

Sparse Sensing in Colocated MIMO Radar: A Matrix Completion Approach

Binary matrix completion

CHAPTER 11. A Revision. 1. The Computers and Numbers therein

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

Linear Algebra (Review) Volker Tresp 2018

Singular Value Decomposition (SVD)

A fast randomized algorithm for overdetermined linear least-squares regression

Deep Linear Networks with Arbitrary Loss: All Local Minima Are Global

Basic Concepts in Linear Algebra

Nonlinear Programming Algorithms Handout

Self-Calibration and Biconvex Compressive Sensing

Adaptive one-bit matrix completion

Transcription:

Matrix Completion from Noisy Entries Raghunandan H. Keshavan, Andrea Montanari, and Sewoong Oh Abstract Given a matrix M of low-rank, we consider the problem of reconstructing it from noisy observations of a small, random subset of its entries. The problem arises in a variety of applications, from collaborative filtering (the Netflix problem ) to structure-from-motion and positioning. We study a low complexity algorithm introduced in [], based on a combination of spectral techniques and manifold optimization, that we call here OPTSPACE. We prove performance guarantees that are order-optimal in a number of circumstances. Introduction Spectral techniques are an authentic workhorse in machine learning, statistics, numerical analysis, and signal processing. Given a matrix M, its largest singular values and the associated singular vectors explain the most significant correlations in the underlying data source. A low-rank approximation of M can further be used for low-complexity implementations of a number of linear algebra algorithms [2]. In many practical circumstances we have access only to a sparse subset of the entries of an m n matrix M. It has recently been discovered that, if the matrix M has rank r, and unless it is too structured, a small random subset of its entries allow to reconstruct it exactly. This result was first proved by Candés and Recht [3] by analyzing a convex relaxation indroduced by Fazel [4]. A tighter analysis of the same convex relaxation was carried out in [5]. A number of iterative schemes to solve the convex optimization problem appeared soon thereafter [6, 7, 8] (also see [9] for a generalization). In an alternative line of work, the authors of [] attacked the same problem using a combination of spectral techniques and manifold optimization: we will refer to their algorithm as OPTSPACE. OPTSPACE is intrinsically of low complexity, the most complex operation being computing r singular values and the corresponding singular vectors of a sparse m n matrix. The performance guarantees proved in [] are comparable with the information theoretic lower bound: roughly nr max{r,log n} random entries are needed to reconstruct M exactly (here we assume m of order n). A related approach was also developed in [0], although without performance guarantees for matrix completion. The above results crucially rely on the assumption that M is exactly a rank r matrix. For many applications of interest, this assumption is unrealistic and it is therefore important to investigate their robustness. Can the above approaches be generalized when the underlying data is well approximated by a rank r matrix? This question was addressed in [] within the convex relaxation approach of [3]. The present paper proves a similar robustness result for OPTSPACE. Remarkably the guarantees we obtain are order-optimal in a variety of circumstances, and improve over the analogus results of []. Department of Electrical Engineering, Stanford University Departments of Statistics, Stanford University

. Model definition Let M be an m n matrix of rank r, that is M = UΣV T. () where U has dimensions m r, V has dimensions n r, and Σ is a diagonal r r matrix. We assume that each entry of M is perturbed, thus producing an approximately low-rank matrix N, with N ij = M ij + Z ij, (2) where the matrix Z will be assumed to be small in an appropriate sense. Out of the m n entries of N, a subset E [m] [n] is revealed. We let N E be the m n matrix that contains the revealed entries of N, and is filled with 0 s in the other positions { Nij E Nij if (i,j) E, = (3) 0 otherwise. The set E will be uniformly random given its size..2 Algorithm For the reader s convenience, we recall the algorithm introduced in [], which we will analyze here. The basic idea is to minimize the cost function F(X,Y ), defined by F(X,Y ) min F(X,Y,S), (4) S Rr r F(X,Y,S) (N ij (XSY T ) ij ) 2. (5) 2 (i,j) E Here X R n r, Y R m r are orthogonal matrices, normalized by X T X = m, Y T Y = n. Minimizing F(X,Y ) is an a priori difficult task, since F is a non-convex function. The key insight is that the singular value decomposition (SVD) of N E provides an excellent initial guess, and that the minimum can be found with high probability by standard gradient descent after this initialization. Two caveats must be added to this decription: () In general the matrix N E must be trimmed to eliminate over-represented rows and columns; (2) For technical reasons, we consider a slightly modified cost function to be denoted by F(X,Y ). OPTSPACE( matrix N E ) : Trim N E, and let ÑE be the output; 2: Compute the rank-r projection of ÑE, T r (ÑE ) = X 0 S 0 Y0 T ; 3: Minimize F(X,Y ) through gradient descent, with initial condition (X 0,Y 0 ). We may note here that the rank of the matrix M, if not known, can be reliably estimated from ÑE. We refer to the journal version of this paper for further details. The various steps of the above algorithm are defined as follows. Trimming. We say that a row is over-represented if it contains more than 2/m revealed entries (i.e. more than twice the average number of revealed entries). Analogously, a column is overrepresented if it contains more than 2/n revealed entries. The trimmed matrix ÑE is obtained from N E by setting to 0 over-represented rows and columns. ME and Z E are defined similarly. Hence, Ñ E = M E + Z E. Rank-r projection. Let Ñ E = min(m,n) i= 2 σ i x i y T i, (6)

be the singular value decomposition of ÑE, with singular vectors σ σ 2... We then define T r (ÑE ) = mn r σ i x i yi T. (7) Apart from an overall normalization, T r (ÑE ) is the best rank-r approximation to ÑE in Frobenius norm. i= Minimization. The modified cost function F is defined as F(X,Y ) = F(X,Y ) + ρg(x,y ) (8) m ( X (i) 2 ) n ( Y (j) 2 ) F(X,Y ) + ρ G + ρ G, (9) 3µ 0 r 3µ 0 r i= where X (i) denotes the i-th row of X, and Y (j) the j-th row of Y. See Section.3 below for the definition of µ 0. The function G : R + R is such that G (z) = 0 if z and G (z) = e (z )2 otherwise. Further, we can choose ρ = Θ(nǫ). Let us stress that the regularization term is mainly introduced for our proof technique to work (and a broad family of functions G would work as well). In numerical experiments we did not find any performance loss in setting ρ = 0. One important feature of OPTSPACE is that F(X,Y ) and F(X,Y ) are regarded as functions of the r-dimensional subspaces of R m and R n generated (respectively) by the columns of X and Y. This interpretation is justified by the fact that F(X,Y ) = F(XA,Y B) for any two orthogonal matrices A, B R r r (the same property holds for F ). The set of r dimensional subspaces of R m is a differentiable Riemannian manifold G(m, r) (the Grassman manifold). The gradient descent algorithm is applied to the function F : M(m,n) G(m,r) G(n,r) R. For further details on optimization by gradient descent on matrix manifolds we refer to [2, 3]..3 Main results Our first result shows that, in great generality, the rank-r projection of ÑE provides a reasonable approximation of M. Throughout this paper, without loss of generality, we assume α m/n. Theorem.. Let N = M + Z, where M has rank r and M ij M max for all (i,j) [m] [n], and assume that the subset of revealed entries E [m] [n] is uniformly random with size. Then there exists numerical constants C and C such that ( ) nrα 3/2 /2 M T r mn (ÑE ) F CM max + C n rα Z E 2, (0) with probability larger than /n 3. Projection onto rank-r matrices through SVD is pretty standard (although trimming is crucial for achieving the above guarantee). The key point here is that a much better approximation is obtained by minimizing the cost F(X,Y ) (step 3 in the pseudocode above), provided M satisfies an appropriate incoherence condition. Let M = UΣV T be a low rank matrix, and assume, without loss of generality, U T U = m and V T V = n. We say that M is (µ 0,µ )-incoherent if the following conditions hold. A. For all i [m], j [n] we have, r k= U2 ik µ 0r, r k= V 2 ik µ 0r. A2. There exists µ such that r k= U ik(σ k /Σ )V jk µ r /2. Theorem.2. Let N = M + Z, where M is a (µ 0,µ )-incoherent matrix of rank r, and assume that the subset of revealed entries E [m] [n] is uniformly random with size. Further, let Σ min = Σ Σ r = Σ max with Σ max /Σ min κ. Let M be the output of OPTSPACE on input N E. Then there exists numerical constants C and C such that if Cn ακ 2 max { µ 0 r α log n; µ 2 0r 2 ακ 4 ; µ 2 r 2 ακ 4}, () j= 3

then, with probability at least /n 3, M M F C κ 2 n αr Z E 2. (2) mn provided that the right-hand side is smaller than Σ min. Apart from capturing the effect of additive noise, these two theorems improve over the work of [] even in the noiseless case. Indeed they provide quantitative bounds in finite dimensions, while the results of [] were only asymptotic..4 Noise models In order to make sense of the above results, it is convenient to consider a couple of simple models for the noise matrix Z: Independent entries model. We assume that Z s entries are independent random variables, with zero mean E{Z ij } = 0 and sub-gaussian tails. The latter means that for some bounded constant σ 2. P{ Z ij x} 2e x2 2σ 2, (3) Worst case model. In this model Z is arbitrary, but we have an uniform bound on the size of its entries: Z ij Z max. The basic parameter entering our main results is the operator norm of Z E, which is bounded as follows. Theorem.3. If Z is a random matrix drawn according to the independent entries model, then there is a constant C such that, ( ) /2 Z αlog E 2 Cσ, (4) n with probability at least /n 3. If Z is a matrix from the worst case model, then for any realization of E. Z E 2 2 n α Z max, (5) Note that for = Ω(nlog n), no row or column is over-represented with high probability. It follows that in the regime of for which the conditions of Theorem.2 are satisfied, we have Z E = Z E. Then, among the other things, this result implies that for the independent entries model the right-hand side of our error estimate, Eq. (2), is with high probability smaller than Σ min, if Crα 3/2 nlog nκ 4 (σ/σ min ) 2. For the worst case model, the same statement is true if Z max Σ min /C rκ 2. Due to space constraints, the proof of Theorem.3 will be given in the journal version of this paper..5 Comparison with related work Let us begin by mentioning that a statement analogous to our preliminary Theorem. was proved in [4]. Our result however applies to any number of revealed entries, while the one of [4] requires (8log n) 4 n (which for n 5 0 8 is larger than n 2 ). As for Theorem.2, we will mainly compare our algorithm with the convex relaxation approach recently analyzed in []. Our basic setting is indeed the same, while the algorithms are rather different. 4

RMSE 0.8 0.6 0.4 Convex Relaxation Lower Bound rank-r projection OptSpace : iteration 2 iterations 3 iterations 0 iterations 0.2 0 0 00 200 300 400 500 600 /n Figure : Root mean square error achieved by OPTSPACE for reconstructing a random rank-2 matrix, as a function of the number of observed entries, and of the number of line minimizations. The performance of nuclear norm minimization and an information theory lower bound are also shown. Figure compares the average root mean square error for the two algorithms as a function of. Here M is a random rank r = 2 matrix of dimension m = n = 600, generated by letting M = ŨṼ T with Ũij,Ṽij i.i.d. N(0,20/ n). The noise is distributed according to the independent entries model with Z ij N(0,). This example is taken from [] Figure 2, from which we took the data for the convex relaxation approach, as well as the information theory lower bound. After one iteration, OPTSPACE has a smaller root mean square error than [], and in about 0 iterations it becomes indistiguishable from the information theory lower bound. Next let us compare our main result with the performance guarantee in [], Theorem 7. Let us stress that we require some bound on the condition number κ, while the analysis of [, 5] requires a stronger incoherence assumption. As far as the error bound is concerned, [] proved n M M F 7 mn ZE F + 2 n α ZE F. (6) (The constant in front of the first term is in fact slightly smaller than 7 in [], but in any case larger than 4 2). Theorem.2 improves over this result in several respects: () We do not have the second term on the right hand side of (6), that actually increases with the number of observed entries; (2) Our error decreases as n/ rather than (n/) /2 ; (3) The noise enters Theorem.2 through the operator norm Z E 2 instead of its Frobenius norm Z E F Z E 2. For E uniformly random, one expects Z E F to be roughly of order Z E 2 n. For instance, within the intependent entries model with bounded variance σ, Z E F = Θ( ) while Z E 2 is of order /n (up to logarithmic terms). 2 Some notations The matrix M to be reconstructed takes the form () where U R m r, V R n r. We write U = [u,u 2,...,u r ] and V = [v,v 2,...,v r ] for the columns of the two factors, with u i = m, v i = n, and u T i u j = 0, v T i v j = 0 for i j (there is no loss of generality in this, since normalizations can be absorbed by redefining Σ). 5

We shall write Σ = diag(σ,...,σ r ) with Σ Σ 2 Σ r > 0. The maximum and minimum singular values will also be denoted by Σ max = Σ and Σ min = Σ r. Further, the maximum size of an entry of M is M max max ij M ij. Probability is taken with respect to the uniformly random subset E [m] [n] given and (eventually) the noise matrix Z. Define ǫ / mn. In the case when m = n, ǫ corresponds to the average number of revealed entries per row or column. Then it is convenient to work with a model in which each entry is revealed independently with probability ǫ/ mn. Since, with high probability [ǫ α n A nlog n,ǫ α n + A nlog n], any guarantee on the algorithm performances that holds within one model, holds within the other model as well if we allow for a vanishing shift in ǫ. We will use C, C etc. to denote universal numerical constants. Given a vector x R n, x will denote its Euclidean norm. For a matrix X R n n, X F is its Frobenius norm, and X 2 its operator norm (i.e. X 2 = sup u 0 Xu / u ). The standard scalar product between vectors or matrices will sometimes be indicated by x,y or X,Y, respectively. Finally, we use the standard combinatorics notation [N] = {,2,...,N} to denote the set of first N integers. 3 Proof of Theorem. As explained in the introduction, the crucial idea is to consider the singular value decomposition of the trimmed matrix ÑE instead of the original matrix N E. Apart from a trivial rescaling, these singular values are close to the ones of the original matrix M. Lemma 3.. There exists a numerical constant C such that, with probability greater than /n 3, σ q ǫ Σ α q CM max ǫ + ǫ Z E 2, (7) where it is understood that Σ q = 0 for q > r. Proof. For any matrix A, let σ q (A) denote the qth singular value of A. Then, σ q (A+B) σ q (A)+ σ (B), whence σ q ǫ Σ q σ q ( M E )/ǫ Σ q + σ ( Z E )/ǫ CM max α ǫ + ǫ Z E 2, where the second inequality follows from the following Lemma as shown in []. Lemma 3.2 (Keshavan, Montanari, Oh, 2009 []). There exists a numerical constant C such that, with probability larger than /n 3, mn 2 α M M E CM max mn ǫ ǫ. (8) We will now prove Theorem.. Proof. (Theorem.) For any matrix A of rank at most 2r, A F 2r A 2, whence 2r M T r mn (ÑE ) F mn (ÑE mn M ) σ i x i yi T ǫ i r+ 2 ( 2r M mn ) M E 2 mn + mn ǫ ǫ Z mn E 2 + σ r+ ǫ This proves our claim. 2CM max 2αr/ǫ + (2 2r/ǫ) ZE 2 ( ) nrα C 3/2 /2 M max + 2 ( ) n rα 2 Z E 2. 6

4 Proof of Theorem.2 Recall that the cost function is defined over the Riemannian manifold M(m,n) G(m,r) G(n,r). The proof of Theorem.2 consists in controlling the behavior of F in a neighborhood of u = (U,V ) (the point corresponding to the matrix M to be reconstructed). Throughout the proof we let K(µ) be the set of matrix couples (X,Y ) R m r R n r such that X (i) 2 µr, Y (j) 2 µr for all i,j 4. Preliminary remarks and definitions Given x = (X,Y ) and x 2 = (X 2,Y 2 ) M(m,n), two points on this manifold, their distance is defined as d(x,x 2 ) = d(x,x 2 ) 2 + d(y,y 2 ) 2, where, letting (cos θ,...,cos θ r ) be the singular values of X T X 2 /m, d(x,x 2 ) = θ 2. (9) Given S achieving the minimum in Eq. (4), it is also convenient to introduce the notations d (x,u) Σ 2 min d(x,u)2 + S Σ 2 F, (20) d + (x,u) Σ 2 maxd(x,u) 2 + S Σ 2 F. (2) 4.2 Auxiliary lemmas and proof of Theorem.2 The proof is based on the following two lemmas that generalize and sharpen analogous bounds in [] (for proofs we refer to the journal version of this paper). Lemma 4.. There exists numerical constants C 0,C,C 2 such that the following happens. Assume ǫ C 0 µ 0 r α max{log n; µ 0 r α(σ min /Σ max ) 4 } and δ Σ min /(C 0 Σ max ). Then, F(x) F(u) C nǫ α d (x,u) 2 C n rα Z E 2 d + (x,u), (22) F(x) F(u) C 2 nǫ α Σ 2 max d(x,u) 2 + C 2 n rα Z E 2 d + (x,u), (23) for all x M(m,n) K(4µ 0 ) such that d(x,u) δ, with probability at least /n 4. Here S R r r is the matrix realizing the minimum in Eq. (4). Corollary 4.2. There exist a constant C such that, under the hypotheses of Lemma 4. r S Σ F CΣ max d(x,u) + C ǫ ZE 2. (24) Further, for an appropriate choice of the constants in Lemma 4., we have r σ max (S) 2Σ max + C ǫ ZE 2, (25) σ min (S) r 2 Σ min C ǫ ZE 2. (26) Lemma 4.3. There exists numerical constants C 0,C,C 2 such that the following happens. Assume ǫ C 0 µ 0 r α (Σ max /Σ min ) 2 max{log n; µ 0 r α(σ max /Σ min ) 4 } and δ Σ min /(C 0 Σ max ). Then, [ grad F(x) 2 C nǫ 2 Σ 4 rσmax Z E ] 2 2 min d(x,u) C 2, (27) ǫσ min Σ min + for all x M(m,n) K(4µ 0 ) such that d(x,u) δ, with probability at least /n 4. (Here [a] + max(a,0).) We can now turn to the proof of our main theorem. 7

Proof. (Theorem.2). Let δ = Σ min /C 0 Σ max with C 0 large enough so that the hypotheses of Lemmas 4. and 4.3 are verified. Call {x k } k 0 the sequence of pairs (X k,y k ) M(m,n) generated by gradient descent. By assumption, the following is true with a large enough constant C: Z E 2 ǫ ( ) 2 C Σmin Σ min. (28) r Σ max Further, by using Corollary 4.2 in Eqs. (22) and (23) we get where F(x) F(u) C nǫ ασ 2 min{ d(x,u) 2 δ 2 0, }, (29) F(x) F(u) C 2 nǫ ασ 2 max{ d(x,u) 2 + δ 2 0,+}, (30) rσmax Z E 2 δ 0, C, ǫσ min Σ min By Eq. (28), we can assume δ 0,+ δ 0, δ/0. rσmax Z E 2 δ 0,+ C. (3) ǫσ min Σ max For ǫ Cαµ 2 r 2 (Σ max /Σ min ) 4 as per our assumptions, using Eq. (28) in Theorem., together with the bound d(u,x 0 ) M X 0 SY T 0 F /n ασ min, we get We make the following claims :. x k K(4µ 0 ) for all k. d(u,x 0 ) δ 0. (32) Indeed without loss of generality we can assume x 0 K(3µ 0 ) (because otherwise we can rescale those lines of X 0, Y 0 that violate the constraint). Therefore F(x 0 ) = F(x 0 ) 4C 2 nǫ ασ 2 maxδ 2 /00. On the other hand F(x) ρ(e /9 ) for x K(4µ 0 ). Since F(x k ) is a non-increasing sequence, the thesis follows provided we take ρ C 2 nǫ ασ 2 min. 2. d(x k,u) δ/0 for all k. Assuming ǫ Cαµ 2 r 2 (Σ max /Σ min ) 6, we have d(x 0,u) 2 (Σ 2 min /C Σ 2 max)(δ/0) 2. Also assuming Eq. (28) with large enough C we can show the following. For all x k such that d(x k,u) [δ/0,δ], we have F(x) F(x) F(x 0 ). This contradicts the monotonicity of F(x), and thus proves the claim. Since the cost function is twice differentiable, and because of the above, the sequence {x k } converges to Ω = { x K(4µ 0 ) M(m,n) : d(x,u) δ,grad F(x) = 0 }. (33) By Lemma 4.3 for any x Ω, which implies the thesis using Corollary 4.2. Acknowledgements rσmax Z E 2 d(x,u) C (34) ǫσ min Σ min This work was partially supported by a Terman fellowship, the NSF CAREER award CCF-0743978 and the NSF grant DMS-08062. 8

References [] R. H. Keshavan, A. Montanari, and S. Oh. Matrix completion from a few entries. arxiv:090.350, January 2009. [2] A. Frieze, R. Kannan, and S. Vempala. Fast monte-carlo algorithms for finding low-rank approximations. J. ACM, 5(6):025 04, 2004. [3] E. J. Candès and B. Recht. Exact matrix completion via convex optimization. arxiv:0805.447, 2008. [4] M. Fazel. Matrix Rank Minimization with Applications. PhD thesis, Stanford University, 2002. [5] E. J. Candès and T. Tao. The power of convex relaxation: Near-optimal matrix completion. arxiv:0903.476, 2009. [6] J-F Cai, E. J. Candès, and Z. Shen. A singular value thresholding algorithm for matrix completion. arxiv:080.3286, 2008. [7] S. Ma, D. Goldfarb, and L. Chen. Fixed point and Bregman iterative methods for matrix rank minimization. arxiv:0905.643, 2009. [8] K. Toh and S. Yun. An accelerated proximal gradient algorithm for nuclear norm regularized least squares problems. http://www.math.nus.edu.sg/ matys, 2009. [9] J. Wright, A. Ganesh, S. Rao, and Y. Ma. Robust principal component analysis: Exact recovery of corrupted low-rank matrices. arxiv:0905.0233, 2009. [0] K. Lee and Y. Bresler. Admira: Atomic decomposition for minimum rank approximation. arxiv:0905.0044, 2009. [] E. J. Candès and Y. Plan. Matrix completion with noise. arxiv:0903.33, 2009. [2] A. Edelman, T. A. Arias, and S. T. Smith. The geometry of algorithms with orthogonality constraints. SIAM J. Matr. Anal. Appl., 20:303 353, 999. [3] P.-A. Absil, R. Mahony, and R. Sepulchrer. Optimization Algorithms on Matrix Manifolds. Princeton University Press, 2008. [4] D. Achlioptas and F. McSherry. Fast computation of low-rank matrix approximations. J. ACM, 54(2):9, 2007. 9