Non-square matrix sensing without spurious local minima via the Burer-Monteiro approach

Similar documents
arxiv: v2 [stat.ml] 27 Sep 2016

COR-OPT Seminar Reading List Sp 18

Research Statement Qing Qu

Overparametrization for Landscape Design in Non-convex Optimization

Low-rank Solutions of Linear Matrix Equations via Procrustes Flow

Ranking from Crowdsourced Pairwise Comparisons via Matrix Manifold Optimization

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Symmetric Factorization for Nonconvex Optimization

Global Optimality in Low-rank Matrix Optimization

How to Escape Saddle Points Efficiently? Praneeth Netrapalli Microsoft Research India

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Incremental Reshaped Wirtinger Flow and Its Connection to Kaczmarz Method

Characterization of Gradient Dominance and Regularity Conditions for Neural Networks

The convex algebraic geometry of rank minimization

A Geometric Analysis of Phase Retrieval

Symmetry, Saddle Points, and Global Optimization Landscape of Nonconvex Matrix Factorization

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Dropping Convexity for Faster Semi-definite Optimization

Linear dimensionality reduction for data analysis

Robust Principal Component Analysis

Fast Algorithms for Robust PCA via Gradient Descent

Analysis of Robust PCA via Local Incoherence

Foundations of Deep Learning: SGD, Overparametrization, and Generalization

Compressed Sensing and Robust Recovery of Low Rank Matrices

Mini-Course 1: SGD Escapes Saddle Points

Gradient Descent Can Take Exponential Time to Escape Saddle Points

Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization

Universal low-rank matrix recovery from Pauli measurements

Implicit Regularization in Nonconvex Statistical Estimation: Gradient Descent Converges Linearly for Phase Retrieval and Matrix Completion

Sparse and low-rank decomposition for big data systems via smoothed Riemannian optimization

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Collaborative Filtering Matrix Completion Alternating Least Squares

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

arxiv: v3 [math.oc] 19 Oct 2017

An iterative hard thresholding estimator for low rank matrix recovery

Sketchy Decisions: Convex Low-Rank Matrix Optimization with Optimal Storage

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

Solving Corrupted Quadratic Equations, Provably

CSC 576: Variants of Sparse Learning

Lecture Notes 10: Matrix Factorization

Convex sets, conic matrix factorizations and conic rank lower bounds

Online Generalized Eigenvalue Decomposition: Primal Dual Geometry and Inverse-Free Stochastic Optimization

Nonconvex Matrix Factorization from Rank-One Measurements

Sparse Solutions of an Undetermined Linear System

Recovering any low-rank matrix, provably

A Conservation Law Method in Optimization

Low-rank Solutions of Linear Matrix Equations via Procrustes Flow

Fast recovery from a union of subspaces

WE study convex minimization problems with respect to

arxiv: v2 [math.oc] 5 Nov 2017

SPARSE signal representations have gained popularity in recent

Provable Alternating Minimization Methods for Non-convex Optimization

Non-convex Robust PCA: Provable Bounds

Convergence of Cubic Regularization for Nonconvex Optimization under KŁ Property

First Efficient Convergence for Streaming k-pca: a Global, Gap-Free, and Near-Optimal Rate

GROUP-SPARSE SUBSPACE CLUSTERING WITH MISSING DATA

Distributed Inexact Newton-type Pursuit for Non-convex Sparse Learning

Breaking the Limits of Subspace Inference

BURER-MONTEIRO GUARANTEES FOR GENERAL SEMIDEFINITE PROGRAMS arxiv: v1 [math.oc] 15 Apr 2019

Median-Truncated Gradient Descent: A Robust and Scalable Nonconvex Approach for Signal Estimation

Optimization methods

Low-Rank Matrix Recovery

arxiv: v1 [math.oc] 1 Jul 2016

Non-Convex Optimization. CS6787 Lecture 7 Fall 2017

Non-Convex Rank/Sparsity Regularization and Local Minima

On Optimal Frame Conditioners

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016

MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications. Class 19: Data Representation by Design

Lecture Notes 9: Constrained Optimization

arxiv: v2 [stat.ml] 22 Nov 2016

Lift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints

Tractable Upper Bounds on the Restricted Isometry Constant

The non-convex Burer Monteiro approach works on smooth semidefinite programs

Tutorial: PART 2. Optimization for Machine Learning. Elad Hazan Princeton University. + help from Sanjeev Arora & Yoram Singer

Sparse and Low-Rank Matrix Decompositions

arxiv: v1 [cs.lg] 2 Mar 2017

Research Statement Figure 2: Figure 1:

Global Optimality in Matrix and Tensor Factorization, Deep Learning & Beyond

Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds

Compressed Sensing and Sparse Recovery

Optimisation Combinatoire et Convexe.

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Optimal Linear Estimation under Unknown Nonlinear Transform

A Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem

A random perturbation approach to some stochastic approximation algorithms in optimization.

A Unified Variance Reduction-Based Framework for Nonconvex Low-Rank Matrix Recovery

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

The Projected Power Method: An Efficient Algorithm for Joint Alignment from Pairwise Differences

Low rank matrix completion by alternating steepest descent methods

arxiv: v1 [math.oc] 9 Oct 2018

Stochastic Variance Reduction for Nonconvex Optimization. Barnabás Póczos

Non-convex optimization. Issam Laradji

Two Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery

Collaborative Filtering

what can deep learning learn from linear regression? Benjamin Recht University of California, Berkeley

COMPRESSED Sensing (CS) is a method to recover a

The Deep Ritz method: A deep learning-based numerical algorithm for solving variational problems

Strengthened Sobolev inequalities for a random subspace of functions

Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison

Transcription:

Non-square matrix sensing without spurious local minima via the Burer-Monteiro approach Dohuyng Park Anastasios Kyrillidis Constantine Caramanis Sujay Sanghavi acebook UT Austin UT Austin UT Austin Abstract We consider the non-square matrix sensing problem, under restricted isometry property RIP) assumptions. We focus on the non-convex formulation, where any rank-r matrix X R m n is represented as UV, where U R m r and V R n r. In this paper, we complement recent findings on the non-convex geometry of the analogous PSD setting [5], and show that matrix factorization does not introduce any spurious local minima, under RIP. 1 Introduction and Problem ormulation Consider the following matrix sensing problem: min X R m n fx) : AX) b subject to rankx) r. 1) Here, b R p denotes the set of observations and A : R m n R p is the sensing linear map. The motivation behind this task comes from several applications, where we are interested in inferring an unknown matrix X R m n from b. Common assumptions are i) p m n, ii) b AX ) + w, i.e., we have a linear measurement system, and iii) X is rank-r, r min{m, n}. Such problems appear in a variety of research fields and include image processing [11, ], data analytics [13, 11], quantum computing [1, 19, 6], systems [3], and sensor localization [3] problems. There are numerous approaches that solve 1), both in its original non-convex form or through its convex relaxation; see [8, 16] and references therein. However, satisfying the rank constraint or any nuclear norm constraints in the convex relaxation) per iteration requires SVD computations, which could be prohibitive in practice for large-scale settings. To overcome this obstacle, recent approaches reside on non-convex parametrization of the variable space and encode the low-rankness directly into the objective [5,,, 3, 9, 1,, 8,, 50,, 35, 6, 37, 36, 7, 3, 9, 33]. In particular, we know that a rank-r matrix X R m n can be written as a product UV, where U R m r and V R n r. Such a re-parametrization technique has a long history [5, 15, 39], and was popularized by Burer and Monteiro [8, 9] for solving semi-definite programs SDPs). Using this observation in 1), we obtain the following non-convex, bilinear problem: min fuv ) : AUV ) b. U R m r,v R n r ) Now, ) has a different form of non-convexity due to the bilinearity of the variable space, which raises the question whether we introduce spurious local minima by doing this transformation. Contributions: The goal of this paper is to answer negatively to this question: We show that, under standard regulatory assumptions on A, UV parametrization does not introduce any spurious local minima. To do so, we non-trivially generalize recent developments for the square, PSD case [5] to the non-square case for X. Our result requires a different but equivalent) problem re-formulation and analysis, with the introduction of an appropriate regularizer in the objective. Appearing in Proceedings of the 0 th International Conference on Artificial Intelligence and Statistics AISTATS) sider similar questions, but for other objectives. [0] Related work. There are several papers that con- 017, ort Lauderdale, lorida, USA. JMLR: W&CP volume 5. Copyright 017 by the authors. characterizes the non-convex geometry of the complete dictionary recovery problem, and proves that

all local minima are global; [6] considers the problem of non-convex phase synchronization where the task is modeled as a non-convex least-squares optimization problem, and can be globally solved via a modified version of power method; [1] show that a nonconvex fourth-order polynomial objective for phase retrieval has no local minimizers and all global minimizers are equivalent; [3, 7] show that the Burer- Monteiro approach works on smooth semidefinite programs, with applications in synchronization and community detection; [17] consider the PCA problem under streaming settings and use martingale arguments to prove that stochastic gradient descent on the factors reaches to the global solution with nonnegligible probability; [0] introduces the notion of strict saddle points and shows that noisy stochastic gradient descent can escape saddle points for generic objectives f; [30] proves that gradient descent converges to local) minimizers almost surely, using arguments drawn from dynamical systems theory. More related to this paper are the works of [1] and [5]: they show that matrix completion and sensing have no spurious local minima, for the case where X is square and PSD. or both cases, extending these arguments for the more realistic non-square case is a non-trivial task. 1.1 Assumptions and Definitions We first state the assumptions we make for the matrix sensing setting. We consider the case where the linear operator A satisfies the Restricted Isometry Property, according to the following definition [1]: Definition 1.1 Restricted Isometry Property RIP)). A linear operator A : R m n R p satisfies the restricted isometry property on rank-r matrices, with parameter δ r, if the following set of inequalities hold for all rank-r matrices X: 1 δ r ) X AX) 1 + δ r ) X. Characteristic examples are Gaussian-based linear maps [18, 38], Pauli-based measurement operators, used in quantum state tomography applications [31], ourier-based measurement operators, which lead to computational gains in practice due to their structure [7, 38], or even permuted and sub-sampled noiselet linear operators, used in image and video compressive sensing applications []. In this paper, we consider sensing mechanisms that can be expressed as: AX)) i A i, X, i 1,..., p, and A i R m n. E.g., for the case of a Gaussian map A, A i are independent, identically distributed i.i.d.) Gaussian matrices; for the case of a Pauli map A, A i R n n are i.i.d. and drawn uniformly at random from a set of scaled Pauli observables P 1 P P d )/ n, where n d and P i is a Pauli observable matrix [31]. A useful property derived from the RIP definition is the following [10]: Proposition 1. Useful property due to RIP). or a linear operator A : R m n R p that satisfies the restricted isometry property on rank-r matrices, the following inequality holds for any two rank-r matrices X, Y R m n : A i, X A i, Y X, Y δ r X Y. An important issue in optimizing f over the factored space is the existence of non-unique possible factorizations for a given X. Since we are interested in obtaining a low-rank solution in the original space, we need a notion of distance to the low-rank solution X over the factors. Among infinitely many possible decompositions of X, we focus on the set of equally-footed factorizations [3]: { Xr U, V ) : U V X, } σ iu ) σ iv ) σ ix ) 1/, i [r]. 3) Given a pair U, V ), we define the distance to X as: ] [ ] Dist U, V ; X ) U U [ V V. min U,V ) X r 1. Problem Re-formulation Before we delve into the main results, we need to further reformulate the objective ) for our analysis. irst, we use a well-known transformation to reduce ) to a semidefinite optimization. Let us define auxiliary variables [ ] U W R m+n) r, V [ ] U W R m+n) r. V Based on the auxiliary variables, we define the linear map B : R m+n) m+n) R p such that BW W )) i B i, W W, and B i R m+n) m+n). To make a connection between the variable spaces U, V ) and W, A and B are related via matrices A i and B i as follows: B i 1 [ 0 Ai A i 0 ].

This further implies that: Main Results BW W )) i 1 B i, W W 1 [ ] [ ] 0 Ai UU A, UV i 0 V U V V A i, UV. Given the above, we re-define f : R m+n) r R such that fw ) : BW W ) b. ) It is important to note that B operates on m + n) m + n) matrices, while we assume RIP on A and m n matrices. Making no other assumptions for B, we cannot directly apply [5] on ), but a rather different analysis is required. In addition to this redefinition, we also introduce a regularizer g : R m+n) r R such that gw ) : λ W W λ U U V V. This regularizer was first introduced in [3] to prove convergence of its algorithm for non-square matrix sensing, and it is also used in this paper to analyze local minima of the problem. After setting λ 1, ) can be equivalently written as: minimize BW W ) b W R m+n) r + 1 W W. 5) By equivalent, we note that the addition of g in the objective does not change the problem, since for any rank-r matrix X there is a pair of factors U, V ) such that gw ) 0. It merely reduces the set of optimal points from all possible factorizations of X to balanced factorizations of X in Xr. U and V have the same set of singular values, which are the square roots of the singular values of X. A key property of the balanced factorizations is the following. Proposition 1.3. or any factorization of the form 3), it holds that W W U U V V 0 Proof. By balanced factorizations of X U V, we mean that factors U and V satisfy U AΣ 1/ R, V BΣ 1/ R 6) where X AΣB is the SVD, and R R r r is an orthonormal matrix. Apply this to W W to get the result. Therefore, we have gw ) 0, and U, V ) is an optimal point of 5). 3 This section describes our main results on the function landscape of the non-square matrix sensing problem. The following theorem bounds the distance of any local minima to the global minimum, by the function value at the global minimum. Theorem.1. Suppose W is any target matrix of the optimization problem 5), under the balanced singular values assumption for U and V. If W is a critical point satisfying the first- and the secondorder optimality conditions, i.e., f + g)w ) 0 and f + g)w ) 0, then we have 1 5δ r 5δ r 1088δrδ r 80+68δ r)1+δ r) W W W W AU V ) b. 7) Observe that for this bound to make sense, the term 1 5δ r 5δ r 1088δrδ r 80+68δ r)1+δ r) needs to be positive. We provide some intuition of this result next. Combined with Lemma 5.1 in [3], we can also obtain the distance between U, V ) and U, V ). [ ] U Corollary.. or W and given the assumptions of Theorem.1, we V have σ r X ) 1 5δr 5δ r 1088δrδ r 100+68δ r)1+δ r) Dist U, V ; X ) AU V ) b. 8) Implications of these results are described next, where we consider specific settings. Remark [ 1 ]Noiseless matrix sensing). Suppose that U W V is the underlying unknown true matrix, i.e., X U V is rank-r and b AU V ). We assume the noiseless setting, w 0. If 0 δ r δ r 0.0363, then 1 5δr 5δ r 1088δrδ r 100+68δ r)1+δ r) > 0 in Corollary.. Since the RHS of 8) is zero, this further implies that Dist U, V ; X ) 0, i.e., any critical point W that satisfies first- and second-order optimality conditions is global minimum. Remark Noisy matrix sensing). Suppose that W is the underlying true matrix, such that X U V and is rank-r, and b AU V ) + w, for some noise term w. If 0 δ r δ r < 0.0, then it follows from 7) that for any local minima W the distance to U V is bounded by 1 W W W W 500 w.

Remark 3 High-rank matrix sensing). Suppose that X is of arbitrary rank and let Xr denote its best rank-r approximation. Let b AX ) + w where w is some noise and let U, V ) be a balanced factorization of Xr. If 0 δ r δ r < 0.005, then it follows from 8) that for any local minima U, V ) the distance to U, V ) is bounded by Dist U, V ; X ) 150 3σ rx ) AX X r ) + w. In plain words, the above remarks state that, given sensing mechanism A with small RIP constants, any critical point of the non-square matrix sensing objective with low rank optimum and no noise is a global minimum. As we describe in Section 3, due to this fact, gradient descent over the factors can converge, with high probability, to or very close to) the global minimum. 3 What About Saddle Points? Theorem 3. [30] - Informal). If the objective is twice differentiable and satisfies the strict saddle property, then gradient descent, randomly initialized and with sufficiently small step size, converges to a local minimum almost surely. In this section, based on the analysis in [5], we show that f + g satisfy the strict saddle property, which implies that gradient descent can avoid saddle points and converge to the global minimum, with high probability. Theorem 3.3. Consider noiseless measurements b AX ), with A satisfying RIP with constant δ r 1 100. Assume that rankx ) r. Let U, V ) be a pair of factors that satisfies the first [ order ] optimality condition fw ) 0, for W, and U V UV X. Then, λ min f + g)w ) ) 1 7 σ rx ). Our discussion so far concentrates on whether UV parametrization introduces spurious local minima. Our main results show that any point U, V ) that satisfies both first- and second-order optimality conditions 1 should be or lie close to) the global optimum. However, we have not discussed what happens with saddle points, i.e., points U, V ) where the Hessian matrix contains both positive and negative eigenvalues. This is important for practical reasons: first-order methods rely on gradient information and, thus, can easily get stuck to saddle points that may be far away from the global optimum. [0] studied conditions that guarantee that stochastic gradient descent randomly initialized converges to a local minimum; i.e., we can avoid getting stuck to non-degenerate saddle points. These conditions include f + g being bounded and smooth, having Lipschitz Hessian, being locally strongly convex, and satisfying the strict saddle property, according to the following definition. Definition 3.1. [0] A twice differentiable function f +g is strict saddle, if all its stationary points, that are not local minima, satisfy λ min f +g) )) < 0. [30] relax some of these conditions and prove the following theorem for standard gradient descent). 1 Note here that the second-order optimality condition includes positive semi-definite second-order information; i.e., Theorem.1 also handles saddle points due to the semi-definiteness of the Hessian at these points. Here, we do not consider the harder case where saddle points have Hessian with positive, negative and zero eigenvalues. Proof. Let Z R m+n) r. Then, by 10), the proof of Theorem.1 and the fact that b AX ) noiseless), f + g)w ) satisfies the following: vecz) f + g)w ) vecz) 13),1) 1 + δ r W e je j W W R) j1 3 8δr W W W W 16 1),15) 1+δ r 1 + 3 16δ ) 16 r) 3 8δ r 16 W W W W 1+5δ r+7δr +5δ3 r 8 W W W W 1 W W W W 9) 10 where the last inequality is due to the requirement δ r 1 100. or the LHS of 9), we can lower bound as follows: vecz) f + g)w ) vecz) Z λ min f + g)w ) ) W W R λ min f + g)w ) ) where the last equality is by setting Z W W R. Combining this expression with 9), we obtain: λ min f + g)w ) ) 1/10 W W W W W W R a) 1 /10 1) σ rx ) W W R W W R 1 7 σrx ),

where a) is due to Lemma 5., [3]. This completes the proof. Proof of Main Results We first describe the first- and second-order optimality conditions for f + g objective with W variable. Then, we provide a detailed proof of the main results: by carefully analyzing the conditions, we study how a local optimum is related to the global optimum..1 Gradient and Hessian of f and g The gradients of f and g w.r.t. W are given by: fw ) B i, W W b i) B i W gw ) 1 W W W 1 [ ] ) U U U V V ) V Regarding Hessian information, we are interested in the positive semi-definiteness of f + g); for this case, it is easier to write the second-order Hessian information with respect to to some matrix direction Z R m+n) r, as follows: vecz) fw ) vecz) lim and t 0 [ fw +tz) fw ) t ], Z B i, ZW + W Z B i, ZW + Bi, W W ) b i Bi, ZZ vecz) gw ) vecz) [ ] lim gw +tz) gw ) t 0 t, Z 1 Z W, ZW + 1 W Z, ZW W implies: [ U V ], we have f + g) W ) 0. This further f + g) W ) 0 Bi, W W ) b i Bi W + 1 W W W 0 11) Second-order optimality condition. or a point W that satisfies the second-order optimality condition f + g)w ) 0, 10) holds for any Z R m+n) r..3 Proof of Theorem.1 Suppose that W is a critical point satisfying the optimality conditions 11) and 10). As in [5], we sum up the above condition for Z 1 W W R)e 1 e 1,..., Z r W W R)e r e r. or simplicity, we first assume Z W W R. Bounding terms A), C) and D) in 10). The following bounds work for any Z. A) B i, ZW + B i, ZW B i, W Z a) 1 B i, ZW b) 1 + δr + A i, Z U V + Z U V Z U V, UZ V ) A i, UZV + 1 + δr UZ V + δ r Z U V UZV c) 1 + δr Z U V + 1 + δr UZ V }{{} + Z U V, UZV A) A1) + 1 W W, ZZ.. Optimality conditions Given the expressions above, we now describe firstand second-order optimality conditions on the composite objective f + g. irst-order optimality condition. By the firstorder optimality condition of a pair U, V ) such that 5 where a) follows from that every B i is symmetric, b) follows from Proposition 1., and c) follows from the AM-GM inequality. We also have C) Z W, ZW Z U U + Z V V Z U V Z V U, A1) + 1 1 + δr C) ZW,

vecz) f + g)w ) vecz) 0 B i, ZW + W Z B i, ZW + B i, W W ) b i B i, ZZ A) Z W, ZW + 1 C) B) W W, ZZ + 1 W Z, ZW + 1 }{{} D) E) 0, Z [ ] ZU R m+n) r. 10) Z V D) W Z, ZW UZU, Z U U + V ZV, Z V V UZV, Z U V V ZU, Z V U, A) + 1 D) 1 W Z, ZW, A) + 1 C) + 1 D) 1 W Z + ZW 8 ZW + δ r. and b) follows from Proposition 1.. Then we have B) E) W W, ZW ZZ W W, W W W RW W R W a) W W R)W W R) W W, W W W W W W, W W W W + W W, W W b) Bounding terms B) and E). B) a) We have B i, W W ) y i B i, ZZ B i, W W ) y i B i, W W W W 1 W W, W W R)W B i, W W 1 W W, W W R)W B i, W W ) y i B i, W W W W b) 1 δ r) UV U V 1 }{{} W W, ZW }{{} B1) B) B i, W W ) y i B i, W W W W B3) where at a) we add the first-order optimality equation B i, W W ) y i B i W, W W R 1 W W W, W W R, 6 c) W W, W W W W + W W, W W W W, W W W W W W, W W W W where a) follows from that W W is symmetric, b) follows from Proposition 1.3, c) follows from that the inner product of two PSD matrices is nonnegative. We then have, B1) + 1 B) 1 E) 1 δ r) UV U V + 1 W W W W, W W W W 1 δ r 1 ) UV U V + 1 UU U U + 1 V V V V 1 δr W W W W or B3), we have B3) B i, W W ) b i B i, W W W W a) AU V ) b 1 ) B i, W W W W b) 1 + δ r AU V ) b W W W W

where a) follows from the Cauchy-Schwarz inequality, and b) follows from Proposition 1.. We get B) + 1 E) 1 δr W W W W + 1 + δ r AU V ) b W W W W 3 8δr W W W W 16 + 161 + δ r) AU V ) b 1) where the last inequality follows from the AM-GM inequality. Summing up the inequalities for Z 1,..., Z r. Now we apply Z j Ze j e j. Since ZZ r j1 Z jzj in 10), the analysis does not change for B) and E). or A), C), and D), we obtain A) + 1 C) + 1 D) { 1 W Zj 8 j1 We have W Zj + Z jw j1 W e je j Z + j1 W e je j Z + j1 W e je j Z + j1 W e je j Z j1 + Z j W + δ r Z j W } W e je j Z, Ze je j W e j Z W e j) Ze j W e j where the inequality follows from the Cauchy- Schwarz inequality. Applying this bound, we get A) + 1 C) + 1 D) 1 + δ r W ej e j W W R). j1 Next, we re-state [5, Lemma.]: 13) Lemma.1. Let W and W be two matrices, and Q is an orthonormal matrix that spans the column space of W. Then, there exists an orthonormal matrix R such that, for any stationary point W of gw ) 7 that satisfies first and second order condition, the following holds: W e j e j W W R) j1 1 8 W W W W + 3 8 W W W W )QQ 1) And we have the following variant of [5, Lemma.]. Lemma.. or any pair of points U, V ) that satisfies the first-order optimality condition, and A be a linear operator satisfying the RIP condition with parameter δ r, the following inequality holds: 1 W W W W )QQ δ r W W W W 1 + δr + AU V ) b 15) Applying the above two lemmas, we can get A) + 1 C) + 1 D) 1+δ r) 1+1088δ r ) 16 W W W W + 31 + δ r)1 + δ r) AU V ) b. 16) inal inequality. Plugging 16) and 1) to 10), we get 1 5δ r 5δ r 1088δrδ r 80+68δ r)1+δ r) W W W W AU V ) b, which completes the proof. 5 Appendix: Proof of Lemma. The first-order optimality condition can be written as 0 f + g)w ), Z B i, W W ) b i B iw, Z + 1 W W W, Z B i, W W W W B i, ZW + + 1 B i, W W ) b i B i, ZW W W, ZW

1 + 1 A i, UV U V A i, Z U V + UZV A i, U V ) b i A i, Z U V + UZV + 1 W W, ZW, [ ] ZU Z R Z m+n) r. Applying Proposition 1. V and the Cauchy-Schwarz inequality to the condition, we obtain 1 UV U V, Z U V + UZV + 1 W W, ZW }{{} }{{} A) B) UV δ r U V ZU V + UZV }{{ } C) 1 + δr + AU V ) b Z U V + UZV }{{ } D) 17) Let Z W W W W )QR 1 where W QR is the QR decomposition. Then we obtain We have [ A) ZW W W W W )QQ. 0 UV U V V U V U 0 ], ZW W W W W ) W W W W ), W W W W )QQ and B) W W, W W W W )QQ W W, W W W W )QQ a) + W W, W W QQ b) W W, W W W W )QQ W W, W W W W )QQ W W W W, W W W W )QQ where a) follows from Proposition 1.3, and b) follows from that the inner product of two PSD matrices is non-negative. Then we obtain A) + B) W W W W, W W W W )QQ W W W W )Q W W W W )QQ 8 or C), we have C) UV U V ZU V + UZV 1 W W W W Z U V + UZ V W W W W ZW W W W W W W W W )QQ Plugging the above bounds into 17), we get 1 W W W W )QQ 1+δ r W δ r W W W W W W W )QQ + AU V ) b W W W W )QQ In either case of W W W W )QQ being zero or positive, we can obtain 1 W W W W )QQ δ r W W W W 1+δ + r AU V ) b This completes the proof. References [1] S. Aaronson. The learnability of quantum states. In Proceedings of the Royal Society of London A: Mathematical, Physical and Engineering Sciences, volume 63, pages 3089 311. The Royal Society, 007. [] A. Anandkumar and R. Ge. Efficient approaches for escaping higher order saddle points in non-convex optimization. arxiv preprint arxiv:160.05908, 016. [3] A. Bandeira, N. Boumal, and V. Voroninski. On the low-rank approach for semidefinite programs arising in synchronization and community detection. arxiv preprint arxiv:160.06, 016. [] S. Bhojanapalli, A. Kyrillidis, and S. Sanghavi. Dropping convexity for faster semi-definite optimization. arxiv preprint arxiv:1509.03917, 015.

[5] S. Bhojanapalli, B. Neyshabur, and N. Srebro. Global optimality of local search for low rank matrix recovery. arxiv preprint arxiv:1605.071, 016. [6] N. Boumal. Nonconvex phase synchronization. arxiv preprint arxiv:1601.0611, 016. [7] N. Boumal, V. Voroninski, and A. Bandeira. The non-convex Burer-Monteiro approach works on smooth semidefinite programs. arxiv preprint arxiv:1606.0970, 016. [8] S. Burer and R. Monteiro. A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization. Mathematical Programming, 95):39 357, 003. [9] S. Burer and R. Monteiro. Local minima and convergence in low-rank semidefinite programming. Mathematical Programming, 1033):7, 005. [10] E. Candes. The restricted isometry property and its implications for compressed sensing. Comptes Rendus Mathematique, 369):589 59, 008. [11] E. Candes, X. Li, Y. Ma, and J. Wright. Robust principal component analysis? Journal of the ACM JACM), 583):11, 011. [1] E. Candes and Y. Plan. Tight oracle inequalities for low-rank matrix recovery from a minimal number of noisy random measurements. Information Theory, IEEE Transactions on, 57):3 359, 011. [13] V. Chandrasekaran, S. Sanghavi, P. Parrilo, and A. Willsky. Sparse and low-rank matrix decompositions. In Communication, Control, and Computing, 009. Allerton 009. 7th Annual Allerton Conference on, pages 96 967. IEEE, 009. [1] Y. Chen and M. Wainwright. ast low-rank estimation by projected gradient descent: General statistical and algorithmic guarantees. arxiv preprint arxiv:1509.0305, 015. [15] A. Christoffersson. The one component model with incomplete data. Uppsala., 1970. [16] M. Davenport and J. Romberg. An overview of low-rank matrix recovery from incomplete observations. IEEE Journal of Selected Topics in Signal Processing, 10):608 6, 016. [17] C. De Sa, K. Olukotun, and C. Re. Global convergence of stochastic gradient descent for some non-convex matrix problems. arxiv preprint arxiv:111.113, 01. 9 [18] M. azel, E. Candes, B. Recht, and P. Parrilo. Compressed sensing and robust recovery of low rank matrices. In Signals, Systems and Computers, 008 nd Asilomar Conference on, pages 103 107. IEEE, 008. [19] S. lammia, D. Gross, Y.-K. Liu, and J. Eisert. Quantum tomography via compressed sensing: Error bounds, sample complexity and efficient estimators. New Journal of Physics, 19):0950, 01. [0] R. Ge,. Huang, C. Jin, and Y. Yuan. Escaping from saddle points online stochastic gradient for tensor decomposition. In Proceedings of The 8th Conference on Learning Theory, pages 797 8, 015. [1] R. Ge, J. Lee, and T. Ma. Matrix completion has no spurious local minimum. arxiv preprint arxiv:1605.077, 016. [] P. Jain, C. Jin, S. Kakade, and P. Netrapalli. Computing matrix squareroot via non convex local search. arxiv preprint arxiv:1507.0585, 015. [3] A. Javanmard and A. Montanari. Localization from incomplete noisy distance measurements. oundations of Computational Mathematics, 133):97 35, 013. [] C. Jin, S. Kakade, and P. Netrapalli. Provable efficient online matrix completion via nonconvex stochastic gradient descent. arxiv preprint arxiv:1605.08370, 016. [5] M. Journée,. Bach, P-A Absil, and R. Sepulchre. Low-rank optimization on the cone of positive semidefinite matrices. SIAM Journal on Optimization, 05):37 351, 010. [6] A. Kalev, R. Kosut, and I. Deutsch. Quantum tomography protocols with positivity are compressed sensing protocols. Nature partner journals npj) Quantum Information, 1:15018, 015. [7]. Krahmer and R. Ward. New and improved Johnson-Lindenstrauss embeddings via the restricted isometry property. SIAM Journal on Mathematical Analysis, 33):169 181, 011. [8] A. Kyrillidis and V. Cevher. Matrix recipes for hard thresholding methods. Journal of mathematical imaging and vision, 8):35 65, 01. [9] L. Le and M. White. Global optimization of factor models using alternating minimization. arxiv preprint arxiv:160.09, 016.

[30] J. Lee, M. Simchowitz, M. Jordan, and B. Recht. Gradient descent converges to minimizers. In Proceedings of The 9th Conference on Learning Theory, 016. [31] Y.-K. Liu. Universal low-rank matrix recovery from Pauli measurements. In Advances in Neural Information Processing Systems, pages 1638 166, 011. [3] Z. Liu and L. Vandenberghe. Interior-point method for nuclear norm approximation with application to system identification. SIAM Journal on Matrix Analysis and Applications, 313):135 156, 009. [33]. Mirzazadeh, Y. Guo, and D. Schuurmans. Convex co-embedding. In AAAI, pages 1989 1996, 01. [3]. Mirzazadeh, M. White, A. György, and D. Schuurmans. Scalable metric learning for co-embedding. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 65 6. Springer, 015. [35] D. Park, A. Kyrillidis, S. Bhojanapalli, C. Caramanis, and S. Sanghavi. Provable non-convex projected gradient descent for a class of constrained matrix optimization problems. arxiv preprint arxiv:1606.01316, 016. [36] D. Park, A. Kyrillidis, C. Caramanis, and S. Sanghavi. inding low-rank solutions to convex smooth problems via the Burer-Monteiro approach. In 5th Annual Allerton Conference on Communication, Control, and Computing, 016. [37] D. Park, A. Kyrillidis, C. Caramanis, and S. Sanghavi. inding low-rank solutions to matrix problems, efficiently and provably. arxiv preprint arxiv:1606.03168, 016. [38] B. Recht, M. azel, and P. Parrilo. Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM review, 53):71 501, 010. [39] A. Ruhe. Numerical computation of principal components when several observations are missing. Univ., 197. [0] J. Sun, Q. Qu, and J. Wright. Complete dictionary recovery over the sphere I: Overview and the geometric picture. arxiv preprint arxiv:1511.03607, 015. [1] J. Sun, Q. Qu, and J. Wright. A geometric analysis of phase retrieval. arxiv preprint arxiv:160.0666, 016. 10 [] R Sun and Z.-Q. Luo. Guaranteed matrix completion via nonconvex factorization. In IEEE 56th Annual Symposium on oundations of Computer Science, OCS 015, pages 70 89, 015. [3] S. Tu, R. Boczar, M. Soltanolkotabi, and B. Recht. Low-rank solutions of linear matrix equations via Procrustes flow. arxiv preprint arxiv:1507.03566v, 016. [] A. Waters, A. Sankaranarayanan, and R. Baraniuk. SpaRCS: Recovering low-rank and sparse matrices from compressive measurements. In Advances in neural information processing systems, pages 1089 1097, 011. [5] H. Wold and E. Lyttkens. Nonlinear iterative partial least squares NIPALS) estimation procedures. Bulletin of the International Statistical Institute, 31), 1969. [6] Xinyang Yi, Dohyung Park, Yudong Chen, and Constantine Caramanis. ast algorithms for robust PCA via gradient descent. arxiv preprint arxiv:1605.0778, 016. [7] X. Zhang, D. Schuurmans, and Y. Yu. Accelerated training for matrix-norm regularization: A boosting approach. In Advances in Neural Information Processing Systems, pages 906 91, 01. [8] T. Zhao, Z. Wang, and H. Liu. A nonconvex optimization framework for low rank matrix estimation. In Advances in Neural Information Processing Systems 8, pages 559 567. 015. [9] Q. Zheng and J. Lafferty. A convergent gradient descent algorithm for rank minimization and semidefinite programming from random linear measurements. In Advances in Neural Information Processing Systems, pages 109 117, 015. [50] Q. Zheng and J. Lafferty. Convergence analysis for rectangular matrix completion using burer-monteiro factorization and gradient descent. arxiv preprint arxiv:1605.07051, 016.