Riemannian Pursuit for Big Matrix Recovery

Size: px
Start display at page:

Download "Riemannian Pursuit for Big Matrix Recovery"

Transcription

1 Riemannian Pursuit for Big Matrix Recovery Mingkui Tan 1, Ivor W. Tsang 2, Li Wang 3, Bart Vandereycken 4, Sinno Jialin Pan 5 1 School of Computer Science, University of Adelaide, Australia 2 Center for Quantum Computation & Intelligent Systems, UTS, Australia 3 Department of Mathematics, University of California, USA 4 Department of Mathematics, Princeton University, USA 5 Institute for Infocomm Research, SG August 18, / 36

2 Outline Introduction to Low Rank Matrix Recovery Problem Formulation Existing MR Methods Riemannian Optimization on Fixed-rank Manifold Riemannian Pursuit for Matrix Recovery Motivations Our Contributions Proposed Algorithm Main Theoretical Results Stopping Conditions of RP Conjugate Gradient Descent on Manifold Experiments on Matrix Completion Conclusions 2 / 36

3 Notations We will use the following notations: Given a linear operator A, denote its adjoint operator by A. For instance, if A is a matrix A, its adjoint operator is A Let X = U(diag(σ))V be the SVD of X R m n. The nuclear norm of X is defined as X = σ 1 = i σ i. The condition number κ r (X) of X w.r.t. a given number r is defined as κ r (X) = σ 1 /σ r 3 / 36

4 Low Rank Assumption Recovering X from partial observations b is impossible in general. Low-rank assumption: rank( X) r, where r min(m, n). Faces from a same person lie on a low-dimensional manifold (Yan, et al., 2007). In collaborative filtering tasks, items from same group may have similar actions. 4 / 36

5 Problem Formulation Given a linear operator A : R m n R l, let b = A( X) + e be l linear measurements of an unknown rank- r matrix X R m n, where e denotes noise. Matrix recover tries to recover X by solving min f(x), s.t. rank(x) r, (1) X where l mn, r r, and f(x) = 1 2 b A(X) 2 2. The definition of the operator A depends on specific application: Matrix completion [Recht et al.(2010)]: Quantum state tomography [Candés & Plan(2010a)] Matrix learning & factorizations [Laue(2012)] Low rank structure learning or clustering [Deng et al.(2013)]... 5 / 36

6 Existing MR Methods MR by minimizing f(x) with rank constraint is known to be NP-hard. How to estimate r? rank(x) r Estimate rank(x) by cross-validation? A good idea, but, how to measure the performance of a specific parameter? Very difficult! Existing methods: 1. Nuclear-norm convex relaxations [Candés & Plan(2010a)]: Replace the rank constraint by X v for some v > Fixed-rank methods (relaxations): Assume rank(x) = r, where r is supposed to be known. 3. Other methods: Max-norm based methods, p-norm non-convex methods (0 < p < 1), and so on. 6 / 36

7 Nuclear-norm Convex Relaxations Two kinds of trace-norm relaxations are usually studied: Nuclear-norm minimization with equality constraint min X X, s.t. A(X) = b, (2) Typical methods: Singular Value Thresholding (SVT) [Cai et al.(2010)] Augmented Lagrangian method (ALM) [Lin et al.(2010)] Matrix lasso problem: min X,ξ λ X ξ 2 2 : ξ = b A(X), (3) Typical methods: Accelerated Proximal Gradient (APG) [Toh & Yun(2010)] Stochastic Gradient methods 7 / 36

8 Nuclear-norm Convex Relaxations Advantages: A global solution can be obtained by convex optimization methods (but may not be unique). Good RIP guarantees for exact matrix recovery, namely, γ r 1/3 (Cai et al., 2013). Disadvantages: Require many high-dimensional SVDs. Very expensive for large-scale problems. Has solution bias due to the nuclear-norm regularization [Vandereycken(2013)]. 8 / 36

9 Challenges The key is how to avoid high dimensional SVDs and estimate the correct rank 9 / 36

10 Fixed-rank Methods The fixed-rank methods solve the following problem: min f(x), s.t. rank(x) = r, (4) X where r is supposed to be known. Non-convex since the constraint rank(x) = r defines a nonlinear smooth matrix manifold [Meyer et al.(2011)]. Many efficient local-search methods have been proposed: Greedy methods: Singular Value Projection (SVP) [Meka et al.(2009b)] Atomic Decomposition for Minimum Rank Approximation (ADMiRA) [Lee & Bresler(2010)] Stochastic Gradient methods [Wen et al.(2012)]. Manifold optimization methods: Low-rank geometric conjugate gradient method (LRGeomCG) [Vandereycken(2013)] The quotient geometric matrix completion method (qgeommc) [Mishra et al.(2012)] 10 / 36

11 Differential Geometry of Fixed-rank Matrices Smooth manifold M r of fixed-rank matrices: M r = {X R m n : rank(x) = r} = {Udiag(σ)V : U St m r, V St n r, σ 0 = r}, Stiefel manifold: m r real and orthonormal matrices: St m r = {U R m r : U U = I}. The tangent space T X M r of M r at X = Udiag(σ)V R m n : T X M r = {UMV +U p V +UVp : M R r r, U p R m r, U p U = 0, V p R n r, Vp V = 0}. 11 / 36

12 Differential Geometry of Fixed-rank Matrices Riemannian gradient of f on M r : Given a smooth function f : M r R, the Riemannian gradient is the orthogonal projection of the gradient of f onto the tangent space. Define P U = UU and P U = I UU for any U St m r, and the orthogonal projection of any Z R m n onto the tangent space at X = Udiag(σ)V is defined as Retraction: E = P TX M r (Z) : Z P U ZP V + P U ZP V + P U ZP V. (5) The Retraction mapping lets a point X + E in the tangent space go back to the manifold: r R X (E) = P Mr (X + E) = σ i p i q i, (6) where r i=1 σ ip i q i denotes the best rank-r approximation to X + E by the SVD. Retraction can be calculated with O((m + n)r 2 ) cost. i=1 12 / 36

13 Fixed-rank Methods Advantages of Fixed-rank Methods: Fixed-rank Methods have superior scalability compared with the nuclear-norm based methods [Boumal & Absil(2012), Mishra et al.(2012), Vandereycken(2013)]. Greedy methods need truncated SVD of known low rank only. Some manifold optimization methods involve simple matrix-products, which is particularly important for parallel computing. Disadvantages How do we know the rank r? It is nontrivial. Greedy methods require restricted conditions to converge. Convergence Issues: If X is in ill-conditioning, existing fixed-rank methods may converge very slowly [Ngo & Saad(2012)]. 13 / 36

14 Motivations Fixed-rank methods have gained great success in solving big MR problems, but it requires the explicit knowledge of the rank r. Three Questions: 1. Can we avoid the rank estimation by iteratively increase the rank by a fixed integer ρ? 2. If using this procedure, how can we stop it? 3. Is this procedure helpful to avoid the convergence issue on ill-conditioned problems? 14 / 36

15 Main Contributions We propose a Riemannian Pursuit (RP) method, which indeed increases the rank by ρ and essentially solves a sequence of fixed-rank minimization problems. RP converges linearly under mild conditions. RP can effectively address the convergence issues that occur with ill-conditioned and large rank problems. RP can automatically estimates the rank under proper stopping conditions. 15 / 36

16 Active Subspace Search Let G = A (A(X) b). The gradient of f(x) on M r can be calculated by gradf(x) = P TX M r (G). (7) The gradient direction orthogonal to M r is Q = G P TX M r (G) Select the subspace of rank ρ from Q Add this subspace into the original manifold 16 / 36

17 Riemannian Pursuit Algorithm 1 Riemannian Pursuit for MR. 1: Inner iteration tolerance ɛ in, and outer iteration tolerance ɛ out. Initialize X 0 = 0, ξ 0 = b and G = A (ξ 0 ). Let t = 1 and r = ρ. 2: Perform an active-subspace search as follows. (2a): Compute Q = G P TX M r (G). (2b): Compute a best rank ρ approximation of Q: H t 1 2 = U ρ diag(σ ρ )V ρ 3: Let H t 1 1 = P TX M r (G) and H t 1 = H1 t 1 + H t 1 (3a): Choose a proper step size τ t and set 2. X intial = R X ( τ t H t 1 ). (Warm Start) (3b): Update X t by X t =NRCG(X intial, ɛ in ). 4: Update r = r + ρ, ξ t = b A(X t ) and G = A (ξ t ). 5: Quit if stopping conditions achieve, otherwise, let t = t+1, and go to Step / 36

18 Main Theoretical Results Lemma 1: Firstly, let {X t } be the sequence generated by RP, then f(x t ) decreases monotonically w.r.t. t. Theorem 1: Let {X t } be the sequence generated by RP, as long as f(x t ) > f( X) = C 2 e 2 (where C > 1), and there exists an integer ι > 0 such that γ r+2ιρ < 1/2, then RP decreases linearly in objective values when t < ι, namely f(x t+1 ) νf(x t ), where ν = ( 1 ρζ 2r ( C(1 2γ (r+2ιρ) ) 2 ( C+1) 2 (1 γ (r+2ιρ) ) ) (1 1 C ) 2 ). 18 / 36

19 Stopping Conditions of RP RP monotonically decreases the objective values. Without proper stopping conditions, it may increase the rank until tρ = min(m, n), leading to over-fitting issue. To avoid this, we propose to use the averaged function value difference between iterations as the stopping condition, namely 2(f(X t 1 ) f(x t ))/(ρ b 2 2) ɛ out, (8) where ɛ out is a predefined tolerance value. If the increasing of rank does not significantly reduce the objective value, we stop it to avoid over-fitting. When it is stopped, a rank is returned! 19 / 36

20 Riemannian Pursuit Algorithm 2 Riemannian Pursuit for MR. 1: Inner iteration tolerance ɛ in, and outer iteration tolerance ɛ out. Initialize X 0 = 0, ξ 0 = b and G = A (ξ 0 ). Let t = 1 and r = ρ. 2: Perform an active-subspace search as follows. (2a): Compute Q = G P TX M r (G). (2b): Compute a best rank ρ approximation of Q: H t 1 2 = U ρ diag(σ ρ )V ρ 3: Let H t 1 1 = P TX M r (G) and H t 1 = H1 t 1 + H t 1 (3a): Choose a proper step size τ t and set 2. X intial = R X ( τ t H t 1 ). (Warm Start) (3b): Update X t by X t =NRCG(X intial, ɛ in ). 4: Update r = r + ρ, ξ t = b A(X t ) and G = A (ξ t ). 5: Quit if stopping conditions achieve, otherwise, let t = t+1, and go to Step / 36

21 Fixed-rank subproblem Step (3b) (Update X t by X t =NRCG(X intial, ɛ in )) is to solving the following problem: min X f(x), s.t. rank(x) = tρ. (9) ρ is much smaller than r. When t is small, tρ < r. Smaller condition number: κ tρ (X) = σ 1 /σ tρ < κ r (X). Faster convergence speed with a small condition number. Selection of ρ is important. Setting ρ = 1 is the simplest way, but ρ > 1 is better. 21 / 36

22 Nonlinear Conjugate Gradient Descent Figure: Basic elements of Riemannian manifold on matrices. 22 / 36

23 Nonlinear Conjugate Gradient Descent The nonlinear conjugate gradient descent algorithm is shown as follows: Algorithm 3 NRCG for solving the fixed-rank minimization problem. 1: Initialize X 1 = X intial and ζ 0 = 0. Let k = 1. 2: Compute the Riemannian gradient E k = gradf(x k ). 3: Compute the conjugate direction ζ k according to ζ k = P k + β t T Xk 1 X k (ζ k 1 ). 4: Choose a step size θ k satisfying the strong Wolfe conditions, and set X k+1 = R Xk (θ k ζ k ). 5: Terminate and output X k+1 if the stopping conditions are achieved; otherwise, let k = k + 1 and go to step / 36

24 Experiments Toy Experiments for Convergence Comparison Real-world Experiments on Collaborative Filtering Tasks 24 / 36

25 Toy Experimental Settings Methods for comparison: LRGeomCG, qgeommc, ScGrassMC, SVP, ADMiRA, SpaRCS, GECO and APG. Except GECO and APG, others are fixed-rank methods. Toy experimental setting We generate ground-truth X by X = Ûdiag( σ) V + e, where σ is a r-sparse vector, Û Stm r, and V St n r. Two kinds of singular values are studied: 1) Gaussian sparse singular value vector s g sampled from the Gaussian distribution N(0, 1000); 2) χ 2 sparse singular value vector s 2 χ, where each entry is the square of s g. 25 / 36

26 Toy Experiments for Convergence Comparison Relative Objective Value GECO ScGrassMC LRGeomCG LMaFit qgeommc RP(η = 1.00) RP(η = 0.75) RP(η = 0.65) RP(η = 0.55) Time (in seconds) Time (in seconds) (a) Relative objective values on (b) Relative objective values on s g, where ρ w.r.t. different η s is s 2 χ, where ρ for different η s is 1, 1, 8, 14, 18, respectively. 4, 6, 12, respectively. Relative Objective Value ScGrassMC LRGeomCG LMaFit qgeommc RP(η = 1.00) RP(η = 0.75) RP(η = 0.65) RP(η = 0.55) Figure: Convergence of comparison methods on s g and s χ 2. RP with different ρ converges well on ill-conditioned problems. Other algorithms have convergence issues. GECO cannot converge within 1 hour on s g, and we omit its results on s χ 2. ScGrassMC gets numerical problems after 50 iterations on s χ / 36

27 Experiments on Real-world Collaborative Filtering Tasks Table: Experimental results on real-world datasets. Result of APG on Netflix is absent. The average ranks estimated by APG, Lmafit-A and RP on Movie-10M are 100, 77 and 10, respectively. The average ranks estimated by Lmafit-A and RP on Netflix are 81 and 12, respectively Dataset Movie-10M Netflix-100M RMSE Time (seconds) RMSE Time (seconds) APG ± LRGeomCG ± ±35 QgeomMC ± ± 74 Lmafit ± ±50 Lmafit-A ± ±165 RP ± ±27 RP performs best among all the methods in terms of RMSE and computational efficiency. We use the rank returned by RP as the rank for LRGeomCG, qgeommc and Lmafit, thus RP is much faster if the model selection cost is considered. 27 / 36

28 Comparison on MovieLens with 10M ratings Table: Performance comparison on Movie-10M dataset. Methods Time (in seconds) SpeedUp RMSE CPU(GHz) GECO 784,941 9,000x Laue 2,663 30x Jaggi 3,120 38x RP / 36

29 Conclusions We propose a Riemannian Pursuit (RP) for tackling Big Matrix Recovery problems. By exploiting the Riemannian geometry on the fixed-rank manifolds, high-dimensional SVDs in the master problem of RP are avoided, which exhibits good scalability for large-scale problems. RP converges linearly under mild conditions. RP can automatically estimate the rank and effectively address the convergence issues on ill-conditioned and large rank matrices. Extensive experimental results show that RP achieves superb scalability on Big Matrices using a single PC; while it maintains similar or better MR performance. 29 / 36

30 Thank you for your attention! Our Poster is at T4! 30 / 36

31 Conjugate Gradient Descent on Manifold Steepest gradient descent in general is very slow. Conjugate gradient descent is a good choice. Let ζ k 1 be a search direction at (k 1) th, we compute the conjugate search direction: ζ t = gradf(x k ) + β t ζ k 1, (10) where β t can be calculated by the Fletcher-Reeves (F-R) rule. But notice: gradf(x k ) and β t ζ k 1 are in different tangent spaces! Therefore, Eq. (10) is not valid! We still need an operator: Vector Transport. 31 / 36

32 Conjugate Gradient Descent on Manifold A Vector Transport T on a manifold M r is a smooth map which transports tangent vectors from one tangent space to another. A vector transport T on a manifold M r is a smooth map [Absil et al.(2008)]: T M r T M r T M r : (ζ X, ν X ) T ζx (ν X ) satisfying the following properties for all X M r : There exists a retraction R associated with T such that for all (ζ X, ν X ) T M r T M r, it holds that T ζx (ν X ) T RX M r. T0 (ν) = ν for all ν T X M r. Tζ (aν X + bω X ) = X atη X (ν X ) + btη X (ω X ) for ν X T M r and ω X T M r. Very tedious definition, but the computation takes only O(m + n)r 2 cost. 32 / 36

33 Conjugate Gradient Descent on Manifold With Vector Transport T, the conjugate search direction can be calculated by ζ k = P k + β t T Xk 1 X k (ζ k 1 ), (11) Figure: Conjugate search direction computation. 33 / 36

34 Other Issues in NRCG Line search for the step size θ: Armijo line search: The convergence cannot be guaranteed. Strong Wolfe line search: The convergence can be guaranteed. Initialization with Warm Start: Set X 1 = X intial, where X intial is from Algorithm / 36

35 Riemannian Pursuit Algorithm 4 Riemannian Pursuit for MR. 1: Inner iteration tolerance ɛ in, and outer iteration tolerance ɛ out. Initialize X 0 = 0, ξ 0 = b and G = A (ξ 0 ). Let t = 1 and r = ρ. 2: Perform an active-subspace search as follows. (2a): Compute Q = G P TX M r (G). (2b): Compute a best rank ρ approximation of Q: H t 1 2 = U ρ diag(σ ρ )V ρ 3: Let H t 1 1 = P TX M r (G) and H t 1 = H1 t 1 + H t 1 (3a): Choose a proper step size τ t and set 2. X intial = R X ( τ t H t 1 ). (Warm Start) (3b): Update X t by X t =NRCG(X intial, ɛ in ). 4: Update r = r + ρ, ξ t = b A(X t ) and G = A (ξ t ). 5: Quit if stopping conditions achieve, otherwise, let t = t+1, and go to Step / 36

36 Other Issues in NRCG Line search for the step size θ: Armijo line search: The convergence cannot be guaranteed. Strong Wolfe line search: The convergence can be guaranteed. Warm start for initialization: Set X 1 = X intial, where X intial is from Algorithm 1. Stopping condition of NRCG: f(x k 1 ) f(x k ) f(x k 1 ) ɛ in, (12) No need to solve the fixed-rank problem with high precision! We set ɛ in = 0.01 in practice. All these techniques are important to improve the efficiency. Please find more details in the paper. 36 / 36

37 Absil, P.-A. and Mahony, R. and Sepulchre, R. Optimization Algorithms on Matrix Manifolds. Princeton University Press, Boumal, N. and Absil, P.-A. Rtrmc: A riemannian trust-region method for low-rank matrix completion. In NIPS, Cai, J., Candés, J., E., and Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM J. on Optim., 20(4): , Candés, E. J. and Plan, Y. Tight oracle bounds for low-rank matrix recovery from a minimal number of random measurements. IEEE Trans. on Inform. Theory, 57(4): , 2010a. Candés, E. J. and Plan, Y. Matrix completion with noise. Proceedings of the IEEE, 98(6): , 2010b. Candés, E. J. and Recht, B. 36 / 36

38 Exact matrix completion via convex optimization. Found. Comput. Math., 9: , Donoho, D. L., Tsaig, Y., Drori, I., and Starck, J. L. Sparse solution of underdetermined systems of linear equations by stagewise orthogonal matching pursuit. IEEE Trans. Info. Theory, 58(2): , Fazel, M. Matrix rank minimization with applications PhD thesis, Stanford University. Golub, G. H. and Van Loan, C. F. Matrix computations. JHU Press, 3rd edition, Hazan, E. Sparse approximate solutions to semidefinite programs. LATIN, pp , Herlocker, J. L., Konstan, J. A., Borchers, A., and Riedl, J. An algorithmic framework for performing collaborative filtering. 36 / 36

39 In SIGIR, Jaggi, M. and Sulovsky, M. A simple algorithm for nuclear norm regularized problems. In ICML, KDDCup. ACM SIGKDD and netflix. In Proceedings of KDD Cup and Workshop, Keshavan, R. H., Montanari, A., and Oh, S. Matrix completion from a few entries. IEEE Trans. on Info. Theory, 56: , 2010a. Keshavan, R. H, Montanari, A., and Oh, S. Matrix completion from noisy entries. JMLR, 99: , 2010b. Laue, S. A hybrid algorithm for convex semidefinite optimization. In ICML, Lee, K. and Bresler, Y. Admira: Atomic decomposition for minimum rank approximation. 36 / 36

40 IEEE Trans. on Inform. Theory, 56(9): , Y. Deng, Q. Dai, R. Liu, Z. Zhang, and S. Hu. Low-rank structure learning via nonconvex heuristic recovery. IEEE Trans. Neural Netw. Learning Syst., 24(3): , Lin, Z., Chen, M., and Ma, Y. The augmented lagrange multiplier method for exact recovery of corrupted low-rank matrices. Technical report, UIUC, Lin, Z., Liu, R., and Su, Z. Linearized alternating direction method with adaptive penalty for low-rank representation. arxiv preprint arxiv: , Meka, R., Jain, P., and Dhillon, I. S. Guaranteed rank minimization via singular value projection. Technical report, 2009a. Meka, R., Jain, P., and I.S.Dhillon. Guaranteed rank minimization via singular value projection. In NIPS, 2009b. 36 / 36

41 Meyer, G., Bonnabel, S., and Sepulchre, R. Linear regression under fixed-rank constraints: A riemannian approach. In ICML, Mishra, B., Apuroop, K. A., and Sepulchre, R. A riemannian geometry for low-rank matrix completion. Technical report, Mishra, B., Meyer, G., Bach, F., and Sepulchre, R. Low-rank optimization with trace norm penalty. SIAM J. Optim., 23(4): , Mitra, K., Sheorey, S., and Chellappa, R. Large-scale matrix factorization with missing data under additional constraints. In NIPS, Negahban, S. and Wainwright, M. J. Restricted strong convexity and weighted matrix completion: Optimal bounds with noise. JMLR, 13: , / 36

42 Ngo, T. T. and Saad, Y. Scaled gradients on grassmann manifolds for matrix completion. In NIPS, Recht, B. A simpler approach to matrix completion. JMLR, pp , Recht, B., Fazel, M., and Parrilo, P. A. Guaranteed minimum rank solutions to linear matrix equations via nuclear norm minimization. SIAM Rev., 52(3), Ring, W. and Wirth, B. Optimization methods on riemannian manifolds and their application to shape space. SIAM J. Optim., 22(2): , Sato, H. and Iwai, T. A new, globally convergent riemannian conjugate gradient method. Optimization: A Journal of Mathematical Programming and Operations Research, (ahead-of-print):1 21, / 36

43 Selvan, S. E., Amato, U., Gallivan, K. A., Qi, Ch., Carfora, M. F., Larobina, M., and Alfano, B. Descent algorithms on oblique manifold for source-adaptive ica contrast. IEEE Trans. Neural Netw. Learning Syst., 23(12): , Shalit, U., Weinshall, D., and Chechik, G. Online learning in the embedded manifold of low-rank matrices. JMLR, 13: , Shwartz, S. S., Gonen, A., and Shamir, O. Large-scale convex minimization with a low-rank constraint. In ICML, Toh, K.-C. and Yun, S. An accelerated proximal gradient algorithm for nuclear norm regularized least squares problems. Pac. J. Optim., 6: , Vandereycken, B. Low-rank matrix completion by Riemannian optimization. Technical report, / 36

44 Siam J. Optim., 23(2): , Waters, A. E., Sankaranarayanan, A. C., and Baraniuk, Richard G. Sparcs: Recovering low-rank and sparse matrices from compressive measurements. In NIPS, Wen, Z., Yin, W., and Zhang, Y. Solving a low-rank factorization model for matrix completion by a non-linear successive over-relaxation algorithm. Math. Program. Comput., 4(4): , Yang, J. and Yuan, X. Linearized augmented lagrangian and alternating direction methods for nuclear norm minimization. Mathematics of Computation, 82(281): , / 36

Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery

Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery Mingkui Tan, School of omputer Science, The University of Adelaide, Australia Ivor W. Tsang, IS, University of Technology Sydney,

More information

Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison

Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison Raghunandan H. Keshavan, Andrea Montanari and Sewoong Oh Electrical Engineering and Statistics Department Stanford University,

More information

Sparse and low-rank decomposition for big data systems via smoothed Riemannian optimization

Sparse and low-rank decomposition for big data systems via smoothed Riemannian optimization Sparse and low-rank decomposition for big data systems via smoothed Riemannian optimization Yuanming Shi ShanghaiTech University, Shanghai, China shiym@shanghaitech.edu.cn Bamdev Mishra Amazon Development

More information

Matrix Completion: Fundamental Limits and Efficient Algorithms

Matrix Completion: Fundamental Limits and Efficient Algorithms Matrix Completion: Fundamental Limits and Efficient Algorithms Sewoong Oh PhD Defense Stanford University July 23, 2010 1 / 33 Matrix completion Find the missing entries in a huge data matrix 2 / 33 Example

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Ranking from Crowdsourced Pairwise Comparisons via Matrix Manifold Optimization

Ranking from Crowdsourced Pairwise Comparisons via Matrix Manifold Optimization Ranking from Crowdsourced Pairwise Comparisons via Matrix Manifold Optimization Jialin Dong ShanghaiTech University 1 Outline Introduction FourVignettes: System Model and Problem Formulation Problem Analysis

More information

Matrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University

Matrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University Matrix completion: Fundamental limits and efficient algorithms Sewoong Oh Stanford University 1 / 35 Low-rank matrix completion Low-rank Data Matrix Sparse Sampled Matrix Complete the matrix from small

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds

Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds Tao Wu Institute for Mathematics and Scientific Computing Karl-Franzens-University of Graz joint work with Prof.

More information

Low Rank Matrix Completion Formulation and Algorithm

Low Rank Matrix Completion Formulation and Algorithm 1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9

More information

An accelerated proximal gradient algorithm for nuclear norm regularized linear least squares problems

An accelerated proximal gradient algorithm for nuclear norm regularized linear least squares problems An accelerated proximal gradient algorithm for nuclear norm regularized linear least squares problems Kim-Chuan Toh Sangwoon Yun March 27, 2009; Revised, Nov 11, 2009 Abstract The affine rank minimization

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,

More information

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Lingchen Kong and Naihua Xiu Department of Applied Mathematics, Beijing Jiaotong University, Beijing, 100044, People s Republic of China E-mail:

More information

Subspace Projection Matrix Completion on Grassmann Manifold

Subspace Projection Matrix Completion on Grassmann Manifold Subspace Projection Matrix Completion on Grassmann Manifold Xinyue Shen and Yuantao Gu Dept. EE, Tsinghua University, Beijing, China http://gu.ee.tsinghua.edu.cn/ ICASSP 2015, Brisbane Contents 1 Background

More information

An iterative hard thresholding estimator for low rank matrix recovery

An iterative hard thresholding estimator for low rank matrix recovery An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical

More information

Fast Hybrid Algorithm for Big Matrix Recovery

Fast Hybrid Algorithm for Big Matrix Recovery Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI-6) Fast Hybrid Algorithm for Big Matrix Recovery Tengfei Zhou, Hui Qian, Zebang Shen, Congfu Xu College of Computer Science

More information

Guaranteed Rank Minimization via Singular Value Projection

Guaranteed Rank Minimization via Singular Value Projection 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 3 3 33 34 35 36 37 38 39 4 4 4 43 44 45 46 47 48 49 5 5 5 53 Guaranteed Rank Minimization via Singular Value Projection Anonymous Author(s) Affiliation Address

More information

Solving Corrupted Quadratic Equations, Provably

Solving Corrupted Quadratic Equations, Provably Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin

More information

Adaptive one-bit matrix completion

Adaptive one-bit matrix completion Adaptive one-bit matrix completion Joseph Salmon Télécom Paristech, Institut Mines-Télécom Joint work with Jean Lafond (Télécom Paristech) Olga Klopp (Crest / MODAL X, Université Paris Ouest) Éric Moulines

More information

Matrix Completion from Noisy Entries

Matrix Completion from Noisy Entries Matrix Completion from Noisy Entries Raghunandan H. Keshavan, Andrea Montanari, and Sewoong Oh Abstract Given a matrix M of low-rank, we consider the problem of reconstructing it from noisy observations

More information

Elaine T. Hale, Wotao Yin, Yin Zhang

Elaine T. Hale, Wotao Yin, Yin Zhang , Wotao Yin, Yin Zhang Department of Computational and Applied Mathematics Rice University McMaster University, ICCOPT II-MOPTA 2007 August 13, 2007 1 with Noise 2 3 4 1 with Noise 2 3 4 1 with Noise 2

More information

A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices

A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices Ryota Tomioka 1, Taiji Suzuki 1, Masashi Sugiyama 2, Hisashi Kashima 1 1 The University of Tokyo 2 Tokyo Institute of Technology 2010-06-22

More information

The convex algebraic geometry of rank minimization

The convex algebraic geometry of rank minimization The convex algebraic geometry of rank minimization Pablo A. Parrilo Laboratory for Information and Decision Systems Massachusetts Institute of Technology International Symposium on Mathematical Programming

More information

Distributed Inexact Newton-type Pursuit for Non-convex Sparse Learning

Distributed Inexact Newton-type Pursuit for Non-convex Sparse Learning Distributed Inexact Newton-type Pursuit for Non-convex Sparse Learning Bo Liu Department of Computer Science, Rutgers Univeristy Xiao-Tong Yuan BDAT Lab, Nanjing University of Information Science and Technology

More information

A Scalable Spectral Relaxation Approach to Matrix Completion via Kronecker Products

A Scalable Spectral Relaxation Approach to Matrix Completion via Kronecker Products A Scalable Spectral Relaxation Approach to Matrix Completion via Kronecker Products Hui Zhao Jiuqiang Han Naiyan Wang Congfu Xu Zhihua Zhang Department of Automation, Xi an Jiaotong University, Xi an,

More information

Low rank matrix completion by alternating steepest descent methods

Low rank matrix completion by alternating steepest descent methods SIAM J. IMAGING SCIENCES Vol. xx, pp. x c xxxx Society for Industrial and Applied Mathematics x x Low rank matrix completion by alternating steepest descent methods Jared Tanner and Ke Wei Abstract. Matrix

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via nonconvex optimization Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

STUDY of genetic variations is a critical task to disease

STUDY of genetic variations is a critical task to disease Haplotype Assembly Using Manifold Optimization and Error Correction Mechanism Mohamad Mahdi Mohades, Sina Majidian, and Mohammad Hossein Kahaei arxiv:9.36v [math.oc] Jan 29 Abstract Recent matrix completion

More information

Rank-Constrainted Optimization: A Riemannian Manifold Approach

Rank-Constrainted Optimization: A Riemannian Manifold Approach Ran-Constrainted Optimization: A Riemannian Manifold Approach Guifang Zhou1, Wen Huang2, Kyle A. Gallivan1, Paul Van Dooren2, P.-A. Absil2 1- Florida State University - Department of Mathematics 1017 Academic

More information

Convex Optimization Algorithms for Machine Learning in 10 Slides

Convex Optimization Algorithms for Machine Learning in 10 Slides Convex Optimization Algorithms for Machine Learning in 10 Slides Presenter: Jul. 15. 2015 Outline 1 Quadratic Problem Linear System 2 Smooth Problem Newton-CG 3 Composite Problem Proximal-Newton-CD 4 Non-smooth,

More information

Low-Rank Matrix Recovery

Low-Rank Matrix Recovery ELE 538B: Mathematics of High-Dimensional Data Low-Rank Matrix Recovery Yuxin Chen Princeton University, Fall 2018 Outline Motivation Problem setup Nuclear norm minimization RIP and low-rank matrix recovery

More information

Bias-free Sparse Regression with Guaranteed Consistency

Bias-free Sparse Regression with Guaranteed Consistency Bias-free Sparse Regression with Guaranteed Consistency Wotao Yin (UCLA Math) joint with: Stanley Osher, Ming Yan (UCLA) Feng Ruan, Jiechao Xiong, Yuan Yao (Peking U) UC Riverside, STATS Department March

More information

Spectral k-support Norm Regularization

Spectral k-support Norm Regularization Spectral k-support Norm Regularization Andrew McDonald Department of Computer Science, UCL (Joint work with Massimiliano Pontil and Dimitris Stamos) 25 March, 2015 1 / 19 Problem: Matrix Completion Goal:

More information

Lecture Notes 10: Matrix Factorization

Lecture Notes 10: Matrix Factorization Optimization-based data analysis Fall 207 Lecture Notes 0: Matrix Factorization Low-rank models. Rank- model Consider the problem of modeling a quantity y[i, j] that depends on two indices i and j. To

More information

Exact Low-Rank Matrix Completion from Sparsely Corrupted Entries via Adaptive Outlier Pursuit

Exact Low-Rank Matrix Completion from Sparsely Corrupted Entries via Adaptive Outlier Pursuit Noname manuscript No. (will be inserted by the editor) Exact Low-Rank Matrix Completion from Sparsely Corrupted Entries via Adaptive Outlier Pursuit Ming Yan Yi Yang Stanley Osher Received: date / Accepted:

More information

Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation

Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation Zhouchen Lin Visual Computing Group Microsoft Research Asia Risheng Liu Zhixun Su School of Mathematical Sciences

More information

Recent Developments in Compressed Sensing

Recent Developments in Compressed Sensing Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline

More information

Stochastic dynamical modeling:

Stochastic dynamical modeling: Stochastic dynamical modeling: Structured matrix completion of partially available statistics Armin Zare www-bcf.usc.edu/ arminzar Joint work with: Yongxin Chen Mihailo R. Jovanovic Tryphon T. Georgiou

More information

On Convergence of the Alternating Projection Method for Matrix Completion and Sparse Recovery Problems

On Convergence of the Alternating Projection Method for Matrix Completion and Sparse Recovery Problems On Convergence of the Alternating Projection Method for Matrix Completion and Sparse Recovery Problems Ming-Jun Lai Department of Mathematics, University of Georgia, Athens, GA, USA 30602 mjlai@ugaedu

More information

GoDec: Randomized Low-rank & Sparse Matrix Decomposition in Noisy Case

GoDec: Randomized Low-rank & Sparse Matrix Decomposition in Noisy Case ianyi Zhou IANYI.ZHOU@SUDEN.US.EDU.AU Dacheng ao DACHENG.AO@US.EDU.AU Centre for Quantum Computation & Intelligent Systems, FEI, University of echnology, Sydney, NSW 2007, Australia Abstract Low-rank and

More information

ACCELERATED LINEARIZED BREGMAN METHOD. June 21, Introduction. In this paper, we are interested in the following optimization problem.

ACCELERATED LINEARIZED BREGMAN METHOD. June 21, Introduction. In this paper, we are interested in the following optimization problem. ACCELERATED LINEARIZED BREGMAN METHOD BO HUANG, SHIQIAN MA, AND DONALD GOLDFARB June 21, 2011 Abstract. In this paper, we propose and analyze an accelerated linearized Bregman (A) method for solving the

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Adrien Todeschini Inria Bordeaux JdS 2014, Rennes Aug. 2014 Joint work with François Caron (Univ. Oxford), Marie

More information

A Randomized Approach for Crowdsourcing in the Presence of Multiple Views

A Randomized Approach for Crowdsourcing in the Presence of Multiple Views A Randomized Approach for Crowdsourcing in the Presence of Multiple Views Presenter: Yao Zhou joint work with: Jingrui He - 1 - Roadmap Motivation Proposed framework: M2VW Experimental results Conclusion

More information

A Brief Overview of Practical Optimization Algorithms in the Context of Relaxation

A Brief Overview of Practical Optimization Algorithms in the Context of Relaxation A Brief Overview of Practical Optimization Algorithms in the Context of Relaxation Zhouchen Lin Peking University April 22, 2018 Too Many Opt. Problems! Too Many Opt. Algorithms! Zero-th order algorithms:

More information

MATRIX RECOVERY FROM QUANTIZED AND CORRUPTED MEASUREMENTS

MATRIX RECOVERY FROM QUANTIZED AND CORRUPTED MEASUREMENTS MATRIX RECOVERY FROM QUANTIZED AND CORRUPTED MEASUREMENTS Andrew S. Lan 1, Christoph Studer 2, and Richard G. Baraniuk 1 1 Rice University; e-mail: {sl29, richb}@rice.edu 2 Cornell University; e-mail:

More information

Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition

Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Gongguo Tang and Arye Nehorai Department of Electrical and Systems Engineering Washington University in St Louis

More information

S 1/2 Regularization Methods and Fixed Point Algorithms for Affine Rank Minimization Problems

S 1/2 Regularization Methods and Fixed Point Algorithms for Affine Rank Minimization Problems S 1/2 Regularization Methods and Fixed Point Algorithms for Affine Rank Minimization Problems Dingtao Peng Naihua Xiu and Jian Yu Abstract The affine rank minimization problem is to minimize the rank of

More information

Structured matrix factorizations. Example: Eigenfaces

Structured matrix factorizations. Example: Eigenfaces Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix

More information

ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering

ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering Filippo Pompili 1, Nicolas Gillis 2, P.-A. Absil 2,andFrançois Glineur 2,3 1- University of Perugia, Department

More information

Key words. Low-Rank Matrix Completion, Riemannian optimization, outliers, smoothing techniques, l 1 norm, non-smooth, fixed-rank manifold.

Key words. Low-Rank Matrix Completion, Riemannian optimization, outliers, smoothing techniques, l 1 norm, non-smooth, fixed-rank manifold. ROBUST LOW-RANK MATRIX COMPLETION BY RIEMANNIAN OPTIMIZATION LÉOPOLD CAMBIER AND P.-A. ABSIL Abstract. Low-rank matrix completion is the problem where one tries to recover a low-rank matrix from noisy

More information

Fast Nonnegative Matrix Factorization with Rank-one ADMM

Fast Nonnegative Matrix Factorization with Rank-one ADMM Fast Nonnegative Matrix Factorization with Rank-one Dongjin Song, David A. Meyer, Martin Renqiang Min, Department of ECE, UCSD, La Jolla, CA, 9093-0409 dosong@ucsd.edu Department of Mathematics, UCSD,

More information

A Simple Algorithm for Nuclear Norm Regularized Problems

A Simple Algorithm for Nuclear Norm Regularized Problems A Simple Algorithm for Nuclear Norm Regularized Problems ICML 00 Martin Jaggi, Marek Sulovský ETH Zurich Matrix Factorizations for recommender systems Y = Customer Movie UV T = u () The Netflix challenge:

More information

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng Robust PCA CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Robust PCA 1 / 52 Previously...

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Low rank matrix completion by alternating steepest descent methods

Low rank matrix completion by alternating steepest descent methods Low rank matrix completion by alternating steepest descent methods Jared Tanner a, Ke Wei b a Mathematical Institute, University of Oxford, Andrew Wiles Building, Radcliffe Observatory Quarter, Woodstock

More information

ECS289: Scalable Machine Learning

ECS289: Scalable Machine Learning ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Sept 29, 2016 Outline Convex vs Nonconvex Functions Coordinate Descent Gradient Descent Newton s method Stochastic Gradient Descent Numerical Optimization

More information

Symmetric Factorization for Nonconvex Optimization

Symmetric Factorization for Nonconvex Optimization Symmetric Factorization for Nonconvex Optimization Qinqing Zheng February 24, 2017 1 Overview A growing body of recent research is shedding new light on the role of nonconvex optimization for tackling

More information

c 2013 Society for Industrial and Applied Mathematics

c 2013 Society for Industrial and Applied Mathematics SIAM J. OPTIM. Vol. 23, No. 2, pp. 1214 1236 c 2013 Society for Industrial and Applied Mathematics LOW-RANK MATRIX COMPLETION BY RIEMANNIAN OPTIMIZATION BART VANDEREYCKEN Abstract. The matrix completion

More information

Subspace Projection Matrix Recovery from Incomplete Information

Subspace Projection Matrix Recovery from Incomplete Information Subspace Projection Matrix Recovery from Incomplete Information Xinyue Shen and Yuantao Gu Abstract Structural signal retrieval from highly incomplete information is a concerning problem in the field of

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 XVI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 A slightly changed ADMM for convex optimization with three separable operators Bingsheng He Department of

More information

CSC 576: Variants of Sparse Learning

CSC 576: Variants of Sparse Learning CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in

More information

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization Panos Parpas Department of Computing Imperial College London www.doc.ic.ac.uk/ pp500 p.parpas@imperial.ac.uk jointly with D.V.

More information

Parallel and Distributed Stochastic Learning -Towards Scalable Learning for Big Data Intelligence

Parallel and Distributed Stochastic Learning -Towards Scalable Learning for Big Data Intelligence Parallel and Distributed Stochastic Learning -Towards Scalable Learning for Big Data Intelligence oé LAMDA Group H ŒÆOŽÅ Æ EâX ^ #EâI[ : liwujun@nju.edu.cn Dec 10, 2016 Wu-Jun Li (http://cs.nju.edu.cn/lwj)

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Lecture 8: Optimization Cho-Jui Hsieh UC Davis May 9, 2017 Optimization Numerical Optimization Numerical Optimization: min X f (X ) Can be applied

More information

Matrix Completion from a Few Entries

Matrix Completion from a Few Entries Matrix Completion from a Few Entries Raghunandan H. Keshavan and Sewoong Oh EE Department Stanford University, Stanford, CA 9434 Andrea Montanari EE and Statistics Departments Stanford University, Stanford,

More information

Riemannian Optimization and its Application to Phase Retrieval Problem

Riemannian Optimization and its Application to Phase Retrieval Problem Riemannian Optimization and its Application to Phase Retrieval Problem 1/46 Introduction Motivations Optimization History Phase Retrieval Problem Summary Joint work with: Pierre-Antoine Absil, Professor

More information

Accelerated Block-Coordinate Relaxation for Regularized Optimization

Accelerated Block-Coordinate Relaxation for Regularized Optimization Accelerated Block-Coordinate Relaxation for Regularized Optimization Stephen J. Wright Computer Sciences University of Wisconsin, Madison October 09, 2012 Problem descriptions Consider where f is smooth

More information

Sparse Solutions of an Undetermined Linear System

Sparse Solutions of an Undetermined Linear System 1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research

More information

Sparse and Low-Rank Matrix Decomposition Via Alternating Direction Method

Sparse and Low-Rank Matrix Decomposition Via Alternating Direction Method Sparse and Low-Rank Matrix Decomposition Via Alternating Direction Method Xiaoming Yuan a, 1 and Junfeng Yang b, 2 a Department of Mathematics, Hong Kong Baptist University, Hong Kong, China b Department

More information

arxiv: v3 [math.oc] 19 Oct 2017

arxiv: v3 [math.oc] 19 Oct 2017 Gradient descent with nonconvex constraints: local concavity determines convergence Rina Foygel Barber and Wooseok Ha arxiv:703.07755v3 [math.oc] 9 Oct 207 0.7.7 Abstract Many problems in high-dimensional

More information

Optimization methods

Optimization methods Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,

More information

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence

More information

Supplementary Material of A Novel Sparsity Measure for Tensor Recovery

Supplementary Material of A Novel Sparsity Measure for Tensor Recovery Supplementary Material of A Novel Sparsity Measure for Tensor Recovery Qian Zhao 1,2 Deyu Meng 1,2 Xu Kong 3 Qi Xie 1,2 Wenfei Cao 1,2 Yao Wang 1,2 Zongben Xu 1,2 1 School of Mathematics and Statistics,

More information

Recovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies

Recovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies July 12, 212 Recovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies Morteza Mardani Dept. of ECE, University of Minnesota, Minneapolis, MN 55455 Acknowledgments:

More information

Linear dimensionality reduction for data analysis

Linear dimensionality reduction for data analysis Linear dimensionality reduction for data analysis Nicolas Gillis Joint work with Robert Luce, François Glineur, Stephen Vavasis, Robert Plemmons, Gabriella Casalino The setup Dimensionality reduction for

More information

Three Generalizations of Compressed Sensing

Three Generalizations of Compressed Sensing Thomas Blumensath School of Mathematics The University of Southampton June, 2010 home prev next page Compressed Sensing and beyond y = Φx + e x R N or x C N x K is K-sparse and x x K 2 is small y R M or

More information

Towards Efficient Large-scale Learning by Exploiting Sparsity

Towards Efficient Large-scale Learning by Exploiting Sparsity Towards Efficient Large-scale Learning by Exploiting Sparsity MINGKUI TAN School of Computer Engineering A thesis submitted to the Nanyang Technological University in partial fulfilment of the requirement

More information

Generalized Conditional Gradient and Its Applications

Generalized Conditional Gradient and Its Applications Generalized Conditional Gradient and Its Applications Yaoliang Yu University of Alberta UBC Kelowna, 04/18/13 Y-L. Yu (UofA) GCG and Its Apps. UBC Kelowna, 04/18/13 1 / 25 1 Introduction 2 Generalized

More information

Lecture 8: February 9

Lecture 8: February 9 0-725/36-725: Convex Optimiation Spring 205 Lecturer: Ryan Tibshirani Lecture 8: February 9 Scribes: Kartikeya Bhardwaj, Sangwon Hyun, Irina Caan 8 Proximal Gradient Descent In the previous lecture, we

More information

A Saddle Point Approach to Structured Low-rank Matrix Learning in Large-scale Applications

A Saddle Point Approach to Structured Low-rank Matrix Learning in Large-scale Applications A Saddle Point Approach to Structured Low-rank Matrix Learning in Large-scale Applications Pratik Jawanpuria Bamdev Mishra April 25, 2017 arxiv:1704.07352v1 [stat.ml] 24 Apr 2017 Abstract We propose a

More information

Low-Rank Factorization Models for Matrix Completion and Matrix Separation

Low-Rank Factorization Models for Matrix Completion and Matrix Separation for Matrix Completion and Matrix Separation Joint work with Wotao Yin, Yin Zhang and Shen Yuan IPAM, UCLA Oct. 5, 2010 Low rank minimization problems Matrix completion: find a low-rank matrix W R m n so

More information

Compressed Sensing and Robust Recovery of Low Rank Matrices

Compressed Sensing and Robust Recovery of Low Rank Matrices Compressed Sensing and Robust Recovery of Low Rank Matrices M. Fazel, E. Candès, B. Recht, P. Parrilo Electrical Engineering, University of Washington Applied and Computational Mathematics Dept., Caltech

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 9 Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 2 Separable convex optimization a special case is min f(x)

More information

Two Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery

Two Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery Two Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery Ming-Jun Lai, Song Li, Louis Y. Liu and Huimin Wang August 14, 2012 Abstract We shall provide a sufficient condition to

More information

Frank-Wolfe Method. Ryan Tibshirani Convex Optimization

Frank-Wolfe Method. Ryan Tibshirani Convex Optimization Frank-Wolfe Method Ryan Tibshirani Convex Optimization 10-725 Last time: ADMM For the problem min x,z f(x) + g(z) subject to Ax + Bz = c we form augmented Lagrangian (scaled form): L ρ (x, z, w) = f(x)

More information

Robust Matrix Completion via Joint Schatten p-norm and l p -Norm Minimization

Robust Matrix Completion via Joint Schatten p-norm and l p -Norm Minimization 0 IEEE th International Conference on Data Mining Robust Matrix Completion via Joint Schatten p-norm and l p -Norm Minimization Feiping Nie, Hua Wang,XiaoCai, Heng Huang and Chris Ding Department of Computer

More information

Uniqueness Conditions For Low-Rank Matrix Recovery

Uniqueness Conditions For Low-Rank Matrix Recovery Claremont Colleges Scholarship @ Claremont CMC Faculty Publications and Research CMC Faculty Scholarship 3-28-2011 Uniqueness Conditions For Low-Rank Matrix Recovery Yonina C. Eldar Israel Institute of

More information

High-dimensional Statistics

High-dimensional Statistics High-dimensional Statistics Pradeep Ravikumar UT Austin Outline 1. High Dimensional Data : Large p, small n 2. Sparsity 3. Group Sparsity 4. Low Rank 1 Curse of Dimensionality Statistical Learning: Given

More information

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725 Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:

More information

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun

More information

Recovering any low-rank matrix, provably

Recovering any low-rank matrix, provably Recovering any low-rank matrix, provably Rachel Ward University of Texas at Austin October, 2014 Joint work with Yudong Chen (U.C. Berkeley), Srinadh Bhojanapalli and Sujay Sanghavi (U.T. Austin) Matrix

More information

1-Bit Matrix Completion

1-Bit Matrix Completion 1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible

More information

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem Kangkang Deng, Zheng Peng Abstract: The main task of genetic regulatory networks is to construct a

More information

Joint Capped Norms Minimization for Robust Matrix Recovery

Joint Capped Norms Minimization for Robust Matrix Recovery Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI-7) Joint Capped Norms Minimization for Robust Matrix Recovery Feiping Nie, Zhouyuan Huo 2, Heng Huang 2

More information

Collaborative Filtering Matrix Completion Alternating Least Squares

Collaborative Filtering Matrix Completion Alternating Least Squares Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade May 19, 2016

More information

Sketchy Decisions: Convex Low-Rank Matrix Optimization with Optimal Storage

Sketchy Decisions: Convex Low-Rank Matrix Optimization with Optimal Storage Sketchy Decisions: Convex Low-Rank Matrix Optimization with Optimal Storage Madeleine Udell Operations Research and Information Engineering Cornell University Based on joint work with Alp Yurtsever (EPFL),

More information

Introduction to Compressed Sensing

Introduction to Compressed Sensing Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral

More information

Recovery of Simultaneously Structured Models using Convex Optimization

Recovery of Simultaneously Structured Models using Convex Optimization Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

Optimization. Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison

Optimization. Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison Optimization Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison optimization () cost constraints might be too much to cover in 3 hours optimization (for big

More information