Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds
|
|
- Anthony Bruce
- 5 years ago
- Views:
Transcription
1 Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds Tao Wu Institute for Mathematics and Scientific Computing Karl-Franzens-University of Graz joint work with Prof. Michael Hintermüller Riem-RPCP (1/19)
2 Low-rank paradigm. Low-rank matrices arise in one way or another: low-degree statistical processes e.g. collaborative filtering, latent semantic indexing. regularization on complex objects e.g. manifold learning, metric learning. approximation of compact operators e.g. proper orthogonal decomposition. Fig.: Collaborative filtering (courtesy of wikipedia.org). Riem-RPCP (2/19)
3 Robust principal component pursuit. Sparse component corresponds to pattern-irrelevant outliers. Robustifies classical principal component analysis. Carries important information in certain applications; e.g. moving objects in surveillance video. Robust principal component pursuit: data low-rank sparse noise Z = A + B + N Introduced in [Candés, Li, Ma, and Wright, 11], [Chandrasekaran, Sanghavi, Parrilo, and Willsky, 11]. tao.wu@uni-graz.at Riem-RPCP (3/19)
4 Convex-relaxation approach. A popular (convex) variational model: min A nuclear + λ B l 1 s.t. A + B Z ε. Considered in [Candés, Li, Ma, and Wright, 11], [Chandrasekaran, Sanghavi, Parrilo, and Willsky, 11],... rank(a) relaxed by nuclear-norm; B 0 relaxed by l 1 -norm. Numerical solvers: proximal gradient method, augmented Lagrangian method,... Efficiency is constrained by SVD in full dimension at each iteration. tao.wu@uni-graz.at Riem-RPCP (4/19)
5 Manifold constrained least-squares model. Our variational model: min 1 A + B Z 2 2 s.t. A M(r) := {A R m n : rank(a) r}, B N (s) := {B R m n : B 0 s}. Our goal is to develop an algorithm such that: globally converges to a stationary point (often a local minimizer). provides exact decomposition with high probability for noiseless data. outperforms solvers based on convex-relaxation approach, especially in large scales. tao.wu@uni-graz.at Riem-RPCP (5/19)
6 Existence of solution and optimality condition. A little quadratic regularization (0 < µ 1) is included for the (theoretical) sake of existence of a solution; i.e. min f(a, B) := 1 2 A + B Z 2 + µ 2 A 2, s.t. (A, B) M(r) N (s). In numerics, choosing µ = 0 seems fine. Stationarity condition as variational inequalities: {, (1 + µ)a + B Z 0, for any T M(r) (A ),, A + B Z 0, for any T N (s) (B ). T M(r) (A ) and T N (s) (B ) refer to tangent cones. tao.wu@uni-graz.at Riem-RPCP (6/19)
7 Constraints of Riemannian manifolds. M(r) is Riemannian manifold around A if rank(a ) = r; N (s) is Riemannian manifold around B if B 0 = s. Optimality condition reduces to: { PTM(r)(A )((1 + µ)a + B Z) = 0, P TN (s) (B )(A + B Z) = 0. P TM(r) (A ) and P TN (s) (B ) are orthogonal projections onto subspaces. Tangent space formulae: T M(r) (A ) = {UMV + U pv + UV p : A = UΣV as compact SVD, M R r r, U p R m r, U p U = 0, V p R n r, V p V = 0}, T N (s) (B ) = { R m n : supp( ) supp(b )}. tao.wu@uni-graz.at Riem-RPCP (7/19)
8 A conceptual alternating minimization scheme. Initialize A 0 M(r), B 0 N (s). Set k := 0 and iterate: 1. A k+1 arg min A M(r) 1 2 A + Bk Z 2 + µ 2 A B k+1 arg min B N (s) 1 2 Ak+1 + B Z 2. Theorem (sufficient descrease + stationarity convergence) Let {(A k, B k )} be generated as above. Suppose that there exists δ > 0, ε k a 0, and ε k b 0 such that for all k: f(a k+1, B k ) f(a k, B k ) δ A k+1 A k 2, f(a k+1, B k+1 ) f(a k+1, B k ) δ B k+1 B k 2,, (1 + µ)a k+1 + B k Z ε k a, for any T M(r) (A k+1 ),, A k+1 + B k+1 Z ε k b, for any T N (s) (B k+1 ). Then any non-degenerate limit point (A, B ), i.e. rank(a ) = r and B 0 = s, satisfies the first-order optimality condition. tao.wu@uni-graz.at Riem-RPCP (8/19)
9 Sparse matrix subproblem. The global solution P N (s) (Z A k+1 ) (as metric projection) can be efficiently calculated from sorting. The global solution may not necessarily fulfill the sufficient descrease condition. Whenever necessary, safeguard by a local solution: { Bij k+1 (Z A k+1 ) = ij, if Bij k 0, 0, otherwise. Given non-degeneracy of B k+1, i.e. B k+1 0 = s, the exact stationarity holds. tao.wu@uni-graz.at Riem-RPCP (9/19)
10 Low-rank matrix subproblem: a Riemannian perspective. Global solution P M(r) ( 1 1+µ (Z Bk )) as metric projection: available due to Eckart-Young theorem; i.e. 1 n 1 + µ (Z Bk ) = σ ju jvj 1 r P M(r) ( 1 + µ (Z Bk )) = σ ju jvj. j=1 j=1 but requires SVD in full dimension expensive for large-scale problems (e.g. m, n 2000). Alternatively resolved by a single Riemannian optimization step on matrix manifold. Riemannian optimization applied to low-rank matrix/tensor problems; see [Simonsson and Eldén, 10], [Savas and Lim, 10], [Vandereycken, 13],... Our goal: The subproblem solver should activate the convergence criteria, i.e. sufficient descrease + stationarity. tao.wu@uni-graz.at Riem-RPCP (10/19)
11 Optimization on Manifolds in one picture Riemannian optimization: an overview. R M f References: [Smith, 93], [Edelman, Arias, and Smith, 98], [Absil, Mahony, and Sepulchre, 08],... Why Riemannian optimization? Local homeomorphism is computationally infeasible/expensive. 3 Intrinsically low dimensionality of the underlying manifold. Further dimension reduction via quotient manifold. Typical Riemannian manifolds in applications: finite-dimensional (matrix manifold): Stiefel manifold, Grassmann manifold, fixed-rank matrix manifold,... infinite-dimensional: shape/curve spaces,... tao.wu@uni-graz.at Riem-RPCP (11/19)
12 Riemannian optimization: a conceptual algorithm. T M(r)(A k ) TxM x A k ξ k retract Rx(ξ) M(r)(A k, k ) M(r) M At the current iterate: 1. Build a quadratic model in the tangent space using Riemannian gradient and Riemannian Hessian. 2. Based on the quadratic model, build a tangential search path. 3. Perform backtracking path search via retraction to determine the step size. 4. Generate the next iterate. tao.wu@uni-graz.at Riem-RPCP (12/19)
13 Riemannian gradient and Hessian. M(r) := {A : rank(a) = r}; f k A : A M(r) f(a, B k ). Riemannian gradient, gradfa k (A) T M(r) (A), is defined s.t. gradfa k(a), = Df A k (A)[ ], T M(r) (A). gradf k A(A) = P T M(r) (A)( f k A(A)). Riemannian Hessian, HessfA k (A) : T M(r) (A) T M(r) (A), is defined s.t. HessfA k(a)[ ] = gradfa k (A), T M (A). Hessf k A(A)[ ] = (I UU ) f k A(A)(I V V ) UΣ 1 V + UΣ 1 V (I UU ) f k A(A)(I V V ) + (1 + µ). See, e.g., [Vandereycken, 12]. tao.wu@uni-graz.at Riem-RPCP (13/19)
14 Dogleg search path and projective retraction. Trust region T M(r)(A k ) TxM x A k Optimal trajectory (σ) p( ) ξ k retract Rx(ξ) M(r)(A k, k ) C pu (unconstrained min along g) dogleg path g N p B ( full step ) M(r) M Dogleg path k (τ k ) as approximation of optimal trajectory of tangential trust-region subproblem (left figure): min f k A(A k ) + g k, + 1 2, Hk [ ] s.t. T M(r) (A k ), σ. Metric projection as retraction (right figure): retract M(r) (A k, k (τ k )) = P M(r) (A k + k (τ k )). Computationally efficient: reduced SVD on 2r-by-2r matrix! tao.wu@uni-graz.at Riem-RPCP (14/19)
15 Low-rank matrix subproblem: projected dogleg step. Given A k M(r), B k N (s): 1. Compute g k, H k, and build the dogleg search path k (τ k ) in T M(r) (A k ). 2. Whenever non-positive definiteness of H k is detected, replace the dogleg search path by the line search path along steepest descent direction, i.e. (τ k ) = τ k g k. 3. Perform backtracking path/line search; i.e. find the largest step size τ k {2, 3/2, 1, 1/2, 1/4, 1/8,...} s.t. the sufficient descrease condition is satisfied: f k A(A k ) f k A(P M(r) (A k + k (τ k ))) δ A k P M(r) (A k + k (τ k )) Return A k+1 = f k A (P M(r) (Ak + k (τ k ))). tao.wu@uni-graz.at Riem-RPCP (15/19)
16 Low-rank matrix subproblem: convergence theory. Backtracking path search: The sufficient descrease condition can always be fulfilled after finitely many trails on τ k. Any accumulation point of {A k } is stationary. Further assume Hessf(A, B ) 0 at a non-degenerate µ=0 accumulation point (A, B ). Then Tangent-space transversality holds, i.e. T M(r) (A ) T N (s) (B ) = {0}. Contractivity of PT M(r) (A ) P TN (s) (B ): κ [0, 1) s.t. (P T M(r) (A ) P TN (s) (B ))( ) κ. q-linear convergence of {A k } towards stationarity: lim sup k A k+1 A A k A κ. tao.wu@uni-graz.at Riem-RPCP (16/19)
17 Numerical implementation. Trimming Adaptive tuning of rank r k+1 and cardinality s k+1 based on the current iterate (A k, B k ). k-means clustering on (nonzero) singular values of A k in logarithmic scale. hard thresholding on entries of B k. q-linear convergence confirmed numerically: A k A * / A * B k B * / B * iteration (a) Convergence of {A k } iteration (b) Convergence of {B k }. tao.wu@uni-graz.at Riem-RPCP (17/19)
18 Comparison with augmented Lagrangian method (m = n = 2000) AMS AMS # fsvd ALM psvd ALM AMS AMS # fsvd ALM psvd ALM relative error on A k relative error on B k CPU time (a) Relative error of {A k } CPU time (b) Relative error of {B k } AMS AMS # fsvd ALM psvd ALM AMS AMS # fsvd ALM psvd ALM rank(a k )/n B k 0 /n CPU time CPU time (c) Phase transition of {A k }. (d) Phase transition of {B k }. tao.wu@uni-graz.at Riem-RPCP (18/19)
19 Application to surveillance video. Problem settings: A sequence of 200 frames taken from a surveillance video at an airport. Each frame is a gray image of resolution Stack 3D-array into a matrix. Results: CPU time: AMS 39.4s; ALM 124.4s. Visual comparison. tao.wu@uni-graz.at Riem-RPCP (19/19)
Robust Principal Component Pursuit via Inexact Alternating Minimization on Matrix Manifolds
Robust Principal Component Pursuit via Inexact Alternating Minimization on Matrix Manifolds Michael Hintermüller Tao Wu Abstract Robust principal component pursuit (RPCP) refers to a decomposition of a
More informationRobust Principal Component Analysis
ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M
More informationRanking from Crowdsourced Pairwise Comparisons via Matrix Manifold Optimization
Ranking from Crowdsourced Pairwise Comparisons via Matrix Manifold Optimization Jialin Dong ShanghaiTech University 1 Outline Introduction FourVignettes: System Model and Problem Formulation Problem Analysis
More informationStructured matrix factorizations. Example: Eigenfaces
Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix
More informationSTUDY of genetic variations is a critical task to disease
Haplotype Assembly Using Manifold Optimization and Error Correction Mechanism Mohamad Mahdi Mohades, Sina Majidian, and Mohammad Hossein Kahaei arxiv:9.36v [math.oc] Jan 29 Abstract Recent matrix completion
More informationLow-Rank Factorization Models for Matrix Completion and Matrix Separation
for Matrix Completion and Matrix Separation Joint work with Wotao Yin, Yin Zhang and Shen Yuan IPAM, UCLA Oct. 5, 2010 Low rank minimization problems Matrix completion: find a low-rank matrix W R m n so
More informationGoDec: Randomized Low-rank & Sparse Matrix Decomposition in Noisy Case
ianyi Zhou IANYI.ZHOU@SUDEN.US.EDU.AU Dacheng ao DACHENG.AO@US.EDU.AU Centre for Quantum Computation & Intelligent Systems, FEI, University of echnology, Sydney, NSW 2007, Australia Abstract Low-rank and
More informationEE 381V: Large Scale Optimization Fall Lecture 24 April 11
EE 381V: Large Scale Optimization Fall 2012 Lecture 24 April 11 Lecturer: Caramanis & Sanghavi Scribe: Tao Huang 24.1 Review In past classes, we studied the problem of sparsity. Sparsity problem is that
More informationDense Error Correction for Low-Rank Matrices via Principal Component Pursuit
Dense Error Correction for Low-Rank Matrices via Principal Component Pursuit Arvind Ganesh, John Wright, Xiaodong Li, Emmanuel J. Candès, and Yi Ma, Microsoft Research Asia, Beijing, P.R.C Dept. of Electrical
More informationLatent Semantic Analysis. Hongning Wang
Latent Semantic Analysis Hongning Wang CS@UVa VS model in practice Document and query are represented by term vectors Terms are not necessarily orthogonal to each other Synonymy: car v.s. automobile Polysemy:
More informationLecture 9: September 28
0-725/36-725: Convex Optimization Fall 206 Lecturer: Ryan Tibshirani Lecture 9: September 28 Scribes: Yiming Wu, Ye Yuan, Zhihao Li Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These
More informationLow Rank Matrix Completion Formulation and Algorithm
1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9
More informationInvariant Subspace Computation - A Geometric Approach
Invariant Subspace Computation - A Geometric Approach Pierre-Antoine Absil PhD Defense Université de Liège Thursday, 13 February 2003 1 Acknowledgements Advisor: Prof. R. Sepulchre (Université de Liège)
More informationStochastic dynamical modeling:
Stochastic dynamical modeling: Structured matrix completion of partially available statistics Armin Zare www-bcf.usc.edu/ arminzar Joint work with: Yongxin Chen Mihailo R. Jovanovic Tryphon T. Georgiou
More informationThe convex algebraic geometry of rank minimization
The convex algebraic geometry of rank minimization Pablo A. Parrilo Laboratory for Information and Decision Systems Massachusetts Institute of Technology International Symposium on Mathematical Programming
More informationRobust PCA by Manifold Optimization
Journal of Machine Learning Research 19 (2018) 1-39 Submitted 8/17; Revised 10/18; Published 11/18 Robust PCA by Manifold Optimization Teng Zhang Department of Mathematics University of Central Florida
More informationSection 3.9. Matrix Norm
3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix
More informationRiemannian Pursuit for Big Matrix Recovery
Riemannian Pursuit for Big Matrix Recovery Mingkui Tan 1, Ivor W. Tsang 2, Li Wang 3, Bart Vandereycken 4, Sinno Jialin Pan 5 1 School of Computer Science, University of Adelaide, Australia 2 Center for
More informationc 2013 Society for Industrial and Applied Mathematics
SIAM J. OPTIM. Vol. 23, No. 2, pp. 1214 1236 c 2013 Society for Industrial and Applied Mathematics LOW-RANK MATRIX COMPLETION BY RIEMANNIAN OPTIMIZATION BART VANDEREYCKEN Abstract. The matrix completion
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationRobust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition
Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Gongguo Tang and Arye Nehorai Department of Electrical and Systems Engineering Washington University in St Louis
More informationSparse and low-rank decomposition for big data systems via smoothed Riemannian optimization
Sparse and low-rank decomposition for big data systems via smoothed Riemannian optimization Yuanming Shi ShanghaiTech University, Shanghai, China shiym@shanghaitech.edu.cn Bamdev Mishra Amazon Development
More informationOn the convergence of higher-order orthogonality iteration and its extension
On the convergence of higher-order orthogonality iteration and its extension Yangyang Xu IMA, University of Minnesota SIAM Conference LA15, Atlanta October 27, 2015 Best low-multilinear-rank approximation
More informationSuppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf.
Maria Cameron 1. Trust Region Methods At every iteration the trust region methods generate a model m k (p), choose a trust region, and solve the constraint optimization problem of finding the minimum of
More information1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?)
Math 35 Exam Review SOLUTIONS Overview In this third of the course we focused on linear learning algorithms to model data. summarize: To. Background: The SVD and the best basis (questions selected from
More informationMatrix Completion from Noisy Entries
Matrix Completion from Noisy Entries Raghunandan H. Keshavan, Andrea Montanari, and Sewoong Oh Abstract Given a matrix M of low-rank, we consider the problem of reconstructing it from noisy observations
More informationLecture 9: Numerical Linear Algebra Primer (February 11st)
10-725/36-725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template
More information1 Computing with constraints
Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)
More informationOptimization. Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison
Optimization Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison optimization () cost constraints might be too much to cover in 3 hours optimization (for big
More informationTrust Region Methods. Lecturer: Pradeep Ravikumar Co-instructor: Aarti Singh. Convex Optimization /36-725
Trust Region Methods Lecturer: Pradeep Ravikumar Co-instructor: Aarti Singh Convex Optimization 10-725/36-725 Trust Region Methods min p m k (p) f(x k + p) s.t. p 2 R k Iteratively solve approximations
More informationLinear dimensionality reduction for data analysis
Linear dimensionality reduction for data analysis Nicolas Gillis Joint work with Robert Luce, François Glineur, Stephen Vavasis, Robert Plemmons, Gabriella Casalino The setup Dimensionality reduction for
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationConvex Optimization Algorithms for Machine Learning in 10 Slides
Convex Optimization Algorithms for Machine Learning in 10 Slides Presenter: Jul. 15. 2015 Outline 1 Quadratic Problem Linear System 2 Smooth Problem Newton-CG 3 Composite Problem Proximal-Newton-CD 4 Non-smooth,
More informationAnalysis of Robust PCA via Local Incoherence
Analysis of Robust PCA via Local Incoherence Huishuai Zhang Department of EECS Syracuse University Syracuse, NY 3244 hzhan23@syr.edu Yi Zhou Department of EECS Syracuse University Syracuse, NY 3244 yzhou35@syr.edu
More informationNumerical Optimization of Partial Differential Equations
Numerical Optimization of Partial Differential Equations Part I: basic optimization concepts in R n Bartosz Protas Department of Mathematics & Statistics McMaster University, Hamilton, Ontario, Canada
More informationRecovering any low-rank matrix, provably
Recovering any low-rank matrix, provably Rachel Ward University of Texas at Austin October, 2014 Joint work with Yudong Chen (U.C. Berkeley), Srinadh Bhojanapalli and Sujay Sanghavi (U.T. Austin) Matrix
More information5 Quasi-Newton Methods
Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min
More informationLatent Semantic Analysis. Hongning Wang
Latent Semantic Analysis Hongning Wang CS@UVa Recap: vector space model Represent both doc and query by concept vectors Each concept defines one dimension K concepts define a high-dimensional space Element
More informationLow-rank matrix completion by Riemannian optimization
Low-rank matrix completion by Riemannian optimization Bart Vandereycken August 24, 2011 Abstract The matrix completion problem consists of finding or approximating a low-rank matrix based on a few samples
More informationRobust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng
Robust PCA CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Robust PCA 1 / 52 Previously...
More informationhttps://goo.gl/kfxweg KYOTO UNIVERSITY Statistical Machine Learning Theory Sparsity Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT OF INTELLIGENCE SCIENCE AND TECHNOLOGY 1 KYOTO UNIVERSITY Topics:
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationLecture 18 Nov 3rd, 2015
CS 229r: Algorithms for Big Data Fall 2015 Prof. Jelani Nelson Lecture 18 Nov 3rd, 2015 Scribe: Jefferson Lee 1 Overview Low-rank approximation, Compression Sensing 2 Last Time We looked at three different
More informationAn extrinsic look at the Riemannian Hessian
http://sites.uclouvain.be/absil/2013.01 Tech. report UCL-INMA-2013.01-v2 An extrinsic look at the Riemannian Hessian P.-A. Absil 1, Robert Mahony 2, and Jochen Trumpf 2 1 Department of Mathematical Engineering,
More informationRecovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm
Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm J. K. Pant, W.-S. Lu, and A. Antoniou University of Victoria August 25, 2011 Compressive Sensing 1 University
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationSparse and Low-Rank Matrix Decompositions
Forty-Seventh Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 30 - October 2, 2009 Sparse and Low-Rank Matrix Decompositions Venkat Chandrasekaran, Sujay Sanghavi, Pablo A. Parrilo,
More informationSupplementary Materials for Riemannian Pursuit for Big Matrix Recovery
Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery Mingkui Tan, School of omputer Science, The University of Adelaide, Australia Ivor W. Tsang, IS, University of Technology Sydney,
More informationRiemannian Optimization and its Application to Phase Retrieval Problem
Riemannian Optimization and its Application to Phase Retrieval Problem 1/46 Introduction Motivations Optimization History Phase Retrieval Problem Summary Joint work with: Pierre-Antoine Absil, Professor
More informationSparse Optimization Lecture: Basic Sparse Optimization Models
Sparse Optimization Lecture: Basic Sparse Optimization Models Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know basic l 1, l 2,1, and nuclear-norm
More informationFantope Regularization in Metric Learning
Fantope Regularization in Metric Learning CVPR 2014 Marc T. Law (LIP6, UPMC), Nicolas Thome (LIP6 - UPMC Sorbonne Universités), Matthieu Cord (LIP6 - UPMC Sorbonne Universités), Paris, France Introduction
More informationNonlinear Optimization for Optimal Control
Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]
More informationbe a Householder matrix. Then prove the followings H = I 2 uut Hu = (I 2 uu u T u )u = u 2 uut u
MATH 434/534 Theoretical Assignment 7 Solution Chapter 7 (71) Let H = I 2uuT Hu = u (ii) Hv = v if = 0 be a Householder matrix Then prove the followings H = I 2 uut Hu = (I 2 uu )u = u 2 uut u = u 2u =
More informationE5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization
E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained
More informationMatrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University
Matrix completion: Fundamental limits and efficient algorithms Sewoong Oh Stanford University 1 / 35 Low-rank matrix completion Low-rank Data Matrix Sparse Sampled Matrix Complete the matrix from small
More informationIntroduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin
1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)
More informationGrassmann Averages for Scalable Robust PCA Supplementary Material
Grassmann Averages for Scalable Robust PCA Supplementary Material Søren Hauberg DTU Compute Lyngby, Denmark sohau@dtu.dk Aasa Feragen DIKU and MPIs Tübingen Denmark and Germany aasa@diku.dk Michael J.
More informationDirect Robust Matrix Factorization for Anomaly Detection
Direct Robust Matrix Factorization for Anomaly Detection Liang Xiong Machine Learning Department Carnegie Mellon University lxiong@cs.cmu.edu Xi Chen Machine Learning Department Carnegie Mellon University
More informationOptimization and Root Finding. Kurt Hornik
Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding
More informationConvex Optimization and Modeling
Convex Optimization and Modeling Introduction and a quick repetition of analysis/linear algebra First lecture, 12.04.2010 Jun.-Prof. Matthias Hein Organization of the lecture Advanced course, 2+2 hours,
More informationEE 381V: Large Scale Learning Spring Lecture 16 March 7
EE 381V: Large Scale Learning Spring 2013 Lecture 16 March 7 Lecturer: Caramanis & Sanghavi Scribe: Tianyang Bai 16.1 Topics Covered In this lecture, we introduced one method of matrix completion via SVD-based
More informationAlgorithms for constrained local optimization
Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Lecture 5: Numerical Linear Algebra Cho-Jui Hsieh UC Davis April 20, 2017 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationLift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints
Lift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints Nicolas Boumal Université catholique de Louvain (Belgium) IDeAS seminar, May 13 th, 2014, Princeton The Riemannian
More informationProximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725
Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:
More informationLow rank matrix completion by alternating steepest descent methods
SIAM J. IMAGING SCIENCES Vol. xx, pp. x c xxxx Society for Industrial and Applied Mathematics x x Low rank matrix completion by alternating steepest descent methods Jared Tanner and Ke Wei Abstract. Matrix
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57
More informationFrom Compressed Sensing to Matrix Completion and Beyond. Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison
From Compressed Sensing to Matrix Completion and Beyond Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison Netflix Prize One million big ones! Given 100 million ratings on a
More informationRobust Asymmetric Nonnegative Matrix Factorization
Robust Asymmetric Nonnegative Matrix Factorization Hyenkyun Woo Korea Institue for Advanced Study Math @UCLA Feb. 11. 2015 Joint work with Haesun Park Georgia Institute of Technology, H. Woo (KIAS) RANMF
More informationLow-rank Matrix Completion with Noisy Observations: a Quantitative Comparison
Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison Raghunandan H. Keshavan, Andrea Montanari and Sewoong Oh Electrical Engineering and Statistics Department Stanford University,
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationPrincipal Component Analysis
Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used
More informationEE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015
EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,
More informationAlgebraic models for higher-order correlations
Algebraic models for higher-order correlations Lek-Heng Lim and Jason Morton U.C. Berkeley and Stanford Univ. December 15, 2008 L.-H. Lim & J. Morton (MSRI Workshop) Algebraic models for higher-order correlations
More informationOptimisation on Manifolds
Optimisation on Manifolds K. Hüper MPI Tübingen & Univ. Würzburg K. Hüper (MPI Tübingen & Univ. Würzburg) Applications in Computer Vision Grenoble 18/9/08 1 / 29 Contents 2 Examples Essential matrix estimation
More informationONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering
ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering Filippo Pompili 1, Nicolas Gillis 2, P.-A. Absil 2,andFrançois Glineur 2,3 1- University of Perugia, Department
More informationSequential Unconstrained Minimization: A Survey
Sequential Unconstrained Minimization: A Survey Charles L. Byrne February 21, 2013 Abstract The problem is to minimize a function f : X (, ], over a non-empty subset C of X, where X is an arbitrary set.
More informationDISSERTATION MEAN VARIANTS ON MATRIX MANIFOLDS. Submitted by. Justin D Marks. Department of Mathematics. In partial fulfillment of the requirements
DISSERTATION MEAN VARIANTS ON MATRIX MANIFOLDS Submitted by Justin D Marks Department of Mathematics In partial fulfillment of the requirements For the Degree of Doctor of Philosophy Colorado State University
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationLecture 5 : Projections
Lecture 5 : Projections EE227C. Lecturer: Professor Martin Wainwright. Scribe: Alvin Wan Up until now, we have seen convergence rates of unconstrained gradient descent. Now, we consider a constrained minimization
More informationCSC 576: Variants of Sparse Learning
CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in
More informationNote on the convex hull of the Stiefel manifold
Note on the convex hull of the Stiefel manifold Kyle A. Gallivan P.-A. Absil July 9, 00 Abstract In this note, we characterize the convex hull of the Stiefel manifold and we find its strong convexity parameter.
More informationOutline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St
Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic
More informationGaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula
Gaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula Larry Goldstein, University of Southern California Nourdin GIoVAnNi Peccati Luxembourg University University British
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationLecture 1: September 25
0-725: Optimization Fall 202 Lecture : September 25 Lecturer: Geoff Gordon/Ryan Tibshirani Scribes: Subhodeep Moitra Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have
More informationRobustness of Principal Components
PCA for Clustering An objective of principal components analysis is to identify linear combinations of the original variables that are useful in accounting for the variation in those original variables.
More informationA Simple Algorithm for Nuclear Norm Regularized Problems
A Simple Algorithm for Nuclear Norm Regularized Problems ICML 00 Martin Jaggi, Marek Sulovský ETH Zurich Matrix Factorizations for recommender systems Y = Customer Movie UV T = u () The Netflix challenge:
More informationIV. Matrix Approximation using Least-Squares
IV. Matrix Approximation using Least-Squares The SVD and Matrix Approximation We begin with the following fundamental question. Let A be an M N matrix with rank R. What is the closest matrix to A that
More informationCoordinate Update Algorithm Short Course Proximal Operators and Algorithms
Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow
More informationGI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil
GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection
More informationMachine Learning - MT & 14. PCA and MDS
Machine Learning - MT 2016 13 & 14. PCA and MDS Varun Kanade University of Oxford November 21 & 23, 2016 Announcements Sheet 4 due this Friday by noon Practical 3 this week (continue next week if necessary)
More informationSingular Value Decomposition (SVD)
School of Computing National University of Singapore CS CS524 Theoretical Foundations of Multimedia More Linear Algebra Singular Value Decomposition (SVD) The highpoint of linear algebra Gilbert Strang
More informationLecture Notes 10: Matrix Factorization
Optimization-based data analysis Fall 207 Lecture Notes 0: Matrix Factorization Low-rank models. Rank- model Consider the problem of modeling a quantity y[i, j] that depends on two indices i and j. To
More informationExact Recoverability of Robust PCA via Outlier Pursuit with Tight Recovery Bounds
Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence Exact Recoverability of Robust PCA via Outlier Pursuit with Tight Recovery Bounds Hongyang Zhang, Zhouchen Lin, Chao Zhang, Edward
More informationc 2010 Society for Industrial and Applied Mathematics
SIAM J. OPTIM. Vol. 20, No. 5, pp. 2327 2351 c 2010 Society for Industrial and Applied Mathematics LOW-RANK OPTIMIZATION ON THE CONE OF POSITIVE SEMIDEFINITE MATRICES M. JOURNÉE, F. BACH, P.-A. ABSIL,
More informationLINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION
LINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION JUNFENG YANG AND XIAOMING YUAN Abstract. The nuclear norm is widely used to induce low-rank solutions for
More informationDivide-and-Conquer Matrix Factorization
Divide-and-Conquer Matrix Factorization Lester Mackey Collaborators: Ameet Talwalkar Michael I. Jordan Stanford University UCLA UC Berkeley December 14, 2015 Mackey (Stanford) Divide-and-Conquer Matrix
More informationShort Course Robust Optimization and Machine Learning. Lecture 4: Optimization in Unsupervised Learning
Short Course Robust Optimization and Machine Machine Lecture 4: Optimization in Unsupervised Laurent El Ghaoui EECS and IEOR Departments UC Berkeley Spring seminar TRANSP-OR, Zinal, Jan. 16-19, 2012 s
More informationarxiv: v1 [cs.cv] 1 Jun 2014
l 1 -regularized Outlier Isolation and Regression arxiv:1406.0156v1 [cs.cv] 1 Jun 2014 Han Sheng Department of Electrical and Electronic Engineering, The University of Hong Kong, HKU Hong Kong, China sheng4151@gmail.com
More information