Low-Rank Factorization Models for Matrix Completion and Matrix Separation
|
|
- Nelson Summers
- 6 years ago
- Views:
Transcription
1 for Matrix Completion and Matrix Separation Joint work with Wotao Yin, Yin Zhang and Shen Yuan IPAM, UCLA Oct. 5, 2010
2 Low rank minimization problems Matrix completion: find a low-rank matrix W R m n so that W ij = M ij for some given M ij with all (i, j) in an index set Ω: min W R m n rank(w ), s.t. W ij = M ij, (i, j) Ω Matrix separation: find a low-rank Z R m n and a sparse matrix S R m n so that Z + S = D for a given D.
3 Convex approximation Nuclear norm W : the summation of singular values of W Matrix completion: Nuclear-norm relaxation: min W, s.t. W ij = M ij, (i, j) Ω, W R m n Nuclear-norm regularized linear least square: Matrix separation: min µ W + 1 W R m n 2 P Ω(W M) 2 F, min Z + µ S 1 s.t. Z + S = D, Z,S R m n Singular value decomposition can be very expensive!
4 Low-rank factorization model for matrix completion Finding a low-rank matrix W so that P Ω (W M) 2 F or the distance between W and {Z R m n, Z ij = M ij, (i, j) Ω} is minimized. Any matrix W R m n with rank(w ) K can be expressed as W = XY where X R m K and Y R K n. New model min X,Y,Z 1 2 XY Z 2 F s.t. Z ij = M ij, (i, j) Ω Advantage: SVD is no longer needed! Related work: the solver OptSpace based on optimization on manifold
5 Low-rank factorization model for matrix separation Consider the model min S 1 s.t. Z + S = D, rank(l) K Z,S Low-rank factorization: Z = UV min Z D 1 s.t. UV Z = 0 U,V,Z Only the entries D ij, (i, j) Ω, are given. P Ω (D) is the projection of D onto Ω. New model min U,V,Z P Ω (Z D) 1 s.t. UV Z = 0 Advantage: SVD is no longer needed!
6 Nonlinear Gauss-Seideal scheme Consider: min X,Y,Z Alternating minimization: 1 2 XY Z 2 F s.t. Z ij = M ij, (i, j) Ω X + = ZY (YY ) 1 = argmin X R m K 2 XY Z 2 F,
7 Nonlinear Gauss-Seideal scheme Consider: min X,Y,Z 1 2 XY Z 2 F s.t. Z ij = M ij, (i, j) Ω First variant of alternating minimization: X + ZY ZY (YY ), Y + (X + ) Z (X + X + ) (X + Z ), Z + X + Y + + P Ω (M X + Y + ). Let P A be the orthogonal projection onto the range space R(A) X + Y + = ( X + (X + X + ) X + ) Z = PX+ Z One can verify that R(X + ) = R(ZY ). X + Y + = P ZY Z = ZY (YZ ZY ) (YZ )Z. idea: modify X + or Y + to obtain the same product X + Y +
8 Nonlinear Gauss-Seideal scheme Consider: min X,Y,Z 1 2 XY Z 2 F s.t. Z ij = M ij, (i, j) Ω Second variant of alternating minimization: X + ZY, Y + (X + ) Z (X + X + ) (X + Z ), Z + X + Y + + P Ω (M X + Y + ).
9 Nonlinear Gauss-Seideal scheme Consider: min X,Y,Z 1 2 XY Z 2 F s.t. Z ij = M ij, (i, j) Ω Second variant of alternating minimization: X + ZY, Y + (X + ) Z (X + X + ) (X + Z ), Z + X + Y + + P Ω (M X + Y + ). Third variant of alternating minimization: V = orth(zy ) X + V, Y + V Z, Z + X + Y + + P Ω (M X + Y + ).
10 Nonlinear SOR The nonlinear GS scheme can be slow Linear SOR: applying extrapolation to the GS method to achieve faster convergence The first implementation: X + ZY (YY ), X + (ω) ωx + + (1 ω)x, Y + (X + (ω) X + (ω)) (X + (ω) Z ), Y + (ω) ωy + + (1 ω)y, Z + (ω) X + (ω)y + (ω) + P Ω (M X + (ω)y + (ω)),
11 Nonlinear SOR Let S = P Ω (M XY ). Then Z = XY + S Let Z ω XY + ωs = ωz + (1 ω)xy Assume Y has full row rank, then Z ω Y (YY ) = ωzy (YY ) + (1 ω)xyy (YY ) = ωx + + (1 ω)x, Second implementation of our nonlinear SOR: X + (ω) Z ω Y or Z ω Y (YY ), Y + (ω) (X + (ω) X + (ω)) (X + (ω) Z ω ), P Ω c(z + (ω)) P Ω c(x + (ω)y + (ω)), P Ω (Z + (ω)) P Ω (M).
12 Reduction of the residual S 2 F S +(ω) 2 F Assume that rank(z ω ) = rank(z ), ω [1, ω 1 ] for some ω 1 1. Then there exists some ω 2 1 such that S 2 F S +(ω) 2 F = ρ 12(ω) + ρ 3 (ω) + ρ 4 (ω) > 0, ω [1, ω 2 ]. ρ 12 (ω) SP 2 F + Q(ω)S(I P) 2 F 0 ρ 3 (ω) P Ω c(sp + Q(ω)S(I P)) 2 F 0 ρ 4 (ω) 1 ω 2 S + (ω) + (ω 1)S 2 F S +(ω) 2 F Whenever ρ 3 (1) > 0 (P Ω c(x + (1)Y + (1) XY ) 0) and ω 1 > 1, then ω 2 > 1 can be chosen so that ρ 4 (ω) > 0, ω (1, ω 2 ].
13 Reduction of the residual S 2 F S +(ω) 2 F ρ 12 (ω) ρ 3 (ω) ρ 4 (ω) 2 S F 2 S+ (ω) F ω (a)
14 Reduction of the residual S 2 F S +(ω) 2 F ρ 12 (ω) ρ 3 (ω) ρ 4 (ω) 2 S F 2 S+ (ω) F ω (b)
15 Nonlinear SOR: convergence guarantee Problem: how can we select a proper weight ω to ensure convergence for a nonlinear model? Strategy: Adjust ω dynamically according to the change of the objective function values. Calculate the residual ratio γ(ω) = S+(ω) F S F A small γ(ω) indicates that the current weight value ω works well so far. If γ(ω) < 1, accept the new point; otherwise, ω is reset to 1 and this procedure is repeated. ω is increased only if the calculated point is acceptable but the residual ratio γ(ω) is considered too large ; that is, γ(ω) [γ 1, 1) for some γ 1 (0, 1).
16 Nonlinear SOR: complete algorithm Algorithm 1: A low-rank matrix fitting algorithm (LMaFit) Input index set Ω, data P Ω (M) and a rank overestimate K r. Set Y 0, Z 0, ω = 1, ω > 1, δ > 0, γ 1 (0, 1) and k = 0. while not convergent do Compute (X + (ω), Y + (ω), Z + (ω)). Compute the residual ratio γ(ω). if γ(ω) 1 then set ω = 1 and go to step 4. Update (X k+1, Y k+1, Z k+1 ) and increment k. if γ(ω) γ 1 then set δ = max(δ, 0.25(ω 1)) and ω = min(ω + δ, ω).
17 nonlinear GS.vs. nonlinear SOR normalized residual SOR: k=12 SOR: k=20 GS: k=12 GS: k=20 normalized residual SOR: k=12 SOR: k=20 GS: k=12 GS: k= iteration (a) n=1000, r=10, SR = iteration (b) n=1000, r=10, SR=0.15
18 Extension to problems with general linear constraints Consider problem: min W R m n Nonconvex relaxation: rank(w ), s.t. A(W ) = b min X,Y,Z 1 2 XY Z 2 F s.t. A(Z ) = b Let S := A (AA ) (b A(XY )) and Z ω = XY + ωs Nonlinear SOR scheme: X + (ω) Z ω Y or Z ω Y (YY ), Y + (ω) (X + (ω) X + (ω)) (X + (ω) Z ω ), Z + (ω) X + (ω)y + (ω) + A (AA ) (b A(X + (ω)y + (ω))).
19 Nonlinear SOR: convergence analysis Main result: Let {(X k, Y k, Z k )} be generated by the nonlinear SOR method and {P Ω c(x k Y k )} be bounded. Then there exists at least a subsequence of {(X k, Y k, Z k )} that satisfies the first-order optimality conditions in the limit. Technical lemmas: ωs (X + (ω)y + (ω) XY ) = X + (ω)y + (ω) XY 2 F. Residual reduction from S 2 F after the first two steps 1 ω 2 X +(ω)y + (ω) Z ω 2 F = (I Q(ω))S(I P) 2 F = S 2 F ρ 12(ω) lim ω 1 + ρ 4(ω) ω 1 = 2 P Ω c(x +(1)Y + (1) XY ) 2 F 0
20 Practical issues Variants to obtain X + (ω) and Y + (ω): linsolve.vs. QR Storage: full Z or partial S = P Ω (M XY ) Stopping criteria: relres = P Ω(M X k Y k ) F tol and reschg = P Ω (M) F 1 P Ω(M X k Y k ) F P Ω (M X k 1 Y k 1 ) F tol/2, Rank estimation: Start from K r then decrease K aggressively QR = XE, E is permutation, d := diag(r) nonincreasing compute the sequnce d i = d i /d i+1, i = 1,, K 1, examine the ratio τ = (K 1) d(p) i p d i, d(p) is the maximal element of { d i } and p is the corresponding index. reset K to p once τ > 10 Start from a small K and increase K to min(k + κ, rank_max) when the alg. stagnates
21 The sensitivity with respect to the rank estimation K SR=0.04 SR=0.08 SR=0.3 iteration number SR=0.04 SR=0.08 SR=0.3 cpu rank estimation (a) Iteration number rank estimation (b) CPU time in seconds
22 Convergence behavior of the residual r=10 r=50 r= SR=0.04 SR=0.08 SR=0.3 normalized residual normalized residual iteration (a) m = n = 1000, sr= iteration (b) m = n = 1000, r = 10
23 Phase diagrams for matrix completion recoverability sampling ratio sampling ratio rank (a) Model solved by FPCA rank (b) Model solved by LMaFit 0
24 Small random problems: m = n = 1000 Problem APGL FPCA LMaFit K = 1.25r K = 1.5r r SR FR µ time rel.err tsvd time rel.err tsvd time rel.err time rel.err e e-03 82% e-01 12% e e e e-04 71% e-04 19% e e e e-04 66% e-04 42% e e e e-04 58% e-04 72% e e e e-03 93% e-04 56% e e e e-04 87% e-04 67% e e e e-04 85% e-04 75% e e e e-04 82% e-04 77% e e e e-03 92% e-04 77% e e e e-04 91% e-04 79% e e e e-04 90% e-04 82% e e e e-04 89% e-04 81% e e-05
25 Large random problems Problem APGL LMaFit (K = 1.25r ) LMaFit (K = 1.5r ) n r SR FR µ iter #sv time rel.err iter #sv time rel.err iter #sv time rel.err e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-4
26 Random low-rank approximation problems σ σ σ index σ index index (a) power-low: σ i = i index (b) exponentially: σ i = e 0.3i
27 Random low-rank approximation problems Problem APGL FPCA LMaFit(est_rank=1) LMaFit(est_rank=2) SR FR µ #sv time rel.err #sv time rel.err #sv time rel.err #sv time rel.err power-low decaying e e e e e e e e e e e e e e e e e e e e-04 exponentially decaying e e e e e e e e e e e e e e e e e e e e-04
28 Video Denoising original video 50% masked original video
29 Video Denoising recovered video by LMaFit recovered video by APGL APGL LMaFit m/n µ iter #sv time rel.err iter #sv time rel.err 76800/ e e e-02
30 Contribution Explore matrix completion by using the low-rank facotrization model Reduce the cost of the nonlinear Gauss-Seideal scheme by eliminating an unecessary least square problem Propose a nonlinear successive over-relaxation (SOR) algorithm with convergence guarantee Adjust the relaxation weight of SOR dynamically Excellent computational performance on a wide range of test problems
31 Matrix separation Consider: min U,V,Z P Ω (Z D) 1 s.t. UV Z = 0 Introduce the augmented Lagrangian function L β (U, V, Z, Λ) = P Ω (Z D) 1 + Λ, UV Z + β 2 UV Z 2 F, Alternating direction augmented Lagrangian framework: U j+1 := arg min U R m k L β (U, V j, Z j, Λ j ), V j+1 := arg min L β (U j+1, V, Z j, Λ j ), V R k n Z j+1 := arg min Z R m n L β(u j+1, V j+1, Z, Λ j ), Λ j+1 := Λ j + γβ(u j+1 V j+1 Z j+1 ).
32 ADM subproblems Let B = Z Λ/β, then U + = BV (VV ) and V + = (U +U + ) U +B Since U + V + = U + (U +U + ) U +B = P U+ B, then: Q := orth(bv ), U + = Q and V + = Q B Variable Z : ( P Ω (Z + ) = P Ω (S U + V + D + Λ β, 1 ) ) + D β ( P Ω c(z + ) = P Ω c U + V + + Λ ) β
33 Theoretical results Let X (U, V, Z, Λ). The KKT conditions are ΛV = 0, U Λ = 0, Z ( Z D 1 ) = Λ, UV Z = 0. The equation Z ( Z D 1 ) = Λ is equivalent to Z D = S(Z D + Λ/β, 1/β) = S(UV D + Λ/β, 1/β), Let {X j } p=1 be an uniformly bounded sequence generated by ADM with lim (X j+1 X j ) = 0. p Then any accumulation point of {X j } p=1 satisfies the KKT conditions.
34 CPU and rel. err vs. sparsity: m = 100 and k = 5 CPU time vs. Sparsity Relative Error vs. Sparsity CPU time 10 1 Rel Error 10 5 LMaFit IALM % of Nonzeros (a) CPU seconds LMaFit IALM % of Nonzeros (b) Relative Error CPU time CPU time vs. Sparsity LMaFit IALM LRSD Rel Error Relative Error vs. Sparsity LMaFit IALM LRSD % of Nonzeros (c) CPU seconds (with noise) % of Nonzeros (d) Relative Error (with noise)
35 Recoverability test: phase plots Digits of accuracy Digits of accuracy Rank Rank Sparsity % (a) IALM Sparsity % (b) LMaFit
36 Performance with respect to size CPU time LRSD LMaFit IALM dimension (a) k =0.05m and S 0 =0.05m 2 CPU time LRSD LMaFit IALM dimension (b) k =0.15m and S 0 =0.15m 2
37 Deterministic low-rank matrices Figure: Recovered results of brickwall". Upper left: original image; upper right: corrupted image; lower left: IALM; lower right: LMaFit.
38 Video separation Figure: original, separated results by IALM and LMaFit.
39 Video separation Figure: original, separated results by IALM and LMaFit.
40 Upcoming Talks: Oct. 1?th Wotao, Yin, Fast Curvilinear Search Algorithms for Optimization with Constraints x 2 = 1 or X X = I p-harmonic flow into spheres polynomial optimization with normalized constraints maxcut SDP relaxation low-rank nearest correlation matrix estimation linear and nonlinear eigenvalue problems quadratic assignment problem
Solving A Low-Rank Factorization Model for Matrix Completion by A Nonlinear Successive Over-Relaxation Algorithm
Solving A Low-Rank Factorization Model for Matrix Completion by A Nonlinear Successive Over-Relaxation Algorithm Zaiwen Wen, Wotao Yin, Yin Zhang 2010 ONR Compressed Sensing Workshop May, 2010 Matrix Completion
More informationSOLVING A LOW-RANK FACTORIZATION MODEL FOR MATRIX COMPLETION BY A NONLINEAR SUCCESSIVE OVER-RELAXATION ALGORITHM
SOLVING A LOW-RANK FACTORIZATION MODEL FOR MATRIX COMPLETION BY A NONLINEAR SUCCESSIVE OVER-RELAXATION ALGORITHM ZAIWEN WEN, WOTAO YIN, AND YIN ZHANG CAAM TECHNICAL REPORT TR10-07 DEPARTMENT OF COMPUTATIONAL
More informationAugmented Lagrangian alternating direction method for matrix separation based on low-rank factorization
Augmented Lagrangian alternating direction method for matrix separation based on low-rank factorization Yuan Shen Zaiwen Wen Yin Zhang January 11, 2011 Abstract The matrix separation problem aims to separate
More informationLecture: Matrix Completion
1/56 Lecture: Matrix Completion http://bicmr.pku.edu.cn/~wenzw/bigdata2017.html Acknowledgement: this slides is based on Prof. Jure Leskovec and Prof. Emmanuel Candes s lecture notes Recommendation systems
More informationarxiv: v2 [cs.na] 24 Mar 2015
Volume X, No. 0X, 200X, X XX doi:1934/xx.xx.xx.xx PARALLEL MATRIX FACTORIZATION FOR LOW-RANK TENSOR COMPLETION arxiv:1312.1254v2 [cs.na] 24 Mar 2015 Yangyang Xu Department of Computational and Applied
More information1 Non-negative Matrix Factorization (NMF)
2018-06-21 1 Non-negative Matrix Factorization NMF) In the last lecture, we considered low rank approximations to data matrices. We started with the optimal rank k approximation to A R m n via the SVD,
More informationRobust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds
Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds Tao Wu Institute for Mathematics and Scientific Computing Karl-Franzens-University of Graz joint work with Prof.
More informationElaine T. Hale, Wotao Yin, Yin Zhang
, Wotao Yin, Yin Zhang Department of Computational and Applied Mathematics Rice University McMaster University, ICCOPT II-MOPTA 2007 August 13, 2007 1 with Noise 2 3 4 1 with Noise 2 3 4 1 with Noise 2
More informationBias-free Sparse Regression with Guaranteed Consistency
Bias-free Sparse Regression with Guaranteed Consistency Wotao Yin (UCLA Math) joint with: Stanley Osher, Ming Yan (UCLA) Feng Ruan, Jiechao Xiong, Yuan Yao (Peking U) UC Riverside, STATS Department March
More informationIterative Methods. Splitting Methods
Iterative Methods Splitting Methods 1 Direct Methods Solving Ax = b using direct methods. Gaussian elimination (using LU decomposition) Variants of LU, including Crout and Doolittle Other decomposition
More informationSparse Optimization Lecture: Basic Sparse Optimization Models
Sparse Optimization Lecture: Basic Sparse Optimization Models Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know basic l 1, l 2,1, and nuclear-norm
More informationLecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University
Lecture Note 7: Iterative methods for solving linear systems Xiaoqun Zhang Shanghai Jiao Tong University Last updated: December 24, 2014 1.1 Review on linear algebra Norms of vectors and matrices vector
More informationON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS
ON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS WEI DENG AND WOTAO YIN Abstract. The formulation min x,y f(x) + g(y) subject to Ax + By = b arises in
More informationLINEAR SYSTEMS (11) Intensive Computation
LINEAR SYSTEMS () Intensive Computation 27-8 prof. Annalisa Massini Viviana Arrigoni EXACT METHODS:. GAUSSIAN ELIMINATION. 2. CHOLESKY DECOMPOSITION. ITERATIVE METHODS:. JACOBI. 2. GAUSS-SEIDEL 2 CHOLESKY
More informationTRANSFORMED SCHATTEN-1 ITERATIVE THRESHOLDING ALGORITHMS FOR LOW RANK MATRIX COMPLETION
TRANSFORMED SCHATTEN- ITERATIVE THRESHOLDING ALGORITHMS FOR LOW RANK MATRIX COMPLETION SHUAI ZHANG, PENGHANG YIN, AND JACK XIN Abstract. We study a non-convex low-rank promoting penalty function, the transformed
More informationCoordinate Update Algorithm Short Course Proximal Operators and Algorithms
Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow
More informationEE 381V: Large Scale Optimization Fall Lecture 24 April 11
EE 381V: Large Scale Optimization Fall 2012 Lecture 24 April 11 Lecturer: Caramanis & Sanghavi Scribe: Tao Huang 24.1 Review In past classes, we studied the problem of sparsity. Sparsity problem is that
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationSolving linear equations with Gaussian Elimination (I)
Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian
More informationComputational Methods. Least Squares Approximation/Optimization
Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the
More informationCourse Notes: Week 1
Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues
More informationBlock Coordinate Descent for Regularized Multi-convex Optimization
Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline
More informationFunctional SVD for Big Data
Functional SVD for Big Data Pan Chao April 23, 2014 Pan Chao Functional SVD for Big Data April 23, 2014 1 / 24 Outline 1 One-Way Functional SVD a) Interpretation b) Robustness c) CV/GCV 2 Two-Way Problem
More informationS 1/2 Regularization Methods and Fixed Point Algorithms for Affine Rank Minimization Problems
S 1/2 Regularization Methods and Fixed Point Algorithms for Affine Rank Minimization Problems Dingtao Peng Naihua Xiu and Jian Yu Abstract The affine rank minimization problem is to minimize the rank of
More informationThe matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.
) Find all solutions of the linear system. Express the answer in vector form. x + 2x + x + x 5 = 2 2x 2 + 2x + 2x + x 5 = 8 x + 2x + x + 9x 5 = 2 2 Solution: Reduce the augmented matrix [ 2 2 2 8 ] to
More informationEUSIPCO
EUSIPCO 013 1569746769 SUBSET PURSUIT FOR ANALYSIS DICTIONARY LEARNING Ye Zhang 1,, Haolong Wang 1, Tenglong Yu 1, Wenwu Wang 1 Department of Electronic and Information Engineering, Nanchang University,
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently
More informationUser s Guide for LMaFit: Low-rank Matrix Fitting
User s Guide for LMaFit: Low-rank Matrix Fitting Yin Zhang Department of CAAM Rice University, Houston, Texas, 77005 (CAAM Technical Report TR09-28) (Versions beta-1: August 23, 2009) Abstract This User
More informationMulti-stage convex relaxation approach for low-rank structured PSD matrix recovery
Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun
More informationA Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations
A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations Chuangchuang Sun and Ran Dai Abstract This paper proposes a customized Alternating Direction Method of Multipliers
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationOrthonormal Transformations and Least Squares
Orthonormal Transformations and Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 30, 2009 Applications of Qx with Q T Q = I 1. solving
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 9 Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 2 Separable convex optimization a special case is min f(x)
More informationCoordinate Update Algorithm Short Course Operator Splitting
Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Lecture 5: Numerical Linear Algebra Cho-Jui Hsieh UC Davis April 20, 2017 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationLINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION
LINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION JUNFENG YANG AND XIAOMING YUAN Abstract. The nuclear norm is widely used to induce low-rank solutions for
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More information1 Computing with constraints
Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)
More informationRobust Principal Component Analysis
ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M
More informationFast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma
Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University 8 September 2003 European Union RTN Summer School on Multi-Agent
More informationBindel, Fall 2009 Matrix Computations (CS 6210) Week 8: Friday, Oct 17
Logistics Week 8: Friday, Oct 17 1. HW 3 errata: in Problem 1, I meant to say p i < i, not that p i is strictly ascending my apologies. You would want p i > i if you were simply forming the matrices and
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationLow rank matrix completion by alternating steepest descent methods
Low rank matrix completion by alternating steepest descent methods Jared Tanner a, Ke Wei b a Mathematical Institute, University of Oxford, Andrew Wiles Building, Radcliffe Observatory Quarter, Woodstock
More informationQR Decomposition. When solving an overdetermined system by projection (or a least squares solution) often the following method is used:
(In practice not Gram-Schmidt, but another process Householder Transformations are used.) QR Decomposition When solving an overdetermined system by projection (or a least squares solution) often the following
More informationC&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz
C&O367: Nonlinear Optimization (Winter 013) Assignment 4 H. Wolkowicz Posted Mon, Feb. 8 Due: Thursday, Feb. 8 10:00AM (before class), 1 Matrices 1.1 Positive Definite Matrices 1. Let A S n, i.e., let
More informationUNIT 6: The singular value decomposition.
UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T
More informationA New Trust Region Algorithm Using Radial Basis Function Models
A New Trust Region Algorithm Using Radial Basis Function Models Seppo Pulkkinen University of Turku Department of Mathematics July 14, 2010 Outline 1 Introduction 2 Background Taylor series approximations
More informationOne Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017
One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 1 One Picture and a Thousand Words Using Matrix Approximations Dianne P. O Leary
More informationPenalty Decomposition Methods for Rank and l 0 -Norm Minimization
IPAM, October 12, 2010 p. 1/54 Penalty Decomposition Methods for Rank and l 0 -Norm Minimization Zhaosong Lu Simon Fraser University Joint work with Yong Zhang (Simon Fraser) IPAM, October 12, 2010 p.
More informationRecent Developments in Compressed Sensing
Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline
More informationAlternating Direction Augmented Lagrangian Algorithms for Convex Optimization
Alternating Direction Augmented Lagrangian Algorithms for Convex Optimization Joint work with Bo Huang, Shiqian Ma, Tony Qin, Katya Scheinberg, Zaiwen Wen and Wotao Yin Department of IEOR, Columbia University
More informationToday s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn
Today s class Linear Algebraic Equations LU Decomposition 1 Linear Algebraic Equations Gaussian Elimination works well for solving linear systems of the form: AX = B What if you have to solve the linear
More informationConditions for Robust Principal Component Analysis
Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and
More informationAn accelerated proximal gradient algorithm for nuclear norm regularized linear least squares problems
An accelerated proximal gradient algorithm for nuclear norm regularized linear least squares problems Kim-Chuan Toh Sangwoon Yun March 27, 2009; Revised, Nov 11, 2009 Abstract The affine rank minimization
More informationMinimizing the Difference of L 1 and L 2 Norms with Applications
1/36 Minimizing the Difference of L 1 and L 2 Norms with Department of Mathematical Sciences University of Texas Dallas May 31, 2017 Partially supported by NSF DMS 1522786 2/36 Outline 1 A nonconvex approach:
More informationCHAPTER 11. A Revision. 1. The Computers and Numbers therein
CHAPTER A Revision. The Computers and Numbers therein Traditional computer science begins with a finite alphabet. By stringing elements of the alphabet one after another, one obtains strings. A set of
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional
More informationLecture 18 Classical Iterative Methods
Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional
More informationSolving large Semidefinite Programs - Part 1 and 2
Solving large Semidefinite Programs - Part 1 and 2 Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Singapore workshop 2006 p.1/34 Overview Limits of Interior
More informationGoDec: Randomized Low-rank & Sparse Matrix Decomposition in Noisy Case
ianyi Zhou IANYI.ZHOU@SUDEN.US.EDU.AU Dacheng ao DACHENG.AO@US.EDU.AU Centre for Quantum Computation & Intelligent Systems, FEI, University of echnology, Sydney, NSW 2007, Australia Abstract Low-rank and
More informationPreliminary Examination, Numerical Analysis, August 2016
Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any
More informationOn l 1 Minimization for Ill-Conditioned Linear Systems with Piecewise Polynomial Solutions
RICE UNIVERSITY On l 1 Minimization for Ill-Conditioned Linear Systems with Piecewise Polynomial Solutions by Jorge A. Castañón A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree
More informationSparse and Low-Rank Matrix Decomposition Via Alternating Direction Method
Sparse and Low-Rank Matrix Decomposition Via Alternating Direction Method Xiaoming Yuan a, 1 and Junfeng Yang b, 2 a Department of Mathematics, Hong Kong Baptist University, Hong Kong, China b Department
More informationBlockMatrixComputations and the Singular Value Decomposition. ATaleofTwoIdeas
BlockMatrixComputations and the Singular Value Decomposition ATaleofTwoIdeas Charles F. Van Loan Department of Computer Science Cornell University Supported in part by the NSF contract CCR-9901988. Block
More informationQALGO workshop, Riga. 1 / 26. Quantum algorithms for linear algebra.
QALGO workshop, Riga. 1 / 26 Quantum algorithms for linear algebra., Center for Quantum Technologies and Nanyang Technological University, Singapore. September 22, 2015 QALGO workshop, Riga. 2 / 26 Overview
More information6. Iterative Methods for Linear Systems. The stepwise approach to the solution...
6 Iterative Methods for Linear Systems The stepwise approach to the solution Miriam Mehl: 6 Iterative Methods for Linear Systems The stepwise approach to the solution, January 18, 2013 1 61 Large Sparse
More information9. Iterative Methods for Large Linear Systems
EE507 - Computational Techniques for EE Jitkomut Songsiri 9. Iterative Methods for Large Linear Systems introduction splitting method Jacobi method Gauss-Seidel method successive overrelaxation (SOR) 9-1
More informationNumerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??
Metodi Numerici M p. 1/?? Numerical Methods Elena loli Piccolomini Civil Engeneering http://www.dm.unibo.it/ piccolom elena.loli@unibo.it Metodi Numerici M p. 2/?? Least Squares Data Fitting Measurement
More informationGauge optimization and duality
1 / 54 Gauge optimization and duality Junfeng Yang Department of Mathematics Nanjing University Joint with Shiqian Ma, CUHK September, 2015 2 / 54 Outline Introduction Duality Lagrange duality Fenchel
More informationCSC 576: Variants of Sparse Learning
CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in
More informationStructured matrix factorizations. Example: Eigenfaces
Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix
More informationMath 471 (Numerical methods) Chapter 3 (second half). System of equations
Math 47 (Numerical methods) Chapter 3 (second half). System of equations Overlap 3.5 3.8 of Bradie 3.5 LU factorization w/o pivoting. Motivation: ( ) A I Gaussian Elimination (U L ) where U is upper triangular
More informationAn algebraic perspective on integer sparse recovery
An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:
More informationJordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS
Jordan Journal of Mathematics and Statistics JJMS) 53), 2012, pp.169-184 A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS ADEL H. AL-RABTAH Abstract. The Jacobi and Gauss-Seidel iterative
More informationarxiv: v1 [cs.it] 26 Oct 2018
Outlier Detection using Generative Models with Theoretical Performance Guarantees arxiv:1810.11335v1 [cs.it] 6 Oct 018 Jirong Yi Anh Duc Le Tianming Wang Xiaodong Wu Weiyu Xu October 9, 018 Abstract This
More informationNumerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization /36-725
Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: proximal gradient descent Consider the problem min g(x) + h(x) with g, h convex, g differentiable, and h simple
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationNumerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization
Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725 Consider Last time: proximal Newton method min x g(x) + h(x) where g, h convex, g twice differentiable, and h simple. Proximal
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationLinear Algebra. Session 12
Linear Algebra. Session 12 Dr. Marco A Roque Sol 08/01/2017 Example 12.1 Find the constant function that is the least squares fit to the following data x 0 1 2 3 f(x) 1 0 1 2 Solution c = 1 c = 0 f (x)
More information7.3 The Jacobi and Gauss-Siedel Iterative Techniques. Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP.
7.3 The Jacobi and Gauss-Siedel Iterative Techniques Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP. 7.3 The Jacobi and Gauss-Siedel Iterative Techniques
More informationPrincipal Component Analysis
Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used
More informationLinear Algebra in Actuarial Science: Slides to the lecture
Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations
More information8. Diagonalization.
8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationComputational Methods. Systems of Linear Equations
Computational Methods Systems of Linear Equations Manfred Huber 2010 1 Systems of Equations Often a system model contains multiple variables (parameters) and contains multiple equations Multiple equations
More informationMatrix Rank-One Decomposition. and Applications
Matrix rank-one decomposition and applications 1 Matrix Rank-One Decomposition and Applications Shuzhong Zhang Department of Systems Engineering and Engineering Management The Chinese University of Hong
More information15-780: LinearProgramming
15-780: LinearProgramming J. Zico Kolter February 1-3, 2016 1 Outline Introduction Some linear algebra review Linear programming Simplex algorithm Duality and dual simplex 2 Outline Introduction Some linear
More informationProbabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms
Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,
More informationSubspace Projection Matrix Completion on Grassmann Manifold
Subspace Projection Matrix Completion on Grassmann Manifold Xinyue Shen and Yuantao Gu Dept. EE, Tsinghua University, Beijing, China http://gu.ee.tsinghua.edu.cn/ ICASSP 2015, Brisbane Contents 1 Background
More informationMath 273a: Optimization Overview of First-Order Optimization Algorithms
Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization
More informationFast and Robust Phase Retrieval
Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National
More informationRank minimization via the γ 2 norm
Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises
More informationThe Singular Value Decomposition
The Singular Value Decomposition An Important topic in NLA Radu Tiberiu Trîmbiţaş Babeş-Bolyai University February 23, 2009 Radu Tiberiu Trîmbiţaş ( Babeş-Bolyai University)The Singular Value Decomposition
More informationACCELERATED LINEARIZED BREGMAN METHOD. June 21, Introduction. In this paper, we are interested in the following optimization problem.
ACCELERATED LINEARIZED BREGMAN METHOD BO HUANG, SHIQIAN MA, AND DONALD GOLDFARB June 21, 2011 Abstract. In this paper, we propose and analyze an accelerated linearized Bregman (A) method for solving the
More informationProcess Model Formulation and Solution, 3E4
Process Model Formulation and Solution, 3E4 Section B: Linear Algebraic Equations Instructor: Kevin Dunn dunnkg@mcmasterca Department of Chemical Engineering Course notes: Dr Benoît Chachuat 06 October
More informationhttps://goo.gl/kfxweg KYOTO UNIVERSITY Statistical Machine Learning Theory Sparsity Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT OF INTELLIGENCE SCIENCE AND TECHNOLOGY 1 KYOTO UNIVERSITY Topics:
More informationUses of duality. Geoff Gordon & Ryan Tibshirani Optimization /
Uses of duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Remember conjugate functions Given f : R n R, the function is called its conjugate f (y) = max x R n yt x f(x) Conjugates appear
More information