Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery
|
|
- Ruth Thornton
- 6 years ago
- Views:
Transcription
1 Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun Bi and Shaohua Pan) May 22, 2014 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 1 / 30
2 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 2 / 30
3 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 3 / 30
4 Matrix recovery Given some measurements for a (structured) matrix X H n + (Hermitian positive semi-definite), how to recover this unknown matrix? b k = Θ k, X + ξ k, k = 1, 2,..., m. b k : observed data (with / without noise) ξ k : additive noise Possible? Yes, with the low-rank structure. Entries = basis coefficients: {e i e T j 1 i n 1, 1 j n 2 }. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 4 / 30
5 Correlation matrix recovery A low-rank correlation matrix X S n +: symmetric positive semidefinite and diag(x) = e. Observations (with / without noise): b k = Θ k, X + ξ k, k = 1,..., m. How to recover the unknown correlation matrix X? All diagonal entries are fixed. Some off-diagonal entries may also be fixed, [e.g., the correlations among pegged currencies]. Entries basis coefficients { e i e T i 1 i n } { 1 2 (e i e T j +e j e T i ) 1 i<j n }. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 5 / 30
6 Density matrix recovery A low-rank density matrix X H n + with n = 2 l : Hermitian positive semidefinite and Tr(X) = 1. Observations Pauli measurements: Θ k, X = Re(Tr(Θ k X)), k = 1,..., m, where Θ k Pauli basis { σ s1 σ sl (s 1,, s l ) {0, 1, 2, 3} l} with σ 0 = ( ) 1 0, σ = How to recover the unknown density matrix X? ( ) ( ) ( ) , σ = 1 0, σ =. 0 1 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 6 / 30
7 Observation model In this talk, we consider the problem of low-rank structured PSD matrix recovery with noise. The observation model takes the following form b k = Θ k, X + ξ k, k = 1,..., m. Define A(X) := ( Θ 1, X,, Θ m, X ) T R m. Then, we have that b = A(X) + ξ. Let X H n + be the unknown low-rank structured true matrix. There exists a constant δ 0, A(X) b δ. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 7 / 30
8 The rank minimization estimator The rank minimization estimator: X arg min Ω H n + { rank(x) : A(X) b δ where Ω is a closed convex set characterizing the structure of X. This is the best possible estimator. But it is very challenging to get the global solution due to the nonconvexity and discontinuity of the rank function. } / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 8 / 30
9 The nuclear norm technique The nuclear norm technique has gained success in many applications: n rank(x) = X := σ i (X), where σ 1 (X) σ n (X) denote all the singular values of X. Nuclear norm convex envelope of the rank function over the unit ball of the spectral norm. y i=1 x /a 1 rank x a 0 a x / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 9 / 30
10 Nuclear norm estimator Nuclear norm estimator: X arg min Ω H n + { } X : A(X) b δ Extra biased (for large singular values). It may fail when certain rows and / or columns are heavily sampled (Salakhutdinov and Srebro, 2010). It may not be rank consistent for general sampling even for the low-dimensional case (Bach, 2008). For correlation and density matrices, X = constant = the nuclear norm technique s power is limited. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 10 / 30
11 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 11 / 30
12 Equivalent MPSDEC Let C(J) be the family of closed convex functions φ : J R with J [0, 1] and φ(0) = 0, 0 < t := arg min 0 t 1 φ(t) and φ (0) < +. Rank minimization estimator is equivalent to the following mathematical program with the PSD equilibrium constraint (MPSDEC) min I, Φ(W) X Ω,W Hn s.t. A(X) b δ (1) W, X = 0, X 0, 0 W I Φ the Löwner operator associated to φ C(J) The bilinear constraint W, X = 0 is the trouble maker. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 12 / 30
13 Exact penalty for MPSDEC To deal with W, X = 0 efficiently, we consider the penalized version min X Ω,W Hn I, Φ(W) + ρ W, X s.t. A(X) b δ (2) X 0, 0 W I Theorem The cost function of (2) has a desired structure though it is nonconvex The penalty problem (2) is exact in the following sense There exists ρ > 0 such that when ρ > ρ, the set of global optimal solutions of the penalty problem (2) coincides with that of the MPSDEC. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 13 / 30
14 Multi-stage convex relaxation approach (S0) Select a function φ C(J). Let W 0 := I and set k := 1. Choose ρ 0 > 0. (S1) Solve the following weighted trace-norm minimization problem X k arg min X Ω H+ n { W k 1, X : A(X) b δ }. (S2) By the information of X k select a suitable ρ k = δ k ρ k 1 with δ k 1. (S3) Solve the following minimization problem W k arg min 0 W I (S4) Set k := k + 1, and go to (S.1). { I, Φ(W) + ρ k W, X k }. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 14 / 30
15 Multi-stage convex relaxation (Examples) The convex relaxation approach is solving (2) in an alternating way. The computational cost is the solution of the weighted trace-norm minimization problem since if X k = U k Diag(λ(X k ))(U k ) T we have W k = U k Diag(w k 1,..., wk n)(u k ) T. 1 φ 1 (t) = t w k i = { 0 if λi (X k ) 1 ρ k 1 otherwise 2 φ 2 (t) = a2 1 4 t 2 at w k i = min( 2 a 2 1 max(a ρ kλ i (X k ), 0), 1). 3 φ 3 (t) = t + 1 t+ɛ 1 ɛ w k 1 i = max( ɛ, 0). ρk λ i (X k )+1 4 φ 4 (t) = t ln(t + ɛ) + ln(ɛ) w k 1 i = max( ρ k λ i ɛ, 0). (X k )+1 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 15 / 30
16 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 16 / 30
17 Notation and assumptions Recall that rank(x) = r. Assume that X = [U 1, U 2 ]Diag(λ(X))[U 1, U 2 ] T, where U 1 O n r and U 2 O n (n r). For any X H n, define P T (X) = U 1 U T 1 X + XU 1 U T 1 U 1 U T 1 XU 1 U T 1. Let Q = A A. Define the restricted eigenvalues of Q by (Tong Zhang, 2010) X, Q(X) X, Q(X) ϑ + (k) := sup 0<rank(X) k X 2 F, ϑ (k) := inf 0<rank(X) k X 2 F Assumption: There exist a c [0, 1) and an integer s (1 s n 2r 2 ) s.t. ϑ + (s) ϑ (2r + 2s) 1 + 2c2 s. r / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 17 / 30
18 General error bound Theorem Let γ k 1 := P T(W k 1 ) λ min ((U 2 ) T W k 1 U 2 ). If 0 γ k 1 1 c, then Xk X F Ξ(γ k 1 ) with Ξ(t) := 1 + rt2 2s 2 ϑ + (2r + s)δ (1 ct) ϑ (2r + s) Ξ(1) the error bound for the nuclear norm estimator. 5 2 ) 3 2 (1 δ 3r) If c < 1 1 δ 3r(1+ Ξ(1) = for t 0., then we have the following inequality 6 ϑ+ (3r)δ (1 c) ϑ (3r) δ3r δ (1 c) (1 δ 3r ) < 3δ 1 + δ3r 1 δ 3r (1 + The last term is the bound given by Mohan and Fazel (2010). 5 2 ). / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 18 / 30
19 Error bounds in the first two stages Let ψ be the conjugate function of ψ(t) := { φ(t) if t [0, 1], + otherwise. Define a 0 := (ψ ) +(0), a 1 := (ψ ) ( ρ1 λ r+1 (X 1 ) ) and b 1 := (ψ ) +( ρ1 λ r (X 1 ) ). Theorem Suppose that λ r (X) = β X 1 X F for some β > 2. If the parameter ρ 1 is chosen such that ρ 1 ( 0, φ + (0) ) b 1 + a 0 β ln λ r+1 (X 1 ) and 2 β 2 < 1, then a 1 a 1 2 ln β β 2 X 2 X F Ξ(γ 1 ) Ξ b 1 + a 0 a 1 a 1 β 2 ln β 2 β 2 ln β 2 < Ξ(1). (3) / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 19 / 30
20 Reduction ratio given by different φ Table: The ratio of the error bounds given by different φ for the first two stages φ 1 ( φ 2 ( φ 3 ( ( φ 4 ( φ 4 ρ 1 β ratio ) Ξ(γ ) λ r+1 (X 1 ) Ξ(1) ) Ξ(γ ) λ r+1 (X 1 ) Ξ(1) ) Ξ(γ ) λ r+1 (X 1 ) Ξ(1) ) Ξ(γ ) λ r+1 (X 1 ) Ξ(1) ) Ξ(γ ) Ξ(1) λ r(x 1 ), 1 2 λ r(x 1 ), 1 3 λ r(x 1 ), λ r(x 1 ), λ r(x 1 ), 0.8 λ r+1 (X 1 ) / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 20 / 30
21 Guaranteed error bound reduction (Cont.) Let b k := (ψ ) +[ ρk (λ r (X ) Ξ( γ k 1 )) ], ã k := (ψ ) ( ρ k Ξ( γ k 1 )), γ k := b k + a 0 2 ln ã k ãk 2 ln λ r(x) λ r(x) 2Ξ( γ k 1 ) λ r(x) λ r(x) 2Ξ( γ k 1 ) for k 1 and γ 0 = 1. Theorem Suppose that λ r (X) = βξ(1) for some β > 2. If the parameter ρ 1 is chosen such that ρ 1 ( 0, φ + (0) ) b 1 + a 0 β ln Ξ(1) and 2 β 2 < 1, and the parameters σ k are chosen ã 1 ã1 2 ln β β 2 such that σ k [ 1, Ξ( γ k 2) Ξ( γ k 1 )], then for all k 2 we have γk 1 γ k 1 < γ k 2 and X k X F Ξ(γ k 1 ) Ξ( γ k 1 ) < Ξ( γ k 2 ). (4) / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 21 / 30
22 Reduction ratio of the error bounds Table: The ratio of the error bounds obtained with φ 2 for the first five stages ( φ Ξ(1), 1.2 ) Ξ(1) ρ 1 ratio Ξ(γ 1 )/Ξ(1) Ξ(γ 2 )/Ξ(1) Ξ(γ 3 )/Ξ(1) Ξ(γ 4 )/Ξ(1) Ξ(γ 5 )/Ξ(1) / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 22 / 30
23 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 23 / 30
24 The performance with a high sample ratio φ 1 φ 2 with a=2.1 φ 3 with ε=0.4 φ 4 with ε= φ 1 φ 2 with a=2.1 φ 3 with ε=0.4 φ 4 with ε= Relative error (%) Computing time (in seconds) (a) The number of stages (b) The number of stages Figure: The performance of multi-stage convex relaxation with p = / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 24 / 30
25 The performance with a low sample ratio φ 1 φ 2 with a=2.1 φ with ε=0.4 3 φ with ε= φ 1 φ 2 with a=2.1 φ with ε=0.4 3 φ with ε= Relative error (%) Computing time (in seconds) (a) The number of stages (b) The number of stages Figure: The performance of multi-stage convex relaxation with p = 3980 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 25 / 30
26 Low-rank correlation matrix recovery Table: Performance for the correlation matrix recovery problems with n = 1000 r 5 10 off diag sr(%) One-stage Two-stage Final result relerr(rank) time relerr(rank) time k relerr(rank) time e-1(1000) e-1(19) e-1(5) e-1(998) e-1(5) e-1(5) e-2(999) e-2(5) e-2(5) e-1(991) e-1(18) e-1(5) e-1(977) e-1(5) e-1(5) e-1(703) e-2(6) e-2(5) e-1(1000) e-1(21) e-1(10) e-1(1000) e-1(10) e-1(10) e-2(1000) e-2(10) e-2(10) e-1(1000) e-1(17) e-1(10) e-1(1000) e-2(10) e-2(10) e-2(991) e-2(10) e-2(10) 787 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 26 / 30
27 Low-rank density matrix recovery Table: Performance for the low-rank density matrix recovery problems r n sr(%) One-stage Two-stage Final result relerr(rank) time relerr(rank) time k relerr(rank) time e-1(31) e-1(3) e-1(3) e-1(31) e-1(3) e-1(3) e-1(31) e-1(3) e-2(3) e-1(41) e-2(3) e-2(3) e-1(41) e-1(3) e-2(3) e-1(40) e-2(3) e-2(3) e-1(42) e-1(5) e-1(5) e-1(43) e-1(5) e-1(5) e-1(43) e-1(17) e-1(5) 1014 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 27 / 30
28 Low-rank covariance matrix recovery Table: Performance for the low-rank covariance matrix recovery with n = 1000 r nfixed sr(%) One-stage Two-stage Final result relerr(rank) time relerr(rank) time k relerr(rank) time (200, 0) e-1(58) e-1(10) e-1(10) 578 (0, 200) e-1(59) e-1(10) e-1(10) 630 (200, 200) e-1(57) e-1(10) e-1(10) 580 (200, 0) e-1(52) e-2(10) e-2(10) 444 (0, 200) e-1(49) e-1(10) e-2(10) 329 (200, 200) e-1(53) e-2(10) e-2(10) 463 (200, 0) e-1(86) e-1(20) e-1(20) 346 (0, 200) e-1(88) e-1(20) e-1(20) 322 (200, 200) e-1(87) e-1(20) e-1(20) 415 (200, 0) e-1(79) e-1(20) e-1(20) 270 (0, 200) e-1(78) e-1(20) e-1(20) 261 (200, 200) e-1(80) e-1(20) e-1(20) 294 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 28 / 30
29 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 29 / 30
30 Conclusion We proposed a multi-stage convex relaxation approach for the rank minimization problem by solving the exact penalty problem of its equivalent MPSDEC in an alternating way. Under a weaker condition than the RIP, we established a tighter error bound for the optimal solution of each subproblem to the global optimal solution of the original problem, and provided a quantitative analysis for the decrease of the error bound. The two-stage convex relaxation can improve the error bound given by the trace norm relaxation method at least 30% for those not nice problems. The general case can be done by using the theory on spectral operators of matrices. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 30 / 30
A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery
A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery Shujun Bi, Shaohua Pan and Defeng Sun March 1, 2017 Abstract This paper concerns with a noisy structured low-rank matrix
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationMatrix Rank Minimization with Applications
Matrix Rank Minimization with Applications Maryam Fazel Haitham Hindi Stephen Boyd Information Systems Lab Electrical Engineering Department Stanford University 8/2001 ACC 01 Outline Rank Minimization
More informationA Rank-Corrected Procedure for Matrix Completion with Fixed Basis Coefficients
A Rank-Corrected Procedure for Matrix Completion with Fixed Basis Coefficients Weimin Miao, Shaohua Pan and Defeng Sun RMI Working Paper No. 2/0 Submitted: 3 October, 202 Abstract In this paper, we address
More informationExact Low-rank Matrix Recovery via Nonconvex M p -Minimization
Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Lingchen Kong and Naihua Xiu Department of Applied Mathematics, Beijing Jiaotong University, Beijing, 100044, People s Republic of China E-mail:
More informationPU Learning for Matrix Completion
Cho-Jui Hsieh Dept of Computer Science UT Austin ICML 2015 Joint work with N. Natarajan and I. S. Dhillon Matrix Completion Example: movie recommendation Given a set Ω and the values M Ω, how to predict
More informationAn iterative hard thresholding estimator for low rank matrix recovery
An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical
More informationRank minimization via the γ 2 norm
Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises
More informationSparse and Low Rank Recovery via Null Space Properties
Sparse and Low Rank Recovery via Null Space Properties Holger Rauhut Lehrstuhl C für Mathematik (Analysis), RWTH Aachen Convexity, probability and discrete structures, a geometric viewpoint Marne-la-Vallée,
More informationRecovery of Simultaneously Structured Models using Convex Optimization
Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)
More informationAn Introduction to Correlation Stress Testing
An Introduction to Correlation Stress Testing Defeng Sun Department of Mathematics and Risk Management Institute National University of Singapore This is based on a joint work with GAO Yan at NUS March
More informationRecovering any low-rank matrix, provably
Recovering any low-rank matrix, provably Rachel Ward University of Texas at Austin October, 2014 Joint work with Yudong Chen (U.C. Berkeley), Srinadh Bhojanapalli and Sujay Sanghavi (U.T. Austin) Matrix
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationRecent Developments in Compressed Sensing
Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline
More informationSemidefinite Programming Basics and Applications
Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent
More informationAssignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition. Name:
Assignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition Due date: Friday, May 4, 2018 (1:35pm) Name: Section Number Assignment #10: Diagonalization
More informationarxiv: v1 [math.st] 10 Sep 2015
Fast low-rank estimation by projected gradient descent: General statistical and algorithmic guarantees Department of Statistics Yudong Chen Martin J. Wainwright, Department of Electrical Engineering and
More informationMATRIX COMPLETION MODELS WITH FIXED BASIS COEFFICIENTS AND RANK REGULARIZED PROBLEMS WITH HARD CONSTRAINTS MIAO WEIMIN. (M.Sc., UCAS; B.Sc.
MATRIX COMPLETION MODELS WITH FIXED BASIS COEFFICIENTS AND RANK REGULARIZED PROBLEMS WITH HARD CONSTRAINTS MIAO WEIMIN (M.Sc., UCAS; B.Sc., PKU) A THESIS SUBMITTED FOR THE DEGREE OF DOCTOR OF PHILOSOPHY
More informationA Unified Theorem on SDP Rank Reduction. yyye
SDP Rank Reduction Yinyu Ye, EURO XXII 1 A Unified Theorem on SDP Rank Reduction Yinyu Ye Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering Stanford
More informationLecture 8 : Eigenvalues and Eigenvectors
CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with
More informationAn Approximation Algorithm for Approximation Rank
An Approximation Algorithm for Approximation Rank Troy Lee Columbia University Adi Shraibman Weizmann Institute Conventions Identify a communication function f : X Y { 1, +1} with the associated X-by-Y
More informationSTRUCTURED LOW RANK MATRIX OPTIMIZATION PROBLEMS: A PENALTY APPROACH GAO YAN. (B.Sc., ECNU)
STRUCTURED LOW RANK MATRIX OPTIMIZATION PROBLEMS: A PENALTY APPROACH GAO YAN (B.Sc., ECNU) A THESIS SUBMITTED FOR THE DEGREE OF DOCTOR OF PHILOSOPHY DEPARTMENT OF MATHEMATICS NATIONAL UNIVERSITY OF SINGAPORE
More informationHigh-dimensional Statistics
High-dimensional Statistics Pradeep Ravikumar UT Austin Outline 1. High Dimensional Data : Large p, small n 2. Sparsity 3. Group Sparsity 4. Low Rank 1 Curse of Dimensionality Statistical Learning: Given
More informationRank Constrained Matrix Optimization Problems
Rank Constrained Matrix Optimization Problems Defeng Sun Department of Mathematics and Risk Management Institute National University of Singapore This talk is based on a joint work with Yan Gao at NUS
More informationLow-Rank Matrix Recovery
ELE 538B: Mathematics of High-Dimensional Data Low-Rank Matrix Recovery Yuxin Chen Princeton University, Fall 2018 Outline Motivation Problem setup Nuclear norm minimization RIP and low-rank matrix recovery
More informationAn Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion
An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion Robert M. Freund, MIT joint with Paul Grigas (UC Berkeley) and Rahul Mazumder (MIT) CDC, December 2016 1 Outline of Topics
More informationCSC 576: Variants of Sparse Learning
CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in
More informationAn Optimization-based Approach to Decentralized Assignability
2016 American Control Conference (ACC) Boston Marriott Copley Place July 6-8, 2016 Boston, MA, USA An Optimization-based Approach to Decentralized Assignability Alborz Alavian and Michael Rotkowitz Abstract
More informationAcyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs
Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,
More information1 Sparsity and l 1 relaxation
6.883 Learning with Combinatorial Structure Note for Lecture 2 Author: Chiyuan Zhang Sparsity and l relaxation Last time we talked about sparsity and characterized when an l relaxation could recover the
More informationA direct formulation for sparse PCA using semidefinite programming
A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationSpectral k-support Norm Regularization
Spectral k-support Norm Regularization Andrew McDonald Department of Computer Science, UCL (Joint work with Massimiliano Pontil and Dimitris Stamos) 25 March, 2015 1 / 19 Problem: Matrix Completion Goal:
More informationQuadratic reformulation techniques for 0-1 quadratic programs
OSE SEMINAR 2014 Quadratic reformulation techniques for 0-1 quadratic programs Ray Pörn CENTER OF EXCELLENCE IN OPTIMIZATION AND SYSTEMS ENGINEERING ÅBO AKADEMI UNIVERSITY ÅBO NOVEMBER 14th 2014 2 Structure
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu
More informationSparse Sensing in Colocated MIMO Radar: A Matrix Completion Approach
Sparse Sensing in Colocated MIMO Radar: A Matrix Completion Approach Athina P. Petropulu Department of Electrical and Computer Engineering Rutgers, the State University of New Jersey Acknowledgments Shunqiao
More informationAdaptive estimation of the copula correlation matrix for semiparametric elliptical copulas
Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas Department of Mathematics Department of Statistical Science Cornell University London, January 7, 2016 Joint work
More informationHigh-dimensional Statistical Models
High-dimensional Statistical Models Pradeep Ravikumar UT Austin MLSS 2014 1 Curse of Dimensionality Statistical Learning: Given n observations from p(x; θ ), where θ R p, recover signal/parameter θ. For
More informationSolving Corrupted Quadratic Equations, Provably
Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin
More informationSome tensor decomposition methods for machine learning
Some tensor decomposition methods for machine learning Massimiliano Pontil Istituto Italiano di Tecnologia and University College London 16 August 2016 1 / 36 Outline Problem and motivation Tucker decomposition
More informationCommon-Knowledge / Cheat Sheet
CSE 521: Design and Analysis of Algorithms I Fall 2018 Common-Knowledge / Cheat Sheet 1 Randomized Algorithm Expectation: For a random variable X with domain, the discrete set S, E [X] = s S P [X = s]
More informationA note on rank constrained solutions to linear matrix equations
A note on rank constrained solutions to linear matrix equations Shravan Mohan September 5, 218 Abstract This preliminary note presents a heuristic for determining rank constrained solutions to linear matrix
More informationSelected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A.
. Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A. Nemirovski Arkadi.Nemirovski@isye.gatech.edu Linear Optimization Problem,
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationLow-Rank Factorization Models for Matrix Completion and Matrix Separation
for Matrix Completion and Matrix Separation Joint work with Wotao Yin, Yin Zhang and Shen Yuan IPAM, UCLA Oct. 5, 2010 Low rank minimization problems Matrix completion: find a low-rank matrix W R m n so
More informationExact penalty decomposition method for zero-norm minimization based on MPEC formulation 1
Exact penalty decomposition method for zero-norm minimization based on MPEC formulation Shujun Bi, Xiaolan Liu and Shaohua Pan November, 2 (First revised July 5, 22) (Second revised March 2, 23) (Final
More information(Part 1) High-dimensional statistics May / 41
Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2
More informationRank-one Generated Spectral Cones Defined by Two Homogeneous Linear Matrix Inequalities
Rank-one Generated Spectral Cones Defined by Two Homogeneous Linear Matrix Inequalities C.J. Argue Joint work with Fatma Kılınç-Karzan October 22, 2017 INFORMS Annual Meeting Houston, Texas C. Argue, F.
More informationConvex relaxation for Combinatorial Penalties
Convex relaxation for Combinatorial Penalties Guillaume Obozinski Equipe Imagine Laboratoire d Informatique Gaspard Monge Ecole des Ponts - ParisTech Joint work with Francis Bach Fête Parisienne in Computation,
More informationCollaborative Filtering Matrix Completion Alternating Least Squares
Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade May 19, 2016
More informationUniversal low-rank matrix recovery from Pauli measurements
Universal low-rank matrix recovery from Pauli measurements Yi-Kai Liu Applied and Computational Mathematics Division National Institute of Standards and Technology Gaithersburg, MD, USA yi-kai.liu@nist.gov
More information8.1 Concentration inequality for Gaussian random matrix (cont d)
MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration
More informationPENNON A Generalized Augmented Lagrangian Method for Nonconvex NLP and SDP p.1/22
PENNON A Generalized Augmented Lagrangian Method for Nonconvex NLP and SDP Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University
More informationReconstruction from Anisotropic Random Measurements
Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013
More informationFast and Robust Phase Retrieval
Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National
More informationLow Rank Matrix Completion Formulation and Algorithm
1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9
More informationTHE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR
THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional
More informationforms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms
Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.
More informationAdaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise
Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Minru Bai(x T) College of Mathematics and Econometrics Hunan University Joint work with Xiongjun Zhang, Qianqian Shao June 30,
More informationComputational and Statistical Aspects of Statistical Machine Learning. John Lafferty Department of Statistics Retreat Gleacher Center
Computational and Statistical Aspects of Statistical Machine Learning John Lafferty Department of Statistics Retreat Gleacher Center Outline Modern nonparametric inference for high dimensional data Nonparametric
More informationSection 3.9. Matrix Norm
3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix
More informationWe describe the generalization of Hazan s algorithm for symmetric programming
ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,
More informationPHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN
PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationConvex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013
Convex Optimization (EE227A: UC Berkeley) Lecture 6 (Conic optimization) 07 Feb, 2013 Suvrit Sra Organizational Info Quiz coming up on 19th Feb. Project teams by 19th Feb Good if you can mix your research
More informationSmall zeros of hermitian forms over quaternion algebras. Lenny Fukshansky Claremont McKenna College & IHES (joint work with W. K.
Small zeros of hermitian forms over quaternion algebras Lenny Fukshansky Claremont McKenna College & IHES (joint work with W. K. Chan) Institut de Mathématiques de Jussieu October 21, 2010 1 Cassels Theorem
More informationBlock Coordinate Descent for Regularized Multi-convex Optimization
Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline
More information(k, q)-trace norm for sparse matrix factorization
(k, q)-trace norm for sparse matrix factorization Emile Richard Department of Electrical Engineering Stanford University emileric@stanford.edu Guillaume Obozinski Imagine Ecole des Ponts ParisTech guillaume.obozinski@imagine.enpc.fr
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57
More informationNew Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit
New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence
More information1-Bit Matrix Completion
1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible
More informationNonlinear Programming Models
Nonlinear Programming Models Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Nonlinear Programming Models p. Introduction Nonlinear Programming Models p. NLP problems minf(x) x S R n Standard form:
More informationSparse PCA with applications in finance
Sparse PCA with applications in finance A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon 1 Introduction
More informationTwo Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery
Two Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery Ming-Jun Lai, Song Li, Louis Y. Liu and Huimin Wang August 14, 2012 Abstract We shall provide a sufficient condition to
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More informationLINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM
LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator
More informationCompression, Matrix Range and Completely Positive Map
Compression, Matrix Range and Completely Positive Map Iowa State University Iowa-Nebraska Functional Analysis Seminar November 5, 2016 Definitions and notations H, K : Hilbert space. If dim H = n
More informationApplications of Robust Optimization in Signal Processing: Beamforming and Power Control Fall 2012
Applications of Robust Optimization in Signal Processing: Beamforg and Power Control Fall 2012 Instructor: Farid Alizadeh Scribe: Shunqiao Sun 12/09/2012 1 Overview In this presentation, we study the applications
More informationMaximum Margin Matrix Factorization
Maximum Margin Matrix Factorization Nati Srebro Toyota Technological Institute Chicago Joint work with Noga Alon Tel-Aviv Yonatan Amit Hebrew U Alex d Aspremont Princeton Michael Fink Hebrew U Tommi Jaakkola
More informationLecture 3. Random Fourier measurements
Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationEE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term
EE 150 - Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko JPL Third Term 2011-2012 Due on Thursday May 3 in class. Homework Set #4 1. (10 points) (Adapted
More information10-725/36-725: Convex Optimization Prerequisite Topics
10-725/36-725: Convex Optimization Prerequisite Topics February 3, 2015 This is meant to be a brief, informal refresher of some topics that will form building blocks in this course. The content of the
More informationRobust Tensor Completion and its Applications. Michael K. Ng Department of Mathematics Hong Kong Baptist University
Robust Tensor Completion and its Applications Michael K. Ng Department of Mathematics Hong Kong Baptist University mng@math.hkbu.edu.hk Example Low-dimensional Structure Data in many real applications
More informationAhmed Mohammed. Harnack Inequality for Non-divergence Structure Semi-linear Elliptic Equations
Harnack Inequality for Non-divergence Structure Semi-linear Elliptic Equations International Conference on PDE, Complex Analysis, and Related Topics Miami, Florida January 4-7, 2016 An Outline 1 The Krylov-Safonov
More informationLecture 4: January 26
10-725/36-725: Conve Optimization Spring 2015 Lecturer: Javier Pena Lecture 4: January 26 Scribes: Vipul Singh, Shinjini Kundu, Chia-Yin Tsai Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationarxiv: v3 [stat.me] 8 Jun 2018
Between hard and soft thresholding: optimal iterative thresholding algorithms Haoyang Liu and Rina Foygel Barber arxiv:804.0884v3 [stat.me] 8 Jun 08 June, 08 Abstract Iterative thresholding algorithms
More informationarxiv: v3 [math.oc] 19 Oct 2017
Gradient descent with nonconvex constraints: local concavity determines convergence Rina Foygel Barber and Wooseok Ha arxiv:703.07755v3 [math.oc] 9 Oct 207 0.7.7 Abstract Many problems in high-dimensional
More informationCanonical Problem Forms. Ryan Tibshirani Convex Optimization
Canonical Problem Forms Ryan Tibshirani Convex Optimization 10-725 Last time: optimization basics Optimization terology (e.g., criterion, constraints, feasible points, solutions) Properties and first-order
More informationRecall : Eigenvalues and Eigenvectors
Recall : Eigenvalues and Eigenvectors Let A be an n n matrix. If a nonzero vector x in R n satisfies Ax λx for a scalar λ, then : The scalar λ is called an eigenvalue of A. The vector x is called an eigenvector
More informationSupplementary Materials for Riemannian Pursuit for Big Matrix Recovery
Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery Mingkui Tan, School of omputer Science, The University of Adelaide, Australia Ivor W. Tsang, IS, University of Technology Sydney,
More informationConstructing Explicit RIP Matrices and the Square-Root Bottleneck
Constructing Explicit RIP Matrices and the Square-Root Bottleneck Ryan Cinoman July 18, 2018 Ryan Cinoman Constructing Explicit RIP Matrices July 18, 2018 1 / 36 Outline 1 Introduction 2 Restricted Isometry
More informationU.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016
U.C. Berkeley CS294: Spectral Methods and Expanders Handout Luca Trevisan February 29, 206 Lecture : ARV In which we introduce semi-definite programming and a semi-definite programming relaxation of sparsest
More informationELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications
ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March
More informationRobust Principal Component Analysis
ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M
More informationHigh Dimensional Inverse Covariate Matrix Estimation via Linear Programming
High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω
More informationTHE NONCOMMUTATIVE TORUS
THE NONCOMMUTATIVE TORUS The noncommutative torus as a twisted convolution An ordinary two-torus T 2 with coordinate functions given by where x 1, x 2 [0, 1]. U 1 = e 2πix 1, U 2 = e 2πix 2, (1) By Fourier
More informationSparse Recovery via Partial Regularization: Models, Theory and Algorithms
Sparse Recovery via Partial Regularization: Models, Theory and Algorithms Zhaosong Lu and Xiaorui Li Department of Mathematics, Simon Fraser University, Canada {zhaosong,xla97}@sfu.ca November 23, 205
More informationPartial regularity for fully nonlinear PDE
Partial regularity for fully nonlinear PDE Luis Silvestre University of Chicago Joint work with Scott Armstrong and Charles Smart Outline Introduction Intro Review of fully nonlinear elliptic PDE Our result
More information