Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery

Size: px
Start display at page:

Download "Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery"

Transcription

1 Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun Bi and Shaohua Pan) May 22, 2014 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 1 / 30

2 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 2 / 30

3 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 3 / 30

4 Matrix recovery Given some measurements for a (structured) matrix X H n + (Hermitian positive semi-definite), how to recover this unknown matrix? b k = Θ k, X + ξ k, k = 1, 2,..., m. b k : observed data (with / without noise) ξ k : additive noise Possible? Yes, with the low-rank structure. Entries = basis coefficients: {e i e T j 1 i n 1, 1 j n 2 }. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 4 / 30

5 Correlation matrix recovery A low-rank correlation matrix X S n +: symmetric positive semidefinite and diag(x) = e. Observations (with / without noise): b k = Θ k, X + ξ k, k = 1,..., m. How to recover the unknown correlation matrix X? All diagonal entries are fixed. Some off-diagonal entries may also be fixed, [e.g., the correlations among pegged currencies]. Entries basis coefficients { e i e T i 1 i n } { 1 2 (e i e T j +e j e T i ) 1 i<j n }. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 5 / 30

6 Density matrix recovery A low-rank density matrix X H n + with n = 2 l : Hermitian positive semidefinite and Tr(X) = 1. Observations Pauli measurements: Θ k, X = Re(Tr(Θ k X)), k = 1,..., m, where Θ k Pauli basis { σ s1 σ sl (s 1,, s l ) {0, 1, 2, 3} l} with σ 0 = ( ) 1 0, σ = How to recover the unknown density matrix X? ( ) ( ) ( ) , σ = 1 0, σ =. 0 1 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 6 / 30

7 Observation model In this talk, we consider the problem of low-rank structured PSD matrix recovery with noise. The observation model takes the following form b k = Θ k, X + ξ k, k = 1,..., m. Define A(X) := ( Θ 1, X,, Θ m, X ) T R m. Then, we have that b = A(X) + ξ. Let X H n + be the unknown low-rank structured true matrix. There exists a constant δ 0, A(X) b δ. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 7 / 30

8 The rank minimization estimator The rank minimization estimator: X arg min Ω H n + { rank(x) : A(X) b δ where Ω is a closed convex set characterizing the structure of X. This is the best possible estimator. But it is very challenging to get the global solution due to the nonconvexity and discontinuity of the rank function. } / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 8 / 30

9 The nuclear norm technique The nuclear norm technique has gained success in many applications: n rank(x) = X := σ i (X), where σ 1 (X) σ n (X) denote all the singular values of X. Nuclear norm convex envelope of the rank function over the unit ball of the spectral norm. y i=1 x /a 1 rank x a 0 a x / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 9 / 30

10 Nuclear norm estimator Nuclear norm estimator: X arg min Ω H n + { } X : A(X) b δ Extra biased (for large singular values). It may fail when certain rows and / or columns are heavily sampled (Salakhutdinov and Srebro, 2010). It may not be rank consistent for general sampling even for the low-dimensional case (Bach, 2008). For correlation and density matrices, X = constant = the nuclear norm technique s power is limited. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 10 / 30

11 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 11 / 30

12 Equivalent MPSDEC Let C(J) be the family of closed convex functions φ : J R with J [0, 1] and φ(0) = 0, 0 < t := arg min 0 t 1 φ(t) and φ (0) < +. Rank minimization estimator is equivalent to the following mathematical program with the PSD equilibrium constraint (MPSDEC) min I, Φ(W) X Ω,W Hn s.t. A(X) b δ (1) W, X = 0, X 0, 0 W I Φ the Löwner operator associated to φ C(J) The bilinear constraint W, X = 0 is the trouble maker. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 12 / 30

13 Exact penalty for MPSDEC To deal with W, X = 0 efficiently, we consider the penalized version min X Ω,W Hn I, Φ(W) + ρ W, X s.t. A(X) b δ (2) X 0, 0 W I Theorem The cost function of (2) has a desired structure though it is nonconvex The penalty problem (2) is exact in the following sense There exists ρ > 0 such that when ρ > ρ, the set of global optimal solutions of the penalty problem (2) coincides with that of the MPSDEC. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 13 / 30

14 Multi-stage convex relaxation approach (S0) Select a function φ C(J). Let W 0 := I and set k := 1. Choose ρ 0 > 0. (S1) Solve the following weighted trace-norm minimization problem X k arg min X Ω H+ n { W k 1, X : A(X) b δ }. (S2) By the information of X k select a suitable ρ k = δ k ρ k 1 with δ k 1. (S3) Solve the following minimization problem W k arg min 0 W I (S4) Set k := k + 1, and go to (S.1). { I, Φ(W) + ρ k W, X k }. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 14 / 30

15 Multi-stage convex relaxation (Examples) The convex relaxation approach is solving (2) in an alternating way. The computational cost is the solution of the weighted trace-norm minimization problem since if X k = U k Diag(λ(X k ))(U k ) T we have W k = U k Diag(w k 1,..., wk n)(u k ) T. 1 φ 1 (t) = t w k i = { 0 if λi (X k ) 1 ρ k 1 otherwise 2 φ 2 (t) = a2 1 4 t 2 at w k i = min( 2 a 2 1 max(a ρ kλ i (X k ), 0), 1). 3 φ 3 (t) = t + 1 t+ɛ 1 ɛ w k 1 i = max( ɛ, 0). ρk λ i (X k )+1 4 φ 4 (t) = t ln(t + ɛ) + ln(ɛ) w k 1 i = max( ρ k λ i ɛ, 0). (X k )+1 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 15 / 30

16 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 16 / 30

17 Notation and assumptions Recall that rank(x) = r. Assume that X = [U 1, U 2 ]Diag(λ(X))[U 1, U 2 ] T, where U 1 O n r and U 2 O n (n r). For any X H n, define P T (X) = U 1 U T 1 X + XU 1 U T 1 U 1 U T 1 XU 1 U T 1. Let Q = A A. Define the restricted eigenvalues of Q by (Tong Zhang, 2010) X, Q(X) X, Q(X) ϑ + (k) := sup 0<rank(X) k X 2 F, ϑ (k) := inf 0<rank(X) k X 2 F Assumption: There exist a c [0, 1) and an integer s (1 s n 2r 2 ) s.t. ϑ + (s) ϑ (2r + 2s) 1 + 2c2 s. r / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 17 / 30

18 General error bound Theorem Let γ k 1 := P T(W k 1 ) λ min ((U 2 ) T W k 1 U 2 ). If 0 γ k 1 1 c, then Xk X F Ξ(γ k 1 ) with Ξ(t) := 1 + rt2 2s 2 ϑ + (2r + s)δ (1 ct) ϑ (2r + s) Ξ(1) the error bound for the nuclear norm estimator. 5 2 ) 3 2 (1 δ 3r) If c < 1 1 δ 3r(1+ Ξ(1) = for t 0., then we have the following inequality 6 ϑ+ (3r)δ (1 c) ϑ (3r) δ3r δ (1 c) (1 δ 3r ) < 3δ 1 + δ3r 1 δ 3r (1 + The last term is the bound given by Mohan and Fazel (2010). 5 2 ). / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 18 / 30

19 Error bounds in the first two stages Let ψ be the conjugate function of ψ(t) := { φ(t) if t [0, 1], + otherwise. Define a 0 := (ψ ) +(0), a 1 := (ψ ) ( ρ1 λ r+1 (X 1 ) ) and b 1 := (ψ ) +( ρ1 λ r (X 1 ) ). Theorem Suppose that λ r (X) = β X 1 X F for some β > 2. If the parameter ρ 1 is chosen such that ρ 1 ( 0, φ + (0) ) b 1 + a 0 β ln λ r+1 (X 1 ) and 2 β 2 < 1, then a 1 a 1 2 ln β β 2 X 2 X F Ξ(γ 1 ) Ξ b 1 + a 0 a 1 a 1 β 2 ln β 2 β 2 ln β 2 < Ξ(1). (3) / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 19 / 30

20 Reduction ratio given by different φ Table: The ratio of the error bounds given by different φ for the first two stages φ 1 ( φ 2 ( φ 3 ( ( φ 4 ( φ 4 ρ 1 β ratio ) Ξ(γ ) λ r+1 (X 1 ) Ξ(1) ) Ξ(γ ) λ r+1 (X 1 ) Ξ(1) ) Ξ(γ ) λ r+1 (X 1 ) Ξ(1) ) Ξ(γ ) λ r+1 (X 1 ) Ξ(1) ) Ξ(γ ) Ξ(1) λ r(x 1 ), 1 2 λ r(x 1 ), 1 3 λ r(x 1 ), λ r(x 1 ), λ r(x 1 ), 0.8 λ r+1 (X 1 ) / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 20 / 30

21 Guaranteed error bound reduction (Cont.) Let b k := (ψ ) +[ ρk (λ r (X ) Ξ( γ k 1 )) ], ã k := (ψ ) ( ρ k Ξ( γ k 1 )), γ k := b k + a 0 2 ln ã k ãk 2 ln λ r(x) λ r(x) 2Ξ( γ k 1 ) λ r(x) λ r(x) 2Ξ( γ k 1 ) for k 1 and γ 0 = 1. Theorem Suppose that λ r (X) = βξ(1) for some β > 2. If the parameter ρ 1 is chosen such that ρ 1 ( 0, φ + (0) ) b 1 + a 0 β ln Ξ(1) and 2 β 2 < 1, and the parameters σ k are chosen ã 1 ã1 2 ln β β 2 such that σ k [ 1, Ξ( γ k 2) Ξ( γ k 1 )], then for all k 2 we have γk 1 γ k 1 < γ k 2 and X k X F Ξ(γ k 1 ) Ξ( γ k 1 ) < Ξ( γ k 2 ). (4) / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 21 / 30

22 Reduction ratio of the error bounds Table: The ratio of the error bounds obtained with φ 2 for the first five stages ( φ Ξ(1), 1.2 ) Ξ(1) ρ 1 ratio Ξ(γ 1 )/Ξ(1) Ξ(γ 2 )/Ξ(1) Ξ(γ 3 )/Ξ(1) Ξ(γ 4 )/Ξ(1) Ξ(γ 5 )/Ξ(1) / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 22 / 30

23 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 23 / 30

24 The performance with a high sample ratio φ 1 φ 2 with a=2.1 φ 3 with ε=0.4 φ 4 with ε= φ 1 φ 2 with a=2.1 φ 3 with ε=0.4 φ 4 with ε= Relative error (%) Computing time (in seconds) (a) The number of stages (b) The number of stages Figure: The performance of multi-stage convex relaxation with p = / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 24 / 30

25 The performance with a low sample ratio φ 1 φ 2 with a=2.1 φ with ε=0.4 3 φ with ε= φ 1 φ 2 with a=2.1 φ with ε=0.4 3 φ with ε= Relative error (%) Computing time (in seconds) (a) The number of stages (b) The number of stages Figure: The performance of multi-stage convex relaxation with p = 3980 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 25 / 30

26 Low-rank correlation matrix recovery Table: Performance for the correlation matrix recovery problems with n = 1000 r 5 10 off diag sr(%) One-stage Two-stage Final result relerr(rank) time relerr(rank) time k relerr(rank) time e-1(1000) e-1(19) e-1(5) e-1(998) e-1(5) e-1(5) e-2(999) e-2(5) e-2(5) e-1(991) e-1(18) e-1(5) e-1(977) e-1(5) e-1(5) e-1(703) e-2(6) e-2(5) e-1(1000) e-1(21) e-1(10) e-1(1000) e-1(10) e-1(10) e-2(1000) e-2(10) e-2(10) e-1(1000) e-1(17) e-1(10) e-1(1000) e-2(10) e-2(10) e-2(991) e-2(10) e-2(10) 787 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 26 / 30

27 Low-rank density matrix recovery Table: Performance for the low-rank density matrix recovery problems r n sr(%) One-stage Two-stage Final result relerr(rank) time relerr(rank) time k relerr(rank) time e-1(31) e-1(3) e-1(3) e-1(31) e-1(3) e-1(3) e-1(31) e-1(3) e-2(3) e-1(41) e-2(3) e-2(3) e-1(41) e-1(3) e-2(3) e-1(40) e-2(3) e-2(3) e-1(42) e-1(5) e-1(5) e-1(43) e-1(5) e-1(5) e-1(43) e-1(17) e-1(5) 1014 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 27 / 30

28 Low-rank covariance matrix recovery Table: Performance for the low-rank covariance matrix recovery with n = 1000 r nfixed sr(%) One-stage Two-stage Final result relerr(rank) time relerr(rank) time k relerr(rank) time (200, 0) e-1(58) e-1(10) e-1(10) 578 (0, 200) e-1(59) e-1(10) e-1(10) 630 (200, 200) e-1(57) e-1(10) e-1(10) 580 (200, 0) e-1(52) e-2(10) e-2(10) 444 (0, 200) e-1(49) e-1(10) e-2(10) 329 (200, 200) e-1(53) e-2(10) e-2(10) 463 (200, 0) e-1(86) e-1(20) e-1(20) 346 (0, 200) e-1(88) e-1(20) e-1(20) 322 (200, 200) e-1(87) e-1(20) e-1(20) 415 (200, 0) e-1(79) e-1(20) e-1(20) 270 (0, 200) e-1(78) e-1(20) e-1(20) 261 (200, 200) e-1(80) e-1(20) e-1(20) 294 / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 28 / 30

29 Outline 1 Problem formulation 2 Multi-stage convex relaxation approach Exact penalty for the equivalent MPSDEC A unified framework for multi-stage convex relaxation 3 The power of multi-stage convex relaxation approach General error bound Error bound comparison for the first two stages Error bound comparison for the first k stages 4 Numerical results 5 Conclusion / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 29 / 30

30 Conclusion We proposed a multi-stage convex relaxation approach for the rank minimization problem by solving the exact penalty problem of its equivalent MPSDEC in an alternating way. Under a weaker condition than the RIP, we established a tighter error bound for the optimal solution of each subproblem to the global optimal solution of the original problem, and provided a quantitative analysis for the decrease of the error bound. The two-stage convex relaxation can improve the error bound given by the trace norm relaxation method at least 30% for those not nice problems. The general case can be done by using the theory on spectral operators of matrices. / Defeng Sun /NUS Adaptive nuclear semi-norm penalization OP14, San Diego 30 / 30

A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery

A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery Shujun Bi, Shaohua Pan and Defeng Sun March 1, 2017 Abstract This paper concerns with a noisy structured low-rank matrix

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Matrix Rank Minimization with Applications

Matrix Rank Minimization with Applications Matrix Rank Minimization with Applications Maryam Fazel Haitham Hindi Stephen Boyd Information Systems Lab Electrical Engineering Department Stanford University 8/2001 ACC 01 Outline Rank Minimization

More information

A Rank-Corrected Procedure for Matrix Completion with Fixed Basis Coefficients

A Rank-Corrected Procedure for Matrix Completion with Fixed Basis Coefficients A Rank-Corrected Procedure for Matrix Completion with Fixed Basis Coefficients Weimin Miao, Shaohua Pan and Defeng Sun RMI Working Paper No. 2/0 Submitted: 3 October, 202 Abstract In this paper, we address

More information

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Lingchen Kong and Naihua Xiu Department of Applied Mathematics, Beijing Jiaotong University, Beijing, 100044, People s Republic of China E-mail:

More information

PU Learning for Matrix Completion

PU Learning for Matrix Completion Cho-Jui Hsieh Dept of Computer Science UT Austin ICML 2015 Joint work with N. Natarajan and I. S. Dhillon Matrix Completion Example: movie recommendation Given a set Ω and the values M Ω, how to predict

More information

An iterative hard thresholding estimator for low rank matrix recovery

An iterative hard thresholding estimator for low rank matrix recovery An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical

More information

Rank minimization via the γ 2 norm

Rank minimization via the γ 2 norm Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises

More information

Sparse and Low Rank Recovery via Null Space Properties

Sparse and Low Rank Recovery via Null Space Properties Sparse and Low Rank Recovery via Null Space Properties Holger Rauhut Lehrstuhl C für Mathematik (Analysis), RWTH Aachen Convexity, probability and discrete structures, a geometric viewpoint Marne-la-Vallée,

More information

Recovery of Simultaneously Structured Models using Convex Optimization

Recovery of Simultaneously Structured Models using Convex Optimization Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)

More information

An Introduction to Correlation Stress Testing

An Introduction to Correlation Stress Testing An Introduction to Correlation Stress Testing Defeng Sun Department of Mathematics and Risk Management Institute National University of Singapore This is based on a joint work with GAO Yan at NUS March

More information

Recovering any low-rank matrix, provably

Recovering any low-rank matrix, provably Recovering any low-rank matrix, provably Rachel Ward University of Texas at Austin October, 2014 Joint work with Yudong Chen (U.C. Berkeley), Srinadh Bhojanapalli and Sujay Sanghavi (U.T. Austin) Matrix

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Recent Developments in Compressed Sensing

Recent Developments in Compressed Sensing Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline

More information

Semidefinite Programming Basics and Applications

Semidefinite Programming Basics and Applications Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent

More information

Assignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition. Name:

Assignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition. Name: Assignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition Due date: Friday, May 4, 2018 (1:35pm) Name: Section Number Assignment #10: Diagonalization

More information

arxiv: v1 [math.st] 10 Sep 2015

arxiv: v1 [math.st] 10 Sep 2015 Fast low-rank estimation by projected gradient descent: General statistical and algorithmic guarantees Department of Statistics Yudong Chen Martin J. Wainwright, Department of Electrical Engineering and

More information

MATRIX COMPLETION MODELS WITH FIXED BASIS COEFFICIENTS AND RANK REGULARIZED PROBLEMS WITH HARD CONSTRAINTS MIAO WEIMIN. (M.Sc., UCAS; B.Sc.

MATRIX COMPLETION MODELS WITH FIXED BASIS COEFFICIENTS AND RANK REGULARIZED PROBLEMS WITH HARD CONSTRAINTS MIAO WEIMIN. (M.Sc., UCAS; B.Sc. MATRIX COMPLETION MODELS WITH FIXED BASIS COEFFICIENTS AND RANK REGULARIZED PROBLEMS WITH HARD CONSTRAINTS MIAO WEIMIN (M.Sc., UCAS; B.Sc., PKU) A THESIS SUBMITTED FOR THE DEGREE OF DOCTOR OF PHILOSOPHY

More information

A Unified Theorem on SDP Rank Reduction. yyye

A Unified Theorem on SDP Rank Reduction.   yyye SDP Rank Reduction Yinyu Ye, EURO XXII 1 A Unified Theorem on SDP Rank Reduction Yinyu Ye Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering Stanford

More information

Lecture 8 : Eigenvalues and Eigenvectors

Lecture 8 : Eigenvalues and Eigenvectors CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with

More information

An Approximation Algorithm for Approximation Rank

An Approximation Algorithm for Approximation Rank An Approximation Algorithm for Approximation Rank Troy Lee Columbia University Adi Shraibman Weizmann Institute Conventions Identify a communication function f : X Y { 1, +1} with the associated X-by-Y

More information

STRUCTURED LOW RANK MATRIX OPTIMIZATION PROBLEMS: A PENALTY APPROACH GAO YAN. (B.Sc., ECNU)

STRUCTURED LOW RANK MATRIX OPTIMIZATION PROBLEMS: A PENALTY APPROACH GAO YAN. (B.Sc., ECNU) STRUCTURED LOW RANK MATRIX OPTIMIZATION PROBLEMS: A PENALTY APPROACH GAO YAN (B.Sc., ECNU) A THESIS SUBMITTED FOR THE DEGREE OF DOCTOR OF PHILOSOPHY DEPARTMENT OF MATHEMATICS NATIONAL UNIVERSITY OF SINGAPORE

More information

High-dimensional Statistics

High-dimensional Statistics High-dimensional Statistics Pradeep Ravikumar UT Austin Outline 1. High Dimensional Data : Large p, small n 2. Sparsity 3. Group Sparsity 4. Low Rank 1 Curse of Dimensionality Statistical Learning: Given

More information

Rank Constrained Matrix Optimization Problems

Rank Constrained Matrix Optimization Problems Rank Constrained Matrix Optimization Problems Defeng Sun Department of Mathematics and Risk Management Institute National University of Singapore This talk is based on a joint work with Yan Gao at NUS

More information

Low-Rank Matrix Recovery

Low-Rank Matrix Recovery ELE 538B: Mathematics of High-Dimensional Data Low-Rank Matrix Recovery Yuxin Chen Princeton University, Fall 2018 Outline Motivation Problem setup Nuclear norm minimization RIP and low-rank matrix recovery

More information

An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion

An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion Robert M. Freund, MIT joint with Paul Grigas (UC Berkeley) and Rahul Mazumder (MIT) CDC, December 2016 1 Outline of Topics

More information

CSC 576: Variants of Sparse Learning

CSC 576: Variants of Sparse Learning CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in

More information

An Optimization-based Approach to Decentralized Assignability

An Optimization-based Approach to Decentralized Assignability 2016 American Control Conference (ACC) Boston Marriott Copley Place July 6-8, 2016 Boston, MA, USA An Optimization-based Approach to Decentralized Assignability Alborz Alavian and Michael Rotkowitz Abstract

More information

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,

More information

1 Sparsity and l 1 relaxation

1 Sparsity and l 1 relaxation 6.883 Learning with Combinatorial Structure Note for Lecture 2 Author: Chiyuan Zhang Sparsity and l relaxation Last time we talked about sparsity and characterized when an l relaxation could recover the

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

Spectral k-support Norm Regularization

Spectral k-support Norm Regularization Spectral k-support Norm Regularization Andrew McDonald Department of Computer Science, UCL (Joint work with Massimiliano Pontil and Dimitris Stamos) 25 March, 2015 1 / 19 Problem: Matrix Completion Goal:

More information

Quadratic reformulation techniques for 0-1 quadratic programs

Quadratic reformulation techniques for 0-1 quadratic programs OSE SEMINAR 2014 Quadratic reformulation techniques for 0-1 quadratic programs Ray Pörn CENTER OF EXCELLENCE IN OPTIMIZATION AND SYSTEMS ENGINEERING ÅBO AKADEMI UNIVERSITY ÅBO NOVEMBER 14th 2014 2 Structure

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu

More information

Sparse Sensing in Colocated MIMO Radar: A Matrix Completion Approach

Sparse Sensing in Colocated MIMO Radar: A Matrix Completion Approach Sparse Sensing in Colocated MIMO Radar: A Matrix Completion Approach Athina P. Petropulu Department of Electrical and Computer Engineering Rutgers, the State University of New Jersey Acknowledgments Shunqiao

More information

Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas

Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas Department of Mathematics Department of Statistical Science Cornell University London, January 7, 2016 Joint work

More information

High-dimensional Statistical Models

High-dimensional Statistical Models High-dimensional Statistical Models Pradeep Ravikumar UT Austin MLSS 2014 1 Curse of Dimensionality Statistical Learning: Given n observations from p(x; θ ), where θ R p, recover signal/parameter θ. For

More information

Solving Corrupted Quadratic Equations, Provably

Solving Corrupted Quadratic Equations, Provably Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin

More information

Some tensor decomposition methods for machine learning

Some tensor decomposition methods for machine learning Some tensor decomposition methods for machine learning Massimiliano Pontil Istituto Italiano di Tecnologia and University College London 16 August 2016 1 / 36 Outline Problem and motivation Tucker decomposition

More information

Common-Knowledge / Cheat Sheet

Common-Knowledge / Cheat Sheet CSE 521: Design and Analysis of Algorithms I Fall 2018 Common-Knowledge / Cheat Sheet 1 Randomized Algorithm Expectation: For a random variable X with domain, the discrete set S, E [X] = s S P [X = s]

More information

A note on rank constrained solutions to linear matrix equations

A note on rank constrained solutions to linear matrix equations A note on rank constrained solutions to linear matrix equations Shravan Mohan September 5, 218 Abstract This preliminary note presents a heuristic for determining rank constrained solutions to linear matrix

More information

Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A.

Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A. . Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A. Nemirovski Arkadi.Nemirovski@isye.gatech.edu Linear Optimization Problem,

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

Low-Rank Factorization Models for Matrix Completion and Matrix Separation

Low-Rank Factorization Models for Matrix Completion and Matrix Separation for Matrix Completion and Matrix Separation Joint work with Wotao Yin, Yin Zhang and Shen Yuan IPAM, UCLA Oct. 5, 2010 Low rank minimization problems Matrix completion: find a low-rank matrix W R m n so

More information

Exact penalty decomposition method for zero-norm minimization based on MPEC formulation 1

Exact penalty decomposition method for zero-norm minimization based on MPEC formulation 1 Exact penalty decomposition method for zero-norm minimization based on MPEC formulation Shujun Bi, Xiaolan Liu and Shaohua Pan November, 2 (First revised July 5, 22) (Second revised March 2, 23) (Final

More information

(Part 1) High-dimensional statistics May / 41

(Part 1) High-dimensional statistics May / 41 Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2

More information

Rank-one Generated Spectral Cones Defined by Two Homogeneous Linear Matrix Inequalities

Rank-one Generated Spectral Cones Defined by Two Homogeneous Linear Matrix Inequalities Rank-one Generated Spectral Cones Defined by Two Homogeneous Linear Matrix Inequalities C.J. Argue Joint work with Fatma Kılınç-Karzan October 22, 2017 INFORMS Annual Meeting Houston, Texas C. Argue, F.

More information

Convex relaxation for Combinatorial Penalties

Convex relaxation for Combinatorial Penalties Convex relaxation for Combinatorial Penalties Guillaume Obozinski Equipe Imagine Laboratoire d Informatique Gaspard Monge Ecole des Ponts - ParisTech Joint work with Francis Bach Fête Parisienne in Computation,

More information

Collaborative Filtering Matrix Completion Alternating Least Squares

Collaborative Filtering Matrix Completion Alternating Least Squares Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade May 19, 2016

More information

Universal low-rank matrix recovery from Pauli measurements

Universal low-rank matrix recovery from Pauli measurements Universal low-rank matrix recovery from Pauli measurements Yi-Kai Liu Applied and Computational Mathematics Division National Institute of Standards and Technology Gaithersburg, MD, USA yi-kai.liu@nist.gov

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

PENNON A Generalized Augmented Lagrangian Method for Nonconvex NLP and SDP p.1/22

PENNON A Generalized Augmented Lagrangian Method for Nonconvex NLP and SDP p.1/22 PENNON A Generalized Augmented Lagrangian Method for Nonconvex NLP and SDP Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

Fast and Robust Phase Retrieval

Fast and Robust Phase Retrieval Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National

More information

Low Rank Matrix Completion Formulation and Algorithm

Low Rank Matrix Completion Formulation and Algorithm 1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9

More information

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional

More information

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.

More information

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Minru Bai(x T) College of Mathematics and Econometrics Hunan University Joint work with Xiongjun Zhang, Qianqian Shao June 30,

More information

Computational and Statistical Aspects of Statistical Machine Learning. John Lafferty Department of Statistics Retreat Gleacher Center

Computational and Statistical Aspects of Statistical Machine Learning. John Lafferty Department of Statistics Retreat Gleacher Center Computational and Statistical Aspects of Statistical Machine Learning John Lafferty Department of Statistics Retreat Gleacher Center Outline Modern nonparametric inference for high dimensional data Nonparametric

More information

Section 3.9. Matrix Norm

Section 3.9. Matrix Norm 3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix

More information

We describe the generalization of Hazan s algorithm for symmetric programming

We describe the generalization of Hazan s algorithm for symmetric programming ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,

More information

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the

More information

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C. Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal

More information

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013 Convex Optimization (EE227A: UC Berkeley) Lecture 6 (Conic optimization) 07 Feb, 2013 Suvrit Sra Organizational Info Quiz coming up on 19th Feb. Project teams by 19th Feb Good if you can mix your research

More information

Small zeros of hermitian forms over quaternion algebras. Lenny Fukshansky Claremont McKenna College & IHES (joint work with W. K.

Small zeros of hermitian forms over quaternion algebras. Lenny Fukshansky Claremont McKenna College & IHES (joint work with W. K. Small zeros of hermitian forms over quaternion algebras Lenny Fukshansky Claremont McKenna College & IHES (joint work with W. K. Chan) Institut de Mathématiques de Jussieu October 21, 2010 1 Cassels Theorem

More information

Block Coordinate Descent for Regularized Multi-convex Optimization

Block Coordinate Descent for Regularized Multi-convex Optimization Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline

More information

(k, q)-trace norm for sparse matrix factorization

(k, q)-trace norm for sparse matrix factorization (k, q)-trace norm for sparse matrix factorization Emile Richard Department of Electrical Engineering Stanford University emileric@stanford.edu Guillaume Obozinski Imagine Ecole des Ponts ParisTech guillaume.obozinski@imagine.enpc.fr

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence

More information

1-Bit Matrix Completion

1-Bit Matrix Completion 1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible

More information

Nonlinear Programming Models

Nonlinear Programming Models Nonlinear Programming Models Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Nonlinear Programming Models p. Introduction Nonlinear Programming Models p. NLP problems minf(x) x S R n Standard form:

More information

Sparse PCA with applications in finance

Sparse PCA with applications in finance Sparse PCA with applications in finance A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon 1 Introduction

More information

Two Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery

Two Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery Two Results on the Schatten p-quasi-norm Minimization for Low-Rank Matrix Recovery Ming-Jun Lai, Song Li, Louis Y. Liu and Huimin Wang August 14, 2012 Abstract We shall provide a sufficient condition to

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator

More information

Compression, Matrix Range and Completely Positive Map

Compression, Matrix Range and Completely Positive Map Compression, Matrix Range and Completely Positive Map Iowa State University Iowa-Nebraska Functional Analysis Seminar November 5, 2016 Definitions and notations H, K : Hilbert space. If dim H = n

More information

Applications of Robust Optimization in Signal Processing: Beamforming and Power Control Fall 2012

Applications of Robust Optimization in Signal Processing: Beamforming and Power Control Fall 2012 Applications of Robust Optimization in Signal Processing: Beamforg and Power Control Fall 2012 Instructor: Farid Alizadeh Scribe: Shunqiao Sun 12/09/2012 1 Overview In this presentation, we study the applications

More information

Maximum Margin Matrix Factorization

Maximum Margin Matrix Factorization Maximum Margin Matrix Factorization Nati Srebro Toyota Technological Institute Chicago Joint work with Noga Alon Tel-Aviv Yonatan Amit Hebrew U Alex d Aspremont Princeton Michael Fink Hebrew U Tommi Jaakkola

More information

Lecture 3. Random Fourier measurements

Lecture 3. Random Fourier measurements Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko

More information

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term EE 150 - Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko JPL Third Term 2011-2012 Due on Thursday May 3 in class. Homework Set #4 1. (10 points) (Adapted

More information

10-725/36-725: Convex Optimization Prerequisite Topics

10-725/36-725: Convex Optimization Prerequisite Topics 10-725/36-725: Convex Optimization Prerequisite Topics February 3, 2015 This is meant to be a brief, informal refresher of some topics that will form building blocks in this course. The content of the

More information

Robust Tensor Completion and its Applications. Michael K. Ng Department of Mathematics Hong Kong Baptist University

Robust Tensor Completion and its Applications. Michael K. Ng Department of Mathematics Hong Kong Baptist University Robust Tensor Completion and its Applications Michael K. Ng Department of Mathematics Hong Kong Baptist University mng@math.hkbu.edu.hk Example Low-dimensional Structure Data in many real applications

More information

Ahmed Mohammed. Harnack Inequality for Non-divergence Structure Semi-linear Elliptic Equations

Ahmed Mohammed. Harnack Inequality for Non-divergence Structure Semi-linear Elliptic Equations Harnack Inequality for Non-divergence Structure Semi-linear Elliptic Equations International Conference on PDE, Complex Analysis, and Related Topics Miami, Florida January 4-7, 2016 An Outline 1 The Krylov-Safonov

More information

Lecture 4: January 26

Lecture 4: January 26 10-725/36-725: Conve Optimization Spring 2015 Lecturer: Javier Pena Lecture 4: January 26 Scribes: Vipul Singh, Shinjini Kundu, Chia-Yin Tsai Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

arxiv: v3 [stat.me] 8 Jun 2018

arxiv: v3 [stat.me] 8 Jun 2018 Between hard and soft thresholding: optimal iterative thresholding algorithms Haoyang Liu and Rina Foygel Barber arxiv:804.0884v3 [stat.me] 8 Jun 08 June, 08 Abstract Iterative thresholding algorithms

More information

arxiv: v3 [math.oc] 19 Oct 2017

arxiv: v3 [math.oc] 19 Oct 2017 Gradient descent with nonconvex constraints: local concavity determines convergence Rina Foygel Barber and Wooseok Ha arxiv:703.07755v3 [math.oc] 9 Oct 207 0.7.7 Abstract Many problems in high-dimensional

More information

Canonical Problem Forms. Ryan Tibshirani Convex Optimization

Canonical Problem Forms. Ryan Tibshirani Convex Optimization Canonical Problem Forms Ryan Tibshirani Convex Optimization 10-725 Last time: optimization basics Optimization terology (e.g., criterion, constraints, feasible points, solutions) Properties and first-order

More information

Recall : Eigenvalues and Eigenvectors

Recall : Eigenvalues and Eigenvectors Recall : Eigenvalues and Eigenvectors Let A be an n n matrix. If a nonzero vector x in R n satisfies Ax λx for a scalar λ, then : The scalar λ is called an eigenvalue of A. The vector x is called an eigenvector

More information

Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery

Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery Supplementary Materials for Riemannian Pursuit for Big Matrix Recovery Mingkui Tan, School of omputer Science, The University of Adelaide, Australia Ivor W. Tsang, IS, University of Technology Sydney,

More information

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

Constructing Explicit RIP Matrices and the Square-Root Bottleneck Constructing Explicit RIP Matrices and the Square-Root Bottleneck Ryan Cinoman July 18, 2018 Ryan Cinoman Constructing Explicit RIP Matrices July 18, 2018 1 / 36 Outline 1 Introduction 2 Restricted Isometry

More information

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016 U.C. Berkeley CS294: Spectral Methods and Expanders Handout Luca Trevisan February 29, 206 Lecture : ARV In which we introduce semi-definite programming and a semi-definite programming relaxation of sparsest

More information

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March

More information

Robust Principal Component Analysis

Robust Principal Component Analysis ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M

More information

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω

More information

THE NONCOMMUTATIVE TORUS

THE NONCOMMUTATIVE TORUS THE NONCOMMUTATIVE TORUS The noncommutative torus as a twisted convolution An ordinary two-torus T 2 with coordinate functions given by where x 1, x 2 [0, 1]. U 1 = e 2πix 1, U 2 = e 2πix 2, (1) By Fourier

More information

Sparse Recovery via Partial Regularization: Models, Theory and Algorithms

Sparse Recovery via Partial Regularization: Models, Theory and Algorithms Sparse Recovery via Partial Regularization: Models, Theory and Algorithms Zhaosong Lu and Xiaorui Li Department of Mathematics, Simon Fraser University, Canada {zhaosong,xla97}@sfu.ca November 23, 205

More information

Partial regularity for fully nonlinear PDE

Partial regularity for fully nonlinear PDE Partial regularity for fully nonlinear PDE Luis Silvestre University of Chicago Joint work with Scott Armstrong and Charles Smart Outline Introduction Intro Review of fully nonlinear elliptic PDE Our result

More information