Matrix Rank Minimization with Applications

Similar documents
Rank Minimization and Applications in System Theory

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Sparse Covariance Selection using Semidefinite Programming

Spectral k-support Norm Regularization

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery

A note on rank constrained solutions to linear matrix equations

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

Distance Geometry-Matrix Completion

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

Lecture 5 : Projections

Relaxations and Randomized Methods for Nonconvex QCQPs

A direct formulation for sparse PCA using semidefinite programming

Reweighted Nuclear Norm Minimization with Application to System Identification

A direct formulation for sparse PCA using semidefinite programming

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Optimization and l 1 -minimization

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

Constrained optimization

Semidefinite Programming Basics and Applications

Lecture 7: Convex Optimizations

Fast Linear Iterations for Distributed Averaging 1

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

10-725/36-725: Convex Optimization Prerequisite Topics

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Conditional Gradient (Frank-Wolfe) Method

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Convex envelopes, cardinality constrained optimization and LASSO. An application in supervised learning: support vector machines (SVMs)

Rank Minimization and Applications in System Theory


Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture 2 Part 1 Optimization

Introduction and Math Preliminaries

An Introduction to Correlation Stress Testing

Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

Frank-Wolfe Method. Ryan Tibshirani Convex Optimization

Lecture 20: November 1st

Rank minimization via the γ 2 norm

Nonlinear Optimization for Optimal Control

Cardinality Optimization Problems

Fundamentals of Matrices

Constrained Optimization and Lagrangian Duality

Tractable Upper Bounds on the Restricted Isometry Constant

Common-Knowledge / Cheat Sheet

Sparse Solutions of an Undetermined Linear System

Determinant maximization with linear. S. Boyd, L. Vandenberghe, S.-P. Wu. Information Systems Laboratory. Stanford University

Network Localization via Schatten Quasi-Norm Minimization

Part I Formulation of convex optimization problems

5. Duality. Lagrangian

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

A Quick Tour of Linear Algebra and Optimization for Machine Learning

Course Outline. FRTN10 Multivariable Control, Lecture 13. General idea for Lectures Lecture 13 Outline. Example 1 (Doyle Stein, 1979)

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term

12. Interior-point methods

Subgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives

Reconstruction from Anisotropic Random Measurements

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Sparse PCA with applications in finance

Conditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint

15. Conic optimization

The Graph Realization Problem

Convex Optimization & Lagrange Duality

Iterative Methods. Splitting Methods

Sparse regression. Optimization-Based Data Analysis. Carlos Fernandez-Granda

Stochastic dynamical modeling:

Sparse Optimization Lecture: Basic Sparse Optimization Models

EE 227A: Convex Optimization and Applications October 14, 2008

Nonconvex? NP! (No Problem!) Ryan Tibshirani Convex Optimization /36-725

The convex algebraic geometry of rank minimization

Interior Point Algorithms for Constrained Convex Optimization

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Lecture: Matrix Completion

High Dimensional Covariance and Precision Matrix Estimation

The Q-parametrization (Youla) Lecture 13: Synthesis by Convex Optimization. Lecture 13: Synthesis by Convex Optimization. Example: Spring-mass System

arxiv: v1 [math.st] 10 Sep 2015

Convex optimization. Javier Peña Carnegie Mellon University. Universidad de los Andes Bogotá, Colombia September 2014

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Convex Optimization M2

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Module 04 Optimization Problems KKT Conditions & Solvers

Structured matrix factorizations. Example: Eigenfaces

Homework 4. Convex Optimization /36-725

Sparsity Matters. Robert J. Vanderbei September 20. IDA: Center for Communications Research Princeton NJ.

Low Rank Matrix Completion Formulation and Algorithm

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

Computational Methods. Eigenvalues and Singular Values

Optimization for Compressed Sensing

Maximum Margin Matrix Factorization

FRTN10 Multivariable Control, Lecture 13. Course outline. The Q-parametrization (Youla) Example: Spring-mass System

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Recovery of Simultaneously Structured Models using Convex Optimization

Lecture 8: February 9

Functional Analysis Review

Lecture 8: Linear Algebra Background

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Transcription:

Matrix Rank Minimization with Applications Maryam Fazel Haitham Hindi Stephen Boyd Information Systems Lab Electrical Engineering Department Stanford University 8/2001 ACC 01

Outline Rank Minimization Problem (RMP) Examples Solution methods ACC 01 1

Rank Minimization Problem (RMP) minimize Rank X subject to X C, X R m n is the optimization variable; C is convex set RMP is difficult nonconvex problem (NP-hard) RMP arises in many application areas usual meaning: find simplest or minimum order system; model with fewest parameters... (Occam s razor) this talk: new methods to (approximately) solve RMP ACC 01 2

Maximum sparsity problem important special case of RMP: variable X = diag(x) then Rank X = Card(x), number of nonzero xi RMP reduces to finding the sparsest vector in convex set C: minimize Card(x) subject to x C meaning: find simplest model, design with fewest components, sparse signal representation; extract sparse signals from noise,... ACC 01 3

Constrained factor analysis find minimum rank covariance matrix close to measured ˆΣ, consistent with prior info minimize Rank Σ subject to Σ ˆΣ ɛ Σ = Σ T 0 Σ C Σ R n n is variable, C is prior information, ɛ is tolerance Rank Σ is number of factors that explain Σ without last constraint, can solve by eigenvalue decomposition; but general case is difficult ACC 01 4

Minimum order controller design find minimum order controller that achieves specified closed-loop performance [ minimize Rank subject to [ X I I Y X I I Y ] 0 ] other LMIs in X, Y with variables X = X T, Y = Y T R n n, possibly others ACC 01 5

Minimum order system realization find minimum order system that satisfies time-domain specs (rise-time, slew-rate, overshoot, settling-time,... ) h1 h2 hn minimize Rank h2 h3 hn+1... hn hn+1 h2n 1 subject to F (h1,..., hn) g variables are impulse response h1,..., h2n 1; linear inequality constraints involve only h1,..., hn ACC 01 6

Euclidean distance matrix (EDM) D R n n is EDM if there are x1,..., xn R r s.t. Dij = xi xj 2 r is called embedding dimension [Schoenberg 35] D = D T R n n is EDM with embedding dimension r iff Dii = 0, VDV 0, where V = I 1 n 11T, Rank VDV r ACC 01 7

Minimum embedding dimension find EDM D with minimum embedding dimension satisfying distance bounds minimize Rank VDV subject to Dii = 0 i = 1,..., n VDV 0 Lij Dij Uij i, j = 1,..., n. applications: statistics/psychometrics (called multidimensional scaling) chemistry (molecular conformation) ACC 01 8

Exact solution methods special cases with analytical solutions via SVD, EV, e.g., minimize Rank X subject to X A b solution: sum of r leading dyads in SVD of A, where σr > b, σr+1 b special cases that reduce to convex problems (e.g., [Mesbahi 97]) global optimization (e.g., branch and bound) ACC 01 9

Heuristic solution methods Factorization methods (e.g., [Iwasaki 99]) Rank X r iff there are F R m r, G R r n s.t. X = F G solve feasibility problem with variables F, G; use bisection on r Analytic anti-centering [David 94] use Newton method in reverse to maximize log barrier or potential function ACC 01 10

Alternating projections [Grigoriadis & Beran 00] alternate between projecting onto constraint set C via convex optimization projecting onto set of matrices of rank r via SVD x0 Rank X r PSfrag replacements C ACC 01 11

Trace heuristic for PSD matrices for X = X T 0, minimizing trace tends to give low rank solution [Mesbahi 97, Pare 00] RMP Trace heuristic minimize Rank X subject to X C minimize Tr X subject to X C convex problem, hence efficiently solved provides lower bound on min rank objective ACC 01 12

variation: weighted trace minimization minimize Tr W X subject to X C where W = W T 0 ACC 01 13

Log-det heuristic for PSD matrices for X = X T 0, (locally) minimize log det(x + δi) (δ > 0 is small constant for regularization) RMP Log-det heuristic minimize Rank X subject to X C minimize log det(x + δi) subject to X C objective is nonconvex (in fact, concave) can use any method to find a local minimum ACC 01 14

Idea behind log-det heuristic n n Rank X = 1(λi > 0) log det(x + δi) = log(λi + δ) i=1 i=1 PSfrag replacements 1 log(λi + δ) log δ λi ACC 01 15

Iterative linearization method linearize (concave) objective at Xk 0: log det(x + δi) log det(xk + δi) + Tr(Xk + δi) 1 (X Xk) minimize linearized objective (a convex problem): Xk+1 = argmin Tr(Xk + δi) 1 X X C i.e., iterative weighted trace minimization with X0 = I, first iteration same as trace heuristic only a few iterations needed (about 5 or 6) ACC 01 16

Semidefinite embedding question: can we extend trace & log-det heuristics to general (nonsquare, non PSD) matrices? let X R m n then Rank X r iff there are Y = Y T R m m, Z = Z T R n n, s.t. [ Rank Y 0 0 Z ] 2r, [ Y X X T Z ] 0 thus, can embed general (non PSD) RMP in a (larger) PSD RMP ACC 01 17

RMP in embedded PSD form minimize Rank X subject to X C equivalent to PSD RMP [ minimize Rank [ subject to Y X X T Z X C, Y 0 0 Z ] ] 0 with variables X R m n, Y = Y T R m m, Z = Z T R n n can now apply any method for symmetric PSD RMP ACC 01 18

Trace heuristic for general matrices [ minimize Tr [ subject to Y 0 0 Z Y X X T Z X C ] ] 0 can show this is equivalent to minimize X subject to X C, where X = n i=1 σ i(x), called nuclear norm of X, is dual of spectral (maximum singular value) norm ACC 01 19

Convex envelope convex envelope of f : C R is largest convex function g s.t. g(x) f(x) for all x C PSfrag replacements f(x) g(x) best convex lower approximation epigraph of g is convex hull of epigraph of f ACC 01 20

Convex envelope of rank fact: X is cvx envelope of Rank X on {X R m n X 1} conclusions: trace heuristic minimizes convex envelope of rank (i.e., the best convex approximation to rank) over ball hence, heuristic provides lower bound on objective provides theoretical support for use of trace/nuclear norm heuristic ACC 01 21

Maximum sparsity problem nuclear norm heuristic for diagonal X becomes minimize x 1 subject to x C well-known l1 heuristic for finding sparse solutions used in LASSO methods in statistics [Tibshirani 94], signal decomposition by basis pursuit [Donoho 96],... x 1 is convex envelope of Card(x) on {x x 1} trace/nuclear norm heuristic is extension of l1 heuristic to matrix case ACC 01 22

Log-det heuristic for general matrices ([ Y 0 minimize log det 0 Z [ ] Y X subject to X T 0 Z X C ] + δi ) can show this is equivalent to n minimize i=1 log(σ i(x) + δ) subject to X C can linearize as before to obtain iterations in X, Y, Z ACC 01 23

Iterative l1 heuristic for maximum sparsity problem log-det heuristic for maximum sparsity problem yields minimize i log( x i + δ) subject to x C. iterative linearization/minimization yields x (k+1) = argmin x C n w (k) i xi, w (k) 1 i = x (k) i + δ i=1 each step is weighted l1 norm minimization when x (k) i small, weight in next step is large; hence, small entries in x are pushed towards zero (subject to x C) ACC 01 24

Example: minimum order system realization with step response constraints find minimum order system that satisfies li si ui, i = 1,..., 16 n = 16 trace/nuclear norm heuristic yields rank 5 log-det heuristic converges in 5 steps, yields rank 4 step response (solid) and specs (dashed) 1.2 1 0.8 0.6 PSfrag replacements 0.4 0.2 0 0.2 0 5 10 15 20 25 30 35 t ACC 01 25

ACC 01 26 iterations 10 1 1.5 2 2.5 3 3.5 4 4.5 5 PSfrag replacements log of non-zero σ i s 2 3 4 5 6 7 8 9 1 0

Conclusions RMP is difficult nonconvex problem, with many applications maximum sparsity problem is special case trace and log-det heuristics for PSD RMP can be extended to general case, via semidefinite embedding generalization of trace heuristic to general matrices is nuclear norm, which is convex envelope of rank ACC 01 27