EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Similar documents
Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

8. Geometric problems

8. Geometric problems

9. Geometric problems

Constrained Optimization and Lagrangian Duality

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 18

12. Interior-point methods

CSCI : Optimization and Control of Networks. Review on Convex Optimization

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

4. Convex optimization problems

Lecture: Convex Optimization Problems

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Optimization M2

A Brief Review on Convex Optimization

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

5. Duality. Lagrangian

Nonlinear Programming Models

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

Convex optimization problems. Optimization problem in standard form

Convex Optimization & Lagrange Duality

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Interior Point Algorithms for Constrained Convex Optimization

Semidefinite Programming

Lecture Note 5: Semidefinite Programming for Stability Analysis

Convex Optimization Boyd & Vandenberghe. 5. Duality

Lecture 1: Background on Convex Analysis

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

Homework Set #6 - Solutions

subject to (x 2)(x 4) u,

4. Convex optimization problems

On John type ellipsoids

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Lecture: Duality of LP, SOCP and SDP

Semidefinite Programming Basics and Applications

Lecture: Duality.

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 4

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

Convex Optimization and Modeling

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

CS-E4830 Kernel Methods in Machine Learning

12. Interior-point methods

Convex Optimization and Modeling

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Linear and non-linear programming

15. Conic optimization

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lecture 7: Weak Duality

Determinant maximization with linear. S. Boyd, L. Vandenberghe, S.-P. Wu. Information Systems Laboratory. Stanford University

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Lagrangian Duality Theory

Constrained optimization

Support Vector Machines

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term

Lecture 9 Sequential unconstrained minimization

ICS-E4030 Kernel Methods in Machine Learning

The proximal mapping

Exercises. Exercises. Basic terminology and optimality conditions. 4.2 Consider the optimization problem

Lecture 6: Conic Optimization September 8

Linear Algebra Massoud Malek

Lecture 3. Optimization Problems and Iterative Algorithms

Chapter 1. Preliminaries

Convex Optimization: Applications

Convex Optimization M2

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

Lecture 14: Optimality Conditions for Conic Problems

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Largest dual ellipsoids inscribed in dual cones

Summer School: Semidefinite Optimization

CS711008Z Algorithm Design and Analysis

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Lecture 8. Strong Duality Results. September 22, 2008

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Contents Real Vector Spaces Linear Equations and Linear Inequalities Polyhedra Linear Programs and the Simplex Method Lagrangian Duality

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

Linear Algebra Review: Linear Independence. IE418 Integer Programming. Linear Algebra Review: Subspaces. Linear Algebra Review: Affine Independence

Assignment 1: From the Definition of Convexity to Helley Theorem

Optimisation convexe: performance, complexité et applications.

Lecture 2: Convex functions

Semidefinite Programming

EE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

EE 227A: Convex Optimization and Applications October 14, 2008

Lecture 1 Introduction

Optimality Conditions for Constrained Optimization

Optimization Theory. A Concise Introduction. Jiongmin Yong

Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS

Lecture 3 January 28

Conic Linear Optimization and its Dual. yyye

4. Algebra and Duality

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Transcription:

EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 17 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory May 29, 2012 Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 1 / 28

Outline 1 Projection Onto Sets Introduction Projection Onto Convex Sets 2 Distance Between Sets Introduction Distance Between Convex Sets 3 Euclidean Distance and Angle Problems Gram Matrix and Realizability Common Constraints and Objectives Angle/Distance Only Problems 4 Extremal Volume Ellipsoids Minimum Volume Covering Ellipsoid Maximum Volume Inscribed Ellipsoid Affine Invariance of Extremal Volume Ellipsoids 5 Centering Problems Chebyshev Center Maximum Volume Ellipsoid Center Analytic Center Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 2 / 28

Projection Onto Sets Introduction Distance and Projection on a Set The distance from a point x 0 R n to a closed set C R n, with respect to the norm, is defined as dist(x 0, C) inf{ x x 0 : x C}. The projection of x 0 on C is any point z C for which z x 0 = dist(x 0, C). The notation P C : R n R n is used to denote any function for which P C(x 0) is a projection of x 0 onto C, i.e., for all x 0, we have P C(x 0) C, P C(x 0) x 0 = dist(x 0, C) P C(x 0) = argmin{ x x 0 : x C}. Properties: If C is closed and convex, and is strictly convex, then for any x 0, there is a unique projection P C(x 0) onto C. If, for every x 0, there is a unique Euclidean projection of x 0 on C, then C is closed and convex. Example: Projection onto rank-k matrices Suppose that we have X 0 R m n, C = { X R m n : rank(x) k }, = 2 = spectral norm. If X 0 has an SVD given by X 0 = r i=1 σiuivt i, where r = rank(x 0), and we have σ 1 σ r > 0, then Y = min{k,r} i=1 σ iu ivi T is a projection of X 0 on C. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 3 / 28

Projection Onto Sets Introduction Projection Via Indicator and Support Functions The problem of finding the distance from a point x 0 to a set C (along with its dual) can be expressed compactly using the indicator function I C and the support function S C: { 0, x C I C(x), S C(x) sup x T y., x C y C The primal problem of finding the distance between x 0 and C can then be expressed as minimize x x 0 subject to I C(x) 0 minimize y subject to I C(x) 0 x 0 x = y Note that the primal problem is convex if C is a closed convex set. The dual problem of the second form of the primal is given by maximize z T x 0 S C(z) subject to z 1 If C is a closed convex set (i.e., the primal is convex), then if z is dual optimal with a positive objective value, then z T x 0 > z T x for all x C, i.e., z defines a separating hyperplane between x 0 and C... Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 4 / 28

Projection Onto Sets Projection Onto Convex Sets Projecting a Point on a Convex Set If C is convex, then P C(x 0) and dist(x 0, C) can be computed via convex optimization. Representing C by a set of convex inequalities f i(x) 0 for i = 1,..., m and affine equalities Ax = b (A R p n, b R p ), dist(x 0, C) and P C(x 0) can be found by solving minimize x x 0 subject to f i(x) 0, i = 1,..., m Ax = b minimize subject to y f i(x) 0, i = 1,..., m Ax = b x 0 x = y The dual problem of the second form of the primal is given by the following. { m } maximize µ T x 0 + inf λ if i(x) + ν T (Ax b) µ T x x subject to λ 0 i=1. µ 1 The primal problem is feasible if and only if C is nonempty. When the primal is feasible, its optimal value is dist(x 0, C), and any optimal point is a projection of x 0 on C. The dual problem leads to insights regarding separation of x 0 and C. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 5 / 28.

Projection Onto Sets Euclidean Projection Examples Projection Onto Convex Sets Euclidean projection on a polyhedron: The projection of x 0 on a polyhedron P {x : Ax b} can be computed by solving the QP minimize x x 0 2 2 subject to Ax b For the hyperplane C hp = { x : a T x = b } and halfspace C hs = { x : a T x b }, we have ( b a T ) { ) x 0 x 0 + (b a T x 0 a/ a 2 P Chp (x 0 ) = x 0 + a 2 a, P Chs (x 0 ) = 2, at x 0 > b 2 x 0, a T x 0 b For the box C box = {x : l x u} (where l u), we have [ P Cbox (x 0 ) ] k = l k for [x 0 ] k l k, [x 0 ] k for l k [x 0 ] k u k, and u k for [x 0 ] k u k. Euclidean projection on a proper cone: The projection of x 0 onto a proper cone K can be obtained by solving the problem.. minimize x x 0 2 2 subject to x K 0 The KKT conditions for this problem can be expressed in terms of a primal nonnegative and dual nonpositive decomposition as follows. x 0 = x + + x, x + K 0, x K 0, x T + x = 0. For K = R n +, we have [P K(x 0 )] k = max { [x 0 ] k, 0 }. For K = S n + and the Frobenius norm F, we have P K(X 0 ) = n i=1 max{λ i, 0} v i v T i, where X 0 = n i=1 λ iv i v T i is the eigenvalue decomposition of X 0 S n. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 6 / 28.

Projection Onto Sets Projection Onto Convex Sets Separating a Point from a Convex Set Suppose C is a closed convex set. Then, we have the following. If x 0 C, then dist(x 0, C) = 0 with an optimal point x 0. If x 0 C, then dist(x 0, C) > 0, and the optimal value of the distance problem is positive. Furthermore, a dual optimal point provides a separating hyperplane between x 0 and C. For the Euclidean norm, if P C (x 0 ) denotes the Euclidean projection of x 0 on C, where x 0 C, then the hyperplane characterized by strictly separates x 0 from C. (P C (x 0 ) x 0 ) T (x (1/2) (x 0 + P C (x 0 ))) = 0, x0 PC(x0) C For a general norm, a separating hyperplane can be obtained via Lagrange duality. From the dual problem, if λ, µ, ν are dual feasible with a positive objective value, then we have m µ T x 0 µ T x + λ i f i (x) + ν T (Ax b) > 0, for all x, i=1 and so µ T x 0 > µ T x for x C. Thus, µ defines a strictly separating hyperplane in this case. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 7 / 28

Distance Between Sets Introduction Expressing the Distance Between Two Sets The distance between two sets C and D, with respect to the norm, is defined as dist(c, D) inf{ x y : x C, y D}. Properties: The sets C and D do not intersect if dist(c, D) > 0. The sets C and D intersect if dist(c, D) = 0 and the infimum is attained. The distance between sets can be expressed as the distance between a point and a set. dist(c, D) = dist(0, C D) = dist(0, D C). Distance and Separation Via Indicator and Support Functions: The problem of finding the distance between sets can be posed in terms of indicator functions. The dual of this problem can be expressed using support functions. Primal: minimize x y subject to I C (x) 0 I D (y) 0 Dual: maximize S C ( z) S D (z) subject to z 1. minimize w subject to I C (x) 0 I D (y) 0 x y = w If C and D are convex sets, then the primal problem is convex and (assuming strong duality holds) the dual problem can be used to find a separating hyperplane. Specifically, if z is dual feasible with a positive objective value, then S D (z) < S C ( z), which means that sup z T x < inf x D x C zt x. In other words, z defines a hyperplane that strictly separates C and D. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 8 / 28

Distance Between Sets Distance Between Convex Sets Computing the Distance Between Convex Sets Suppose C and D are described by two sets of convex inequalities C = {x : f i(x) 0, i = 1,..., m}, D = {x : g i(x) 0, i = 1,..., p}. This can admit affine equalities by including two convex inequalities for each affine equality. Then, we can find dist(c, D) by solving the convex optimization problem minimize subject to x y f i(x) 0, i = 1,..., m g i(y) 0, i = 1,..., p. Example: Euclidean distance between polyhedra C = {x : A 1x b 1}, D = {x : A 1x b 1}. minimize x y 2 subject to A 1x b 1. A 2x b 2 C D Squaring the objective, we obtain an equivalent QP. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 9 / 28

Distance Between Sets Separating Convex Sets Distance Between Convex Sets The dual of the problem of finding the distance between convex sets has an interesting interpretation in terms of separating hyperplanes. Starting from the equivalent form of the primal problem minimize subject to w f i(x) 0, i = 1,..., m g i(y) 0, i = 1,..., p x y = w we obtain the dual problem { m } maximize inf x λ if i(x) + z T x i=1 subject to z 1, λ 0, µ 0 + inf y, { p } µ ig i(y) z T y If λ and µ are dual feasible with a positive objective value (i.e., dist(c, D) > 0), then m p λ if i(x) + z T x + µ ig i(y) z T y > 0, i=1 for all x and y. In particular, for x C and y D, we have z T x > z T y, and so z defines a hyperplane that strictly separates C and D. i=1 Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 10 / 28 i=1.

Euclidean Distance and Angle Problems Gram Matrix and Realizability Configurations and the Gram Matrix Consider a set of vectors a 1,..., a n R n with known Euclidean lengths l 1 = a 1 2,..., l n = a n 2. We refer to this set as a configuration in general and as a basis when they are linearly independent. The lengths, distances, correlation coefficients, and angles between vectors of the configuration can be expressed in terms of the Gram matrix G R n n given by Length: (of a i ) G A T A, A [ a 1 a n ]. l i a i 2 = (G i,i ) 1/2, G i,i = l 2 i, i = 1,..., n. Distance: (between a i and a j ) d i,j a i a j 2 = ( li 2 + l2 j 2G ) 1/2 i,j, Gi,j = l2 i + l2 j d2 i,j, i, j = 1,..., n. 2 Correlation coefficient: (between a i and a j ) ρ i,j Angle: (between a i and a j ) a T i a j a i 2 a j 2 = G i,j l i l j, G i,j = l i l j ρ i,j, i, j = 1,..., n. θ i,j arccos ρ i,j = arccos(g i,j / (l i l j )), G i,j = l i l j cos θ i,j, i, j = 1,..., n. These quantities are invariant under orthogonal transformations. Namely, if Q R n n is orthogonal, then the configuration Qa 1,..., Qa n has the same Gram matrix as the original one. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 11 / 28

Euclidean Distance and Angle Problems Realizability of the Gram Matrix Gram Matrix and Realizability Euclidean distance problems can be posed in terms of the Gram matrix G = A T A. From linear algebra, a matrix G S n is the Gram matrix of a configuration {a 1,..., a n} (i.e., G = A T A with A = [ a 1 a n ] ) if and only if G 0. When G 0, we can construct a configuration with Gram matrix G by taking A = G 1/2, using an eigenvalue decomposition of G. When G 0, a configuration can be obtained via the Cholesky decomposition G = LL T, where L is a lower triangular matrix, by taking A = L T. All configurations with Gram matrix G can be constructed by orthogonal transformation, given any one solution A. Namely, if ÃT à = G is any solution, then à = QA for some orthogonal matrix Q. A set of lengths, distances, correlation coefficients, and/or angles is said to be realizable, i.e., those of some configuration, if and only if we have G 0, G i,i = l i, i = 1,..., n. In other words, G is realizable for a set of length, distance, correlation coefficient, and/or angle constraints if and only if the above condition is satisfied. This realizability constraint can be used to pose several geometric problems as convex optimization problems, with G S n as the optimization variable. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 12 / 28

Euclidean Distance and Angle Problems Configuration Constraints Common Constraints and Objectives Angle constraints: A lower and upper bound on an angle, α θ i,j β, is equivalent to the affine inequality constraints l il j cos α G i,j l il j cos β. Distance constraints: A lower and upper bound on a distance, d min d i,j d max, is equivalent to the affine inequality constraints d 2 min l 2 i + l 2 j 2G i,j d 2 max. Singular value constraints: The singular values σ 1 σ n of A are the square roots of the eigenvalues λ 1 λ n of G. Thus, σ 2 1 is a convex function of G, whereas σ 2 n is a concave function of G. So, the following constraints are convex. σ 2 1(G) σ 2 max, σ 2 n(g) σ 2 min. Condition number constraints: The condition number κ σ 1/σ n of A is a quasiconvex function of G. Hence, the following constraint is convex. κ(g) κ max. Instead of this, we can minimize κ(g) over all configurations satisfying other convex geometric constraints, by quasiconvex optimization. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 13 / 28

Euclidean Distance and Angle Problems Dual Basis and Volume Objectives Common Constraints and Objectives Dual Basis: When G 0, the vectors a 1,..., a n form a basis for R n. The associated dual basis b 1,..., b n is one that satisfies b T i a j = δ i,j, where δ i,j is the Kronecker delta function. The dual basis vectors b 1,..., b n are simply the rows of A 1. As such, the Gram matrix associated with the dual basis is G 1. The squared lengths of the dual basis vectors are given by b i 2 2 = et i G 1 e i, i = 1,..., n, and are convex functions of G. So, they can be minimized or upper bound constrained. Also, n i=1 b i 2 2 = tr( G 1), is a measure of a well conditioned basis and is convex in G. Volume Under Linear Transformation: The volume of a set C R n under a linear transformation A (i.e., C lt {Au : u C}) is ( ( 1/2 vol(c lt ) = vol(c) det A A)) T = vol(c) (det G) 1/2. Example: Ellipsoid (E { Au : u 2 1 } ) vol(e) = γ ub (det G) 1/2, γ ub = πn/2 Γ ( n 2 + 1). Example: Simplex (S {Au : u conv{0, e 1,..., e n}}) vol(s) = γ ub (det G) 1/2, γ us = 1/n!. The volume of the warped set C lt can be maximized by equivalently maximizing log det G. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 14 / 28

Euclidean Distance and Angle Problems Angle/Distance Only Problems Bounding Correlations for Angle Only Problems For problems which only involve angles (or equivalently correlation coefficients), many constraints on the Gram matrix G can be simplified as they will not depend on the actual lengths of the vectors of the configuration. This follows from the fact that where l is the vector of lengths given by and C is the correlation matrix given by G = diag(l) C diag(l), l = [ l 1 l n ] T, C i,j = ρ i,j = cos θ i,j, i, j = 1,..., n. If G 0 for any l 0, then G 0 for all l 0, which occurs if and only if C 0. A set of angles θ i,j [0, π] for i, j = 1,..., n is realizable if and only if C 0, which is an LMI in ρ i,j for i, j = 1,..., n. As an example, given lower and upper bounds on some angles, we can find the minimum and maximum possible value of some other angle, over all configurations, by solving two SDPs. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 15 / 28

Euclidean Distance and Angle Problems Euclidean Distance Problems Angle/Distance Only Problems In a Euclidean distance problem, the only concern is the set of distances between vectors d i,j. Distances are invariant under orthogonal transformations, i.e., the configuration à = QA has the same distances as the original one A, for any orthogonal matrix Q. Distances are invariant under translation, i.e., the configuration à = A b1t has the same distances as the original one A, for any b R n. If b is the centroid given by b = 1 n a i = (1/n) A1, n i=1 then we have Ã1 = 0. Hence, without loss of generality, we can assume that A1 = 0, i.e., A is orthogonal to 1 and lies solely in 1. Defining z R n as z i = li 2, and D Sn by D i,j = d 2 i,j (with D i,i = 0), the condition G 0 for some set of lengths is equivalent to ( ) G = z1 T + 1z T D /2 0 for some z 0. A matrix D S n with nonnegative elements and zero diagonal that satisfies the above relation is called a Euclidean distance matrix. Assuming A1 = 0, we can show that G 0 if and only if the following conditions on D hold: D i,i = 0, i = 1,..., n, D i,j 0, i, j = 1,..., n, ( ) ( ) I (1/n) 11 T D I (1/n) 11 T 0. Therefore, any Euclidean distance problem that is convex in the squared distances can be expressed as convex problem with variable D S n. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 16 / 28

Extremal Volume Ellipsoids The Löwner-John Ellipsoid Minimum Volume Covering Ellipsoid Let C R n be some bounded set with nonempty interior. Suppose we wish to find the minimum volume ellipsoid that covers C, which is called the Löwner-John ellipsoid. To characterize this ellipsoid, it will be convenient to use the following parametrization: E = { v : Av + b 2 1 }, where we can assume without loss of generality that A S n ++. Then, the Löwner-John ellipsoid E lj can be found by solving the following problem. Minimum Volume Covering Ellipsoid Problem minimize subject to log det A 1 sup Av + b 2 1, v C with variables A S n and b R n, and there is an implicit constraint that A 0. This is a convex optimization problem as the objective and constraint functions are both convex in A and b. Evaluating the constraint function involves solving a convex maximization problem and is only tractable in certain special cases. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 17 / 28

Examples Extremal Volume Ellipsoids Minimum Volume Covering Ellipsoid Minimum volume ellipsoid covering a finite set: Suppose C = {x 1,..., x m}. The Löwner-John ellipsoid can be found by solving the problem: minimize log det A 1 subject to Ax i + b 2 1, i = 1,..., m. Norm constraint Ax i + b 2 1 equivalent to convex quadratic one Ax i + b 2 2 1. As an ellipsoid covers any set C if and only if it covers its convex hull, this problem yields the Löwner-John ellipsoid for the polyhedron P = conv{x 1,..., x m}. Minimum volume ellipsoid covering a union of ellipsoids: Suppose C = m i=1 E i, where E 1,..., E m are ellipsoids of the following form. { } E i = x : x T A i x + 2b T i x + c i 0, A i S n ++, i = 1,..., m. The Löwner-John ellipsoid can be found by solving the following convex problem. minimize log det A 1 subject to τ 1 0,..., τ m 0 A 2 τ i A i b τi b i 0 ( b τi b i ) T 1 τ i c i bt 0 b A 2, 0, i = 1,..., m with variables A 2 S n, b = Ab R n, and τ i R for i = 1,..., m. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 18 / 28

Extremal Volume Ellipsoids Minimum Volume Covering Ellipsoid Efficiency of Löwner-John Ellipsoidal Approximation Let E lj be the Löwner-John ellipsoid of a convex set C R n, which is bounded with nonempty interior, and whose center is x 0. Then, it can be shown that the Löwner-John ellipsoid, shrunk by a factor of n, lies inside C. Namely, we have Example: Polyhedron in R 2 x 0 + (1/n) (E lj x 0) C E lj. Further Extensions: If C is symmetric, then the factor of n can be improved to n. Namely, we have x 0 + ( 1/ n ) (E lj x 0) C E lj. This allows us to approximate any norm by a quadratic norm. In other words, if is any norm, C {x : x 1} is its unit ball, and E lj = { x : x T Ax 1 } (where A S n ++) is the Löwner-John ellipsoid of C, then we have z lj z ( 1/2 n z lj, where z lj z Az) T. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 19 / 28

Extremal Volume Ellipsoids Maximum Volume Inscribed Ellipsoid The Maximum Volume Inscribed Ellipsoid Let C R n be a convex set that is bounded with nonempty interior. To find the ellipsoid of maximum volume that lies inside C, we will use the parametrization E = { Bu + d : u 2 1 }, where we can assume without loss of generality that B S n ++. Then, the maximum volume inscribed ellipsoid E mvie can be obtained by solving the following convex optimization problem. Maximum Volume Inscribed Ellipsoid Problem maximize subject to log det B sup I C (Bu + d) 0, u 2 1 with variables B S n and d R n, and there is an implicit constraint that B 0. Efficiency of Ellipsoidal Inner Approximations: The ellipsoid E mvie, expanded by a factor of n about x 0, the center of C, lies inside C. Thus, E mvie C x 0 + n (E mvie x 0 ). If C is symmetric, the factor improves to n. Hence, E mvie C x 0 + n (E mvie x 0 ). Example: Polyhedron in R 2 Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 20 / 28

Examples Extremal Volume Ellipsoids Maximum Volume Inscribed Ellipsoid Maximum volume ellipsoid in a polyhedron: Suppose C = { x : a T i x b i, i = 1,..., m }. Then, we have sup I C (Bu + d) 0 Ba i 2 + a T i d b i b i, i = 1,..., m. u 2 1 The maximum volume ellipsoid inscribed in C can then be found by solving the convex problem minimize log det B 1 subject to Ba i 2 + a T i d b i, i = 1,..., m. Maximum volume ellipsoid in an intersection of ellipsoids: Suppose C = m i=1 E i, where E 1,..., E m are ellipsoids of the following form. { } E i = x : x T A i x + 2b T i x + c i 0, A i S n ++, i = 1,..., m. The maximum volume inscribed ellipsoid can be found by solving the following convex problem. minimize subject to log det B 1 ( ) T λ i c i + b T i A 1 i b i 0 d + A 1 i b i 0 λ i I B 0, i = 1,..., m, d + A 1 i b i B A 1 i with variables B S n, d R n, and λ R m. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 21 / 28

Extremal Volume Ellipsoids Affine Invariance of Extremal Volume Ellipsoids Affine Invariance of Extremal Volume Ellipsoids Affine Invariance Property of Extremal Volume Ellipsoids: Let C, D R n each be a bounded set with nonempty interior, where D is additionally convex. Also, let T R n n be any nonsingular matrix. Then, we have the following. If E lj is the Löwner-John ellipsoid of C, then the Löwner-John ellipsoid of TC is TE lj. If E mvie is the maximum volume inscribed ellipsoid of D, then the maximum volume inscribed ellipsoid of TD is TE mvie. Proof of Affine Invariance: Let E be any ellipsoid that covers C (or is inscribed in D). Then the ellipsoid TE covers TC (or is inscribed in TD). Conversely, every ellipsoid that covers TC (or is inscribed in TD) is of the form TE, where E is an ellipsoid that covers C (or is inscribed in D). Therefore, the relation Ẽ = TE gives a one-to-one correspondence between ellipsoids covering TC (or inscribed in TD) and ellipsoids covering C (or inscribed in D). Moreover, the volumes of the corresponding ellipsoids are related by the ratio det T. Hence, if E has minimum volume among ellipsoids covering C, then TE has minimum volume among ellipsoids covering TC. Similarly, if E has maximum volume among ellipsoids inscribed in D, then TE has maximum volume among ellipsoids inscribed in TD. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 22 / 28

Centering Problems Chebyshev Center Depth and the Chebyshev Center Let C R n be a bounded set with nonempty interior, and x C. The depth of x is depth(x, C) dist(x, R n \ C). Intuitively, the depth gives the radius of the largest ball (for the specific norm), centered at x, that lies in C. A Chebyshev center of C is any point of maximum depth in C: x cheb (C) = argmax x depth(x, C) = argmax dist(x, R n \ C). x Chebyshev Center of a Convex Set: When C is convex, the depth is a concave function for x C, so computing the Chebyshev center is a convex optimization problem. More specifically, suppose C is defined by a set of convex inequalities: C = {x : f 1(x) 0,..., f m(x) 0}. Then, a Chebyshev center can be found by solving the following problem. maximize R subject to g i(x, R) 0, i = 1,..., m, where gi(x, R) sup f i(x + Ru). u 1 Each function g i is convex, since it is the pointwise supremum of a family of convex functions of x and R. Evaluating g i entails solving a convex maximization problem, which may be hard. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 23 / 28

Examples Centering Problems Chebyshev Center Chebyshev center of a polyhedron: Suppose C = { x : a T i x b i, i = 1,..., m }. Then, we have g i (x, R) = sup a T i (x + Ru) = at i x + R a i b i, u 1 if R 0. Hence, the Chebyshev center can be found by solving the LP maximize subject to R a T i x + R a i b i, i = 1,..., m R 0 Euclidean Chebyshev center of intersection of ellipsoids: Suppose C is an intersection of m ellipsoids, defined by quadratic inequalities, { } C = x : x T A i x + 2b T i x + c i 0, i = 1,..., m, where A i S n ++. The Chebyshev centering problem can be shown to be the following. maximize subject to. R ( ) T λ i c i + b T i A 1 i b i 0 x + A 1 i b i 0 λ i I RI 0, i = 1,..., m, x + A 1 i b i RI A 1 i which is an SDP with variables R R, λ R m, and x R n. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 24 / 28 xcheb

Centering Problems Maximum Volume Ellipsoid Center Maximum Volume Ellipsoid Center Recall that the Chebyshev center x cheb of a set C R n is the center of the largest ball that lies in C. As an extension of this idea, we define the maximum volume ellipsoid center of C, denoted x mve, as the center of the maximum volume ellipsoid that lies in C. We can compute x mve by finding the maximum volume inscribed ellipsoid for C. To do this, we first solve the optimization problem maximize subject to log det B sup I C(Bu + d) 0, u 2 1 with variables B S n and d R n, and an implicit constraint that B 0. Then we simply set x mve = d. xmve Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 25 / 28

Centering Problems Analytic Center Analytic Center of a Set of Inequalities The analytic center x ac of a set of convex inequalities and affine equalities, f i (x) 0, i = 1,..., m, Fx = g, is defined as an optimal point for the convex optimization problem m minimize log( f i (x)) i=1, subject to Fx = g with variable x R n and implicit constraints f i (x) < 0 for i = 1,..., m. The objective of this problem is called the logarithmic barrier associated with the set of inequalities. The analytic center x ac is the point that maximizes the product (or equivalently the geometric mean) of the slacks f i (x), subject to Fx = g and f i (x) < 0. This center is not a function of the feasible set C = {x : f i (x) < 0 i, Fx = g}: two sets of inequalities can describe the same set, but have different analytic centers. The analytic center x ac is independent of affine changes of coordinates. This center is also invariant under positive scalings of the inequality functions and reparametrization of the equality constraints. In other words, if α 1,..., α m > 0, and F and g are such that Fx = g if and only if Fx = g, then the analytic center of is the same as the analytic center of α i f i (x) 0, i = 1,..., m, Fx = g, f i (x) 0, i = 1,..., m, Fx = g. Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 26 / 28

Centering Problems Analytic Center Ellipsoids from Analytic Center of Linear Inequalities The analytic center x ac of a set of linear inequalities a T i x b i, i = 1,..., m, is the solution of the unconstrained minimization problem m ) minimize φ(x) log (b i a T i x, i=1 with implicit constraint b i a T i x > 0, i = 1,..., m. It can be shown that the analytic center here defines an inscribed and a covering ellipsoid, in terms of H 2 φ(x ac). Namely, we have where P = E inner = E outer = E inner P E outer, { } x : a T i x b i, i = 1,..., m, { } x : (x x ac) T H (x x ac) 1, { } x : (x x ac) T H (x x ac) m (m 1). Here, H is the Hessian of φ(x) at x ac, given by m H = d 2 i a ia T i, d 1 i = b i=1 i a T, i = 1,..., m. i xac Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 27 / 28 xac

Centering Problems Analytic Center Analytic Center of a Linear Matrix Inequality The definition of an analytic center can be extended to sets described by generalized inequalities with respect to a cone K, if we define a generalized logarithm for K. Generalized Logarithm for a Proper Cone: Let K R q is a proper cone. We say ψ : R q R is a generalized logarithm for K if The function ψ is concave, closed, twice continuously differentiable, dom ψ = int K, and 2 ψ(y) 0 for y int K. There is a constant θ > 0 (the degree of ψ) such that for all y K 0 and all s > 0, ψ(sy) = ψ(y) + θ log s. In other words, ψ behaves like a logarithm along any ray in the cone K. For example, to find the analytic center of the LMI x 1A 1 + x na n B, we can take ψ(x) = log det X as a generalized logarithm for S p + with degree p, and solve the problem minimize log det(b x 1A 1 x na n). Andre Tkacenko (JPL) EE/ACM 150 - Lecture 17 May 29, 2012 28 / 28