Lecture: Cone programming. Approximating the Lorentz cone.

Similar documents
Lecture 6: Conic Optimization September 8

Lecture: Introduction to LP, SDP and SOCP

Agenda. 1 Cone programming. 2 Convex cones. 3 Generalized inequalities. 4 Linear programming (LP) 5 Second-order cone programming (SOCP)

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

Convex Optimization M2

Lecture 8. Strong Duality Results. September 22, 2008

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

12. Interior-point methods

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Lecture 9 Monotone VIs/CPs Properties of cones and some existence results. October 6, 2008

Optimality Conditions for Constrained Optimization

Global Optimization of Polynomials

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

Convex Optimization and l 1 -minimization

Lecture: Duality of LP, SOCP and SDP

Convex Optimization and Modeling

Additional Homework Problems

Semidefinite Programming Basics and Applications

15. Conic optimization

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

4. Convex optimization problems

Assignment 1: From the Definition of Convexity to Helley Theorem

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

Lecture: Convex Optimization Problems

SEMIDEFINITE PROGRAM BASICS. Contents

Convex relaxation. In example below, we have N = 6, and the cut we are considering

IE 521 Convex Optimization

Convexification of Mixed-Integer Quadratically Constrained Quadratic Programs

Interior Point Algorithms for Constrained Convex Optimization

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 12 Luca Trevisan October 3, 2017

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

IE 521 Convex Optimization

We describe the generalization of Hazan s algorithm for symmetric programming

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

EE 227A: Convex Optimization and Applications October 14, 2008

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

12. Interior-point methods

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

LECTURE 10 LECTURE OUTLINE

Lecture: Examples of LP, SOCP and SDP

Lecture: Duality.

EE364a Review Session 5

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

Lecture 6 - Convex Sets

What can be expressed via Conic Quadratic and Semidefinite Programming?

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

A Review of Linear Programming

Lecture Note 5: Semidefinite Programming for Stability Analysis

5. Duality. Lagrangian

Lecture 5. The Dual Cone and Dual Problem

Summer School: Semidefinite Optimization

ORF 523 Lecture 9 Spring 2016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Thursday, March 10, 2016

Convex relaxation. In example below, we have N = 6, and the cut we are considering

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Convex optimization problems. Optimization problem in standard form

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

POLARS AND DUAL CONES

Robust linear optimization under general norms

Conic Linear Programming. Yinyu Ye

2.098/6.255/ Optimization Methods Practice True/False Questions

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Structure of Valid Inequalities for Mixed Integer Conic Programs

Relations between Semidefinite, Copositive, Semi-infinite and Integer Programming

You should be able to...

Optimization for Machine Learning

Lecture 7: Convex Optimizations

l p -Norm Constrained Quadratic Programming: Conic Approximation Methods

Supplementary lecture notes on linear programming. We will present an algorithm to solve linear programs of the form. maximize.

Lecture 9 Sequential unconstrained minimization

Asteroide Santana, Santanu S. Dey. December 4, School of Industrial and Systems Engineering, Georgia Institute of Technology

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

1 Review of last lecture and introduction

Linear and non-linear programming

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

1 Robust optimization

Duality Theory of Constrained Optimization

Linear Programming Inverse Projection Theory Chapter 3

Linear Programming. Larry Blume Cornell University, IHS Vienna and SFI. Summer 2016

Preliminaries Overview OPF and Extensions. Convex Optimization. Lecture 8 - Applications in Smart Grids. Instructor: Yuanzhang Xiao

Integer Programming ISE 418. Lecture 12. Dr. Ted Ralphs

On polyhedral approximations of the second-order cone Aharon Ben-Tal 1 and Arkadi Nemirovski 1 May 30, 1999

Topic: Balanced Cut, Sparsest Cut, and Metric Embeddings Date: 3/21/2007

Lecture 20: November 1st

Continuous Optimisation, Chpt 9: Semidefinite Problems

15.083J/6.859J Integer Optimization. Lecture 10: Solving Relaxations

Linear Programming Duality P&S Chapter 3 Last Revised Nov 1, 2004

Applications of Linear Programming

BCOL RESEARCH REPORT 07.04

Chapter 1. Preliminaries

Advances in Convex Optimization: Theory, Algorithms, and Applications

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 4

ICS-E4030 Kernel Methods in Machine Learning

Convex Optimization Boyd & Vandenberghe. 5. Duality

Optimization methods NOPT048

On self-concordant barriers for generalized power cones

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Lecture 7: Semidefinite programming

Transcription:

Strong relaxations for discrete optimization problems 10/05/16 Lecture: Cone programming. Approximating the Lorentz cone. Lecturer: Yuri Faenza Scribes: Igor Malinović 1 Introduction Cone programming is a broad generalization of linear programming. We focus on two well studied subclasses: second-order cone programming (SOCP) and semidefinite programing (SDP). In what follows, we will provide an overview of cone programming, skipping most proof, that can be found e.g. in []. As seen in previous classes, there are two equivalent ways to look at a polyhedral cone K R n : K = {x : Ax 0} {r 1,..., r m } R n : K = {x = We generalize this notion as follows. m λ i r i, λ 0}. Definition 1. Let V be a real finite-dimensional Hilbert space and K V be a nonempty closed set. K is called a closed convex cone if for any nonnegative scalars α, β and any x, y K one has that αx + βy belongs to K. i=1 Examples: 1. K = R n + = {x : Ix 0};. Lorentz cones (also called ice cream cones or second-order cones ) L n = {(x, t) R n R : x t}; t x x 1 3. Semidefinite cones S n + = {X R n n : X = X T, v T Xv 0 v R n }. 1

The fact that the first two are closed convex cone (with V = R n ) is immediate to check. For the last example, we defer the proof to next lecture. Given closed convex cones L and K, the corresponding cone programming (CP) problem is max c, x s.t. Ax b L x K where, is the inner product of the Hilbert space. Example of those problems are: LP: SOCP: max c, x max c, x s.t. Ax = b s.t. Ax b x 0 A i x b i c i x d i i = 1,..., m Here SOCP stands for second-order cone programming. It is immediate to check that LP can be formulated as a cone programming problem, setting L = {0} and K = R n +. Standard transformations also show that SOCP can be formulated as a cone programming problem, by taking L = {0} and K a Lorentz cone of appropriate size. In order to define the dual cone program, it is useful to introduce the notion of a dual cone. Definition. Let K V be a closed convex cone. Its dual cone is given by K := {y V : x, y 0 x K}. Exercise 3. If K is a closed convex cone then K is also a closed convex cone. We can develop a duality theory for cone programming in an analogous way to linear programming: CP: LP: (P ) max c, x max c, x s.t. Ax b L s.t. Ax = b x K x 0 (D) min b, y min b, y s.t. A T y c K s.t. A T y c y L The difference between general CP and LP is that Strong duality does not necessarily hold, i.e. it can be the case that both the dual and the primal are feasible but their optimal values do not coincide. The most famous sufficient conditions are given bellow, and we will need them in next lecture. Theorem 4 (Slater condition for Strong duality). Let (P ) be the primal problem with L = {0} and suppose there exists a primal feasible point x int(k). Then (D) is feasible and the optima have the same value.

Modelling problems as SOCPs A class of problems that can be remodelled in terms of SOCP are quadratically constrained convex quadratic programs (QCCQP): min x T Q T Qx b }{{} T x Qx,Qx = Qx s.t. x T Q T i Q i x }{{} b T i x c i i = 1,..., m. Q ix The optimum value of the above program coincides with the optimum value of the SOCP: min s.t. t Qx b T x+t 1 Q i x b T i x+t 1 bt x+t+1 b, T i x+t+1 i = 1,..., m. It is easy to see that the first constraint can be squared to obtain the following equivalent form: and similarly for the others Qx 1 (bt x + t) 1 (bt x + t) Qx b T x t Q i x b T i x c i. Hence, if (t, Q, {Q i } i [m] ) is an optimum solution to the SOCP above, then (Q, {Q i } i [m] ) is an optimum solution to the QCQP problem and t its optimum value. For another interesting application, imagine a linear program with some incertitude on the coefficient vector of the constraints. For each inequality a T x b, we know that a can take any value in the ellipsoid E(c, R) := {c + Ru : u 1}. We want therefore our vector x to satisfy and the optimum is reached for u = which is again a second-order constraint. b max a E(c,R) at x = c T x + max u: u 1 xt Ru. RT x R T x having the value xt R R, T x R T x = RT x. So finally b c T x + R T x. 3

3 Approximating Second-order cones Second-order cone programming seems to be much more general then linear programming. Surprisingly, the feasible region of an SOCP program can be well-approximated by a polyhedral cone 1. Definition 5. Let Q be a feasible region of a SOCP, i.e. Q := {x : Ax b, A i x b i c i x d i i = 1,..., m} and let Q ɛ := {x : Ax b, A i x b i (c i x d i )(1 + ɛ) i = 1,..., m}. A polychedral cone P is an ɛ-approximation of Q if Q P Q ɛ. Our goal here is to find an ɛ-approximation of Q with small extension complexity. perform the following two steps: 1. Give an ɛ-approximation of the 3-dimensional Lorentz cone L ; We will. Show how this approximation can be extended to any Lorentz cone, and then to an arbitrary Q as in Definition 5 above. 3.1 An ɛ-approximation of L Recall that L = {(x, t) R R : x t}. Given x = ( x 1, x, t) R 3, we would like to understand whether it lies in L. Assume wlog t = 1, x 0. Ideally we would like to rotate the point ( x 1, x, 1) to ( x 1, 0, 1). Then x L iff x 1 1. Unfortunately, this rotation cannot be done with a linear map, but there is a simple linear transformation that reduces the angle between x and the x 1 axis by at least half. Iterating, we can make this angle arbitrarily small. We can now check whether the rotated point x has x 1 1. If not, we know that x / L. If yes, then still it may be that x / L, but x L ɛ for an appropriate small ɛ, since the component x is very close to 0. x ( x 1, x ) ( x 1, 0) x 1 cut at t = 1 The intuition above is formalized by the following polyhedron. T := { (x 1, x, ξ 0, η 0,..., ξ ν, η ν, t) : ξ 0 x 1, (1) η 0 x () ξ j = cos( j+1 )ξ j 1 + sin( j+1 )η j 1, j = 1,..., ν (3) η j sin( j+1 )ξ j 1 + cos( j+1 )η j 1, j = 1,..., ν (4) ξ ν t, (5) η ν tan( ν+1 )ξ ν}. (6) where (ξ 0, η 0 ) rotates (x 1, x ) so that x 1, x 0, and we have two additional variables ξ j, η j for each iteration j [ν]. Note that the fact that only appears in inequalities allows as to deduce via standar techniques that T as a polyhedron. 1 Note that this does not imply in general that the optimum solution of an SOCP can be well-approximated by an LP. For instance, Q may be empty, but Q ɛ and P may not. 4

Theorem 6. proj x1,x,t(t ) is a polyhedral ɛ-approximation to L for ɛ = Ω( 1 4 ν ) and xc(t ) = O(log 1 ɛ ). Proof. The statement on the extension complexity is immediate, so we focus on the rest of the proof. Part I: Q proj x1,x,t(t ). Let ( x 1, x, t) Q, and extend it to a point (x 1, x, ξ 0, η 0,..., ξ ν, η ν, t) by setting to equality constraints (1) (4). Note that this extension is unique. Before showing that this point is in T, let us study how the sequence of points (ξ 0, η 0 ),..., (ξ ν, η ν, ) looks like. We set (ξ 0, η 0 ) = ( x 1, x ), thus reflecting the initial point to the nonnegative orthant. So the angle between (ξ 0, η 0 ) and the x 1 axis verifies α(ξ 0, η 0 ). Standard trigonometry implies that, if there was no modulus in (4), then (ξ 1, η 1 ) would be exactly the clockwise rotation by 4 of (ξ 0, η 0 ). Through the modulus we are actually reflecting over the x 1 -axis if η 1 < 0. So α(ξ 1, η 1 ) 4, and we can repeat the argument, deducing α(ξ ν, η ν ). ν+1 It remains to show that the last two constraints hold after repeating the operation ν times. (5) holds since: ξ ν (ξ ν, η ν ) = = (ξ 0, η 0 ) = ( x 1 + x ) 1/ t. Using that α(ξ ν, η ν ) ν+1 and by the definition ην ξ ν = tan(α(ξ ν, η ν )) we obtain (6): η ν = tan(α(ξ ν, η ν ))ξ ν tan( ν+1 )ξ ν. Part II: proj x1,x,t(t ) Q ɛ. Given ( x 1, x, ξ 0, η 0,..., ξ ν, η ν, t) T, we would like to prove that ( x 1, x, t) Q ɛ. Since in each of (1),(),(4), we have, while (3) is an equation, the norm gets never decreased. We conclude ( x 1, x ) ( ξν, η ν ) = ( ξ ν + η ν) (6) 1/ ξ ν (1 + tan ( ))1/ ν+1 where we used the trigonometric identity 1 + tan (γ) = 1/ cos (γ). (5) 1 t (1 + cos( ) 1 ), }{{ ν+1 } O( 1 4 ν ) x 0,1 x 0, x 0,3 x 0,4 (x 0,1 + x 0,) 1/ x 1,1 (x 0,3 + x 0,4) 1/ x 1, x 1,1 x 1, (x 1,1 + x 1,) 1/ t x,1 = t 3. Extending the ɛ-approximation Assume wlog n = k for some k N. We show that we can give an extended formulation for L n {(x 0,1,..., x 0,n ) R n R : (x 0,1,..., x 0,n ) t} 5

using O(n) L cones. Rename original variables as x 0,1,..., x 0,n. Note that (x 0,1,..., x 0,n, t) L n if and only if there exits a value x 1,1 such that (x 1,1, x 0,1,..., x 0,n, t, ) satisfy (x 1,1, x 0,3, x 0,4,..., x 0,n, t) L n 1 and (x 0,1, x 0,, x 1,1 ) L. We can now repeat the argument: split all the original variables into pairs (x 0,j 1, x 0,j ) for j [n/] and associate to each pair a new 1st level variable x 1,j. Apply the same procedure to 1st level variables, introducing nd level variables, etc. We construct therefore a binary tree where each i-th level variable has as children two (i 1)-th level variables. The same structure is preserved up to the level log (n) which has a single node x log (n),1 = t. At each level of the tree we have a ɛ-approximation of the L cones with the construction from the previous section. This gives in total a ((1 + ɛ) log (n) 1)-approximation for L n. Last, observe that, if P n is a polyhedral ɛ-approximation to L n, then {x : Ax b, (A i x, c i x b i ) P ki for i = 1,..., m} is a polyhedral ɛ-approximation to Q as in Definition 5, where k i is the number of row of matrix A i. 4 Notes Results from Section 3 appeared in [1]. It is also shown there that, by carefully choosing different ɛ for different levels, one can obtain ɛ-approximation of L n (hence of Q as in Definition 5) with slighty smaller extension complexity. References [1] A. Ben-Tal and A. Nemirovski. On polyhedral approximations of the second-order cone. Math. Oper. Res., 6():193 05, May 001. [] B. Gärtner and J. Matousek. Approximation Algorithms and Semidefinite Programming. Springer-Verlag, Berlin, Heidelberg, 01. 6