Tamás Terlaky George N. and Soteria Kledaras 87 Endowed Chair Professor. Chair, Department of Industrial and Systems Engineering Lehigh University

Similar documents
Tamás Terlaky George N. and Soteria Kledaras 87 Endowed Chair Professor. Chair, Department of Industrial and Systems Engineering Lehigh University

Disjunctive conic cuts: The good, the bad, and implementation

A conic representation of the convex hull of disjunctive sets and conic cuts for integer second order cone optimization

Second-order cone programming

Disjunctive Cuts for Cross-Sections of the Second-Order Cone

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

BCOL RESEARCH REPORT 07.04

15. Conic optimization

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Interior Point Methods for Nonlinear Optimization

10 Numerical methods for constrained problems

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Interior Point Methods for Mathematical Programming

III. Applications in convex optimization

Lifting for conic mixed-integer programming

IMPLEMENTATION OF INTERIOR POINT METHODS

Sparse Optimization Lecture: Basic Sparse Optimization Models

Convex hull of two quadratic or a conic quadratic and a quadratic inequality


Cuts for Conic Mixed-Integer Programming

Research overview. Seminar September 4, Lehigh University Department of Industrial & Systems Engineering. Research overview.

On the Sandwich Theorem and a approximation algorithm for MAX CUT

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

The Ongoing Development of CSDP

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma

Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A.

Interior Point Methods in Mathematical Programming

4TE3/6TE3. Algorithms for. Continuous Optimization

Semidefinite Programming, Combinatorial Optimization and Real Algebraic Geometry

Two-Term Disjunctions on the Second-Order Cone

Lecture Note 5: Semidefinite Programming for Stability Analysis

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Summer School: Semidefinite Optimization

Sparse Matrix Theory and Semidefinite Optimization

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

A ten page introduction to conic optimization

Lecture: Introduction to LP, SDP and SOCP

Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming*

A priori bounds on the condition numbers in interior-point methods

Lecture: Algorithms for LP, SOCP and SDP

Robust and Optimal Control, Spring 2015

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

Semidefinite Programming

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Split Cuts for Convex Nonlinear Mixed Integer Programming

A new primal-dual path-following method for convex quadratic programming

The Solution of Euclidean Norm Trust Region SQP Subproblems via Second Order Cone Programs, an Overview and Elementary Introduction

The moment-lp and moment-sos approaches

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

The Ellipsoid Algorithm

Structure of Valid Inequalities for Mixed Integer Conic Programs

4TE3/6TE3. Algorithms for. Continuous Optimization

Cuts for mixed 0-1 conic programs

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Lecture: Examples of LP, SOCP and SDP

More First-Order Optimization Algorithms

Structure in Mixed Integer Conic Optimization: From Minimal Inequalities to Conic Disjunctive Cuts

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

18. Primal-dual interior-point methods

A New Class of Polynomial Primal-Dual Methods for Linear and Semidefinite Optimization

15-780: LinearProgramming

On implementing a primal-dual interior-point method for conic quadratic optimization

A full-newton step feasible interior-point algorithm for P (κ)-lcp based on a new search direction

Interior-Point Methods

Introduction to Semidefinite Programs

Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming*

Interior Point Algorithms for Constrained Convex Optimization

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

Split Cuts for Convex Nonlinear Mixed Integer Programming

On Valid Inequalities for Quadratic Programming with Continuous Variables and Binary Indicators

A Full Newton Step Infeasible Interior Point Algorithm for Linear Optimization

The Q Method for Second-Order Cone Programming

2.1. Jordan algebras. In this subsection, we introduce Jordan algebras as well as some of their basic properties.

Continuous Optimisation, Chpt 9: Semidefinite Problems

Module 04 Optimization Problems KKT Conditions & Solvers

Conic Linear Programming. Yinyu Ye

The Q Method for Symmetric Cone Programmin

3. Linear Programming and Polyhedral Combinatorics

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

A path following interior-point algorithm for semidefinite optimization problem based on new kernel function. djeffal

4. Algebra and Duality

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

SEMIDEFINITE PROGRAM BASICS. Contents

Continuous Optimisation, Chpt 9: Semidefinite Optimisation

Primal-dual IPM with Asymmetric Barrier

Convex Optimization in Classification Problems

A direct formulation for sparse PCA using semidefinite programming

Continuous Optimisation, Chpt 7: Proper Cones

A Quick Tour of Linear Algebra and Optimization for Machine Learning

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Convex Optimization M2

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

CO 250 Final Exam Guide

Lecture 17: Primal-dual interior-point methods part II

ORF 523 Lecture 9 Spring 2016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Thursday, March 10, 2016

Transcription:

5th SJOM Bejing, 2011 Cone Linear Optimization (CLO) From LO, SOCO and SDO Towards Mixed-Integer CLO Tamás Terlaky George N. and Soteria Kledaras 87 Endowed Chair Professor. Chair, Department of Industrial and Systems Engineering Lehigh University

Outline Motivation: Data uncertainty in linear inequalities-robust LO Data uncertainty in linear inequalities Ellipsoidal uncertainty set norm constraint Eigen- and singular value optimization problems Relaxing integer variables Conic Linear Optimization (CLO): General Convex Cones Second Order Conic Optimization (SOCO) Semidefinite Optimization (SDO) Interior Point Algorithms for SOCO and SDO: MISOCO: Mixed Integer Second Order Conic Optimization Sisyphus got stuck with a suboptimal solution; don t let it happen to you! http://www.optimize.com 1

Robust Linear Optimization Classic Polyhedral (scenario) approach (P ) min c T x a j b j n j i=1 s.t. a T j x b j 0 j n j = ai j b i j i=1 λ i j ai j b i j T x 1 λ i j n j i=1 0 Let (a j, b j ) be uncertain, it is coming from a polyhedral set (e.g. convex combination of scenario data points): λ i j = 1, λ i j 0 for all n j i=1 The inequality a T j x b j must be true for all possible values of (a T j, b j): λ i j = 1, λ i j 0 Infinitely many constraints! iff [a i j] T x b i j 0 for i = 1,..., n Finally the problem stays linear as: j (RP ) min c T x s.t. [a i j ]T x b i j 0 for i = 1,..., n j j Disadvantages: Huge number of linear inequalities Polyhedral uncertainty set not realistic. 2

Robust Linear Optimization a0 j b 0 j (P ) min c T x s.t. a T j x b j 0 j a j = a0 j + P u b u IRk, u T u 1 j T + P u x 1 b 0 j Let (a j, b j ) be uncertain, it is coming from an ellipsoid (e.g. level set of a distribution): The inequality a T j x b j must be true for all possible values of (a T j, b j): 0 u : u T u 1 iff [a 0 j ] T x b 0 j + min (P u)t x u T u 1 1 0 [a 0 j ] T x b 0 j P T x 1 0 2 This is a nondifferentiable norm constraint: (See second order cones) P T x [a 0 j 1 ]T x b 0 j. 2 Single nonlinear, norm-constraint! 3

Eigenvalue Optimization Given n n symmetric matrices A 1,..., A m. Problem: Find a nonnegative combination of the matrices that has the maximal smallest eigenvalue. Solution: max λ m i=1 A i y i λi is positive semidefinite y i 0 i = 1,..., m Problem: Find a nonnegative combination of the matrices that has the smallest maximal eigenvalue. m λi A i y i Solution: min λ i=1 is positive semidefinite y i 0 i = 1,..., m The semidefiteness constraint is not differentiable, not easy to calculate when formulated by explicit functions, e.g., min-eigenvalue, determinant (of minors). See Semidefinite Optimization. 4

Relaxing Binary Variables Given z 1,..., z n binary, i.e., {0, 1} variables with other, e.g., linear constraints. Problem: Find convex, continuous relaxations of the binary constraints. Old solution: Let 0 z i 1 for all i = 1,..., n and use branch and bound, branch and cut schemes. New opportunity to get tighter relaxations: Let x i = 2z 1 1 2 for all i = 1,..., n. Thus gives x i as a { 1, 1} variable. Now n = n x 2 i = x T x = Tr(x T x) = Tr(xx T ) = Tr(X), i=1 where X = xx T is a rank-1 positive semidefinite matrix with diag(x) = e. Thus, for x i { 1, 1} i we may use the relaxation: X ii = 1 i and X is positive semidefinite The semidefiteness constraint is not differentiable, not easy to calculate when formulated by explicit functions, e.g., min-eigenvalue, determinant (of minors). See Semidefinite Optimization. 5

Stability of Optimal Power Flow The stability of the Power Flow is ensured by a lower bound on the singular value of the Jacobian J pf of the power flow equations: σ min (J pf ) σ cpf equivalently λ min (J pf Jpf T ) σ cpf by substitution X J pf Jpf T = 0 and X σ cpf I is positive semidefinite Again a semidefinite constraint! One may want maximize σ cpf. 6

New/Old Convex Optimization Problems Cone Linear Optimization Problems Primal-dual pair of CLO problems is given as (P ) min c T x (D) max b T y s.t. Ax b C 1 s.t. c A T y C 2 x C 2 y C1, where b, y IR m, c, x IR n, A : m n matrix, C 1, C 2 are convex cones and Ci = {s IRn : x T s 0, x C i } are the dual cones for i = 1, 2. These are solvable efficiently (in polynomial time) by using Interior Point Methods. LO is based on polyhedral cones. Be careful! Perfect duality, strict complementarity lost. Are all convex cones good??? NOT 7

New Convex Optimization Problems Second Order Conic Optimization (SOCO) The second order cone in IR n is defined as S2 n := x IRn : x 2:n = n x 2 i x 1 i=2. The name ice cream cone is coming from the 3-dimensional shape of the cone. The second order cone is self-dual: (S n 2 ) = S n 2. Optimization problems, where cones C 1 and C 2 are polyhedral or products of second order cones, are second order cone optimization (SOCO) problems. Significance Norm minimization, robust optimization, quadratic, and thus portfolio optimization. 8

SOCO Optimality The primal-dual SOCO problem is defined as (SP ) min c T x s.t. Ax = b, (SD) max b T y s.t. A T y + s = c x k j=1 Sn j 2 s k j=1 Sn j 2. x T = ((x 1 ) T,..., (x j ) T,..., (x k ) T ) IR n ; and s T = ((s 1 ) T,..., (s j ) T,..., (s k ) T ) IR n. Ax = b, Optimality: x k j=1 Sn j 2, AT y + s = c, s k j=1 Sn j 2 (x j ) T s j = 0 x j s j = 0 Arr(x j )s j = Arr(s j )x j = 0 j Here we have used the notation: u v = u T v u 1 v 2:n + v 1 u 2:n and Arr(u) = u 1 u 2... u n u 2 u 1.... u n u 1 9

Notes on SOCO Convex (conic) optimization problem with vectors Vector calculus not associative Duality gap may exists (next page) Strong duality with interior point (Slater) condition Second order cones cannot be combined into larger second order cones, i.e., S n 1 2 Sn 2 2 Sn 1+n 2 2 Generalization of LO: S 1 2 = IR1 + Rotated cone x 2:n 2 x 0 x 1, x 0, x 1 0 n ( ) x 2 x0 i + x 2 ( 1 x0 x 1 2 2 i=2 Efficiently solvable by IPMs. ) 2 10

SOCO: Duality gap example Primal Problem (SP ) min x 2 s.t. x 1 x 3 = 0, x 2 2 + x2 3 x 1 (SD) max 0 y s.t. y + s 1 = 0 0 y + s 2 = 1 y + s 3 = 0 s 2 2 + s2 3 s 1 Dual Problem Primal Optimal solutions: x 1 = x 3 0; x 2 = 0 Dual problem is infeasible!! Slater not satisfied = Zero duality gap may not hold 11

New Convex Optimization Problems Semidefinite Optimization I The semidefinite cone in IR n n is defined as S n := { X IR n n : X = X T, z T Xz 0 z IR n} i.e. the matrices X are symmetric and positive semidefinite, denoted as X 0. The semidefinite cone is self-dual: (S n ) = S n. Optimization problems where the cones C 1 and C 2 are either polyhedral, second order or semidefinite cones are called semidefinite optimization (SDO) problems. 12

New Convex Optimization Problems Semidefinite Optimization Let A i, i = 1,, n and C, X be n n symmetric real matrices, b, y IR m and let Tr( ) denote the trace of a matrix. The primal-dual SDO problem is defined as (SP ) min Tr(CX) s.t. Tr(A i X) b i 0, i X 0 (SD) max b T y s.t. Optimality: Tr(CX) b T y = Tr(XS) = 0 XS = 0 Significance C m i=1 A i y i 0 y 0. Robust optimization, eigenvalue and singular value optimization Linear matrix inequalities, trust design Convex relaxation of nonconvex/integer problems 13

Notes on SDO Convex (conic) optimization problem with matrices Matrix calculus not commutative Product of symmetric matrices is not smmetric Duality gap may exists Strong duality with interior point (Slater) condition Semidefinite cones can be combined into larger semidefinite cones, i.e., S n 1 S n 2 S n 1+n 2 Generalization of LO: S 1 = IR 1 + diagonal matrices Not proper generalization of SOCO: x S n 2 Arr(x) Sn, BUT arrow-head structure cannot be preserved Efficiently solvable by IPMs. 14

LO Linear Optimization v/s Conic LO Conic LO linear objective linear equality constraints linear inequality constraints perfect duality strictly complementary opt.sol. Euclidean linear algebra x T s = 0 xs = 0 linear objective linear equality constraints conic inequality constraints perfect duality only with IPC maximally complementary opt.sol. matrix and Jordan algebra x T s = 0 x s = 0 (SOCO) Tr(XS) = 0 XS = 0 (SDO) Tr(XS) = 0 X 1 2SX 1 2 = S 2XS 1 1 2 = 0 (P XP T ) 1 2(P T SP 1 )(P XP T ) 1 2 = µi (P SP T ) 1 2(P T XP 1 )(P SP T ) 1 2 = µi 15

Solvability of CLO problems Use IPMs Classic Linear Optimization Large scale LO problems are solved efficiently. High performance packages, like (CPLEX, GuRoBi, XPRESS-MP, MOSEK) offer simplex and interior point solvers as well. Problems solved with 10 7 variables. SOCO and SDO Polynomial solvability established. Traditional software is unable to handle conic constraints. Specialized software is developed. (SeDuMi, SDPA, SDPT3, CSDP, DSDP, SDPpack, MOSEK etc.) SOCO: MOSEK - commercial SDO: SDPA, CSDP, DSDP LO-SOCO-SDO: SeDuMi, SDPT3 SOCO: Problems solved with O(10 6 ) variables. SDO: solved with O(10 4 ) dimensional matrices. http://sedumi.ie.lehigh.edu 16

The Primal Dual LO Problems, Central Path The primal-dual LO problems is given as: min c T x max b T y Ax = b, x 0, A T y + s = c, s 0, where c, x, s R n, b, y R m, A R m n, rank(a) = m. Optimality conditions and the central path are given as: Ax = b, x 0, A T y + s = c, s 0, Ax = b, x 0, A T y + s = c, s 0, xs = 0, xs = µe, where e = (1,..., 1) T R n. We assume that the Interior Point Condition holds. 17

Primal-Dual Search-directions for LO The central path and the Classical Newton direction: Ax = b, x 0, A T y + s = c, s 0, xs = µe. Scaled Newton direction: Āp x = 0, Ā T y + p s = 0, p x + p s = v 1 v A x = 0, A T y + s = 0, s x + x s = µe xs, Proximity Functions: ( ) n v 2 Ψ(v) = i 1 log v i 2 i=1 Ψ(v) = 1 2 v v 1 2. where Ā = 1 µ AV 1 X, V = diag (v), X = diag (x) with v := xs µ, v 1 := µ xs, p x := v x x, p s := v s s. 18

The Newton System for SDO (SP ) min Tr(CX) s.t. Tr(A i X) = b i, i X 0 (SD) max b T y s.t. m i=1 A i y i + S = C S 0. The Newton System for the NT-direction: P = S 1 2(S 1 2XS 1 2) 1 2S 1 2 m i=1 Tr(A i X) = b i, i X 0 A i y i + S = C S 0 Tr(XS) = 0 XS = µi H P ( ) is a symmetrization: m i=1 Tr(A i X) = 0, i A i y i + S = 0 X S + XS = µi XS X + X SS 1 = µs 1 X H P (X S + XS) = µi H P (XS) 19

The Newton System for SOCO (SOP ) min c T x s.t. Ax = b, x k j=1 Sn j 2 (SOD) max b T y s.t. A T y + s = c s k j=1 Sn j 2. The Newton System for the NT-direction in arrow-head formulation: P = S 1 2(S 1 2XS 1 2) 1 2S 1 2 Ax = b, i x k j=1 Sn j 2 A x = 0, A T y + s = c, s k j=1 Sn j 2 A T y + s = 0, x j s j = 0 x j s j = µe j x j s j + x j s j = µe j x j s j H P (x j s j + x j s j ) = µe j H P (x j s j ) H P ( ) is a symmetrization operator. 20

Primal-Dual Interior Point Methods with small and large updates Input: A proximity parameter τ; an accuracy parameter ϵ > 0; an update parameter 0 < θ < 1; a variable damping factor α; (x 0, s 0 ), µ 0 = 1 s.t. Ψ(v 0 ) τ. begin x := x 0 ; s := s 0 ; µ := µ 0 ; while nµ ϵ do begin µ := (1 θ)µ; while Ψ(v) τ do begin Calculate x, s; Do line search for Ψ(v(α)); x := x + α x; s := s + α s; end end end 21

Complexity of IPMs for LO Method Practice Large update Small update θ adaptive 1 1/100 1/ n Iter. bound max 100 O(ϑ log ϑ ϵ ) O( ϑ log ϑ ϵ ) Performance Efficient Efficient Very poor Almost constant (< 100) number of iterations in practice! CLO ϑ = cost/iteration LO # of variables O(n 3 ) sparse SOCO # of second order cones O(n 3 )+update SDO dimension of matrix X O(n 2 m 3 ) dense 22

Several IPM Solvers for CLO Problems What made this major advance possible? Advances in Computers and Software Computers processor speed memory disk space floating point arithmetic architecture (cash...) Software component/algorithms presolve LINEAR ALGEBRA sparse factorizations symmetric square root IPMs, predictor-corrector dense and sparse versions SeDuMi, SDPT3 SDPA-xxx tuned to all three cones CSDP, DSDP tuned to SDO only MOSEK, CPLEX are commercial solves for SOCO. Parallel implementations exist coming. MODELING languages YALMIP (Löfberg); CVx (Boyd/Grant). 23

Further notes Norm and convex quadratic (including portfolio) optimization prob s can be solved with almost the same efficiency as LO. Efficient tools to eigenvalue, singular value optimization, LMI s CLO based approximation algorithms for nonconvex and combinatorial optimization problems. Lots of activity in exploring special structure of conic problems and developing modeling systems that support conic formulation First HPC-massively parallel implementations Cheap first order methods for very large scale SDO problems. Warm start and decomposition/cutting plane algorithms. http://sedumi.ie.lehigh.edu Now: Adding logarithmic objective. 24

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work Disjunctive Conic Cuts for Mixed Integer Second Order Cone Optimization (MISOCO) Tamás Terlaky Joint work with: Pietro Belotti, Julio C. Góez, Imre Pólik, Ted Ralphs With thanks for the AFOSR Award #: FA9550-10-1-0404

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work CONTENTS Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work MIXED INTEGER SECOND ORDER CONE OPTIMIZATION (MISOCO) minimize: c T x subject to: Ax = b x K x Z d R n d, (MISOCO) where, A R m n, c R n, b R m L n = {x x 1 x 2:n } K = L n 1 1 Ln k k Rows of A are linearly independent

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work OBJECTIVES Obtain the convex hull after applying a linear disjunction to a Mixed Integer Second Order Conic Optimization (MISOCO) problem. Design Disjunctive Conic Cuts for MISOCO.

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work PREVIOUS WORK Atamtürk and Narayanan (2010), conic cuts for general MISOCO problems. Drewes (2009), nonlinear cuts for 0-1 MISOCO problems. Krokhmal and Soberanisin(2010), Drewes (2009), Vielma et al. (2008), branch and bound algorithm based on linear outer approximations for Second Order Cones. Drewes (2009), Atamtürk and Narayanan (2009), lifting techniques for MISOCO problems. Çezik and Iyengar (2005), cuts for mixed 0-1 conic programming. Stubbs and Mehrotra (1999), lift-and-project method for 01 mixed convex programming.

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work APPLICATIONS Turbine balancing problems can be modeled as MISOCOs, White (1996). The euclidean Steiner tree problem can be formulated as a MISOCO, Fampa and Maculan (2004) Computer Vision and Pattern Recognition, Kumar, Torr, and Zisserman (2006). Cardinality-constrained portfolio optimization problems, Bertsimas and Shioda (2009).

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work SINGLE CONE PROBLEM Let us consider the special case: minimize: c T x subject to: Ax = b x L n (MISOCO) x Z d R n d, This problem has a single second order cone All the variables are in the single second order cone

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work STEP 1: SOLVE THE RELAXED PROBLEM Find the optimal solution x for the continuous relaxation of the MISOCO problem minimize: 3x 1 +2x 2 +2x 3 +x 4 subject to: 9x 1 +x 2 +x 3 +x 4 = 10 (x 1, x 2, x 3, x 4 ) L 4 x 4 Z. Relaxing the integrality constraint we get the optimal solution: x = (1.36, 0.91, 0.91, 0.45), with and optimal objective value: z = 0.00.

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work STEP 2: FIND A DISJUNCTION a T x β a T x β VIOLATED BY x = (1.36, 0.91, 0.91, 0.45) The disjunction x 4 1 x 4 0 is violated by x 1.5 1 0.5 x 4 0 0 0.5 x 4 1 1.5 2 2.5 3 2 x 0 3 Relaxed optimal solution x 4 1 2 2 1.5 1 0.5 0 0.5 1 1.5 2 x 2

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work STEP 3: APPLY THE DISJUNCTION AND CONVEXIFY min: 3x 1 +2x 2 +2x 3 +x 4 s.t: 9x 1 +x 2 +x 3 +x 4 = 10 0.04x 2 0.04x 3 3.56x 4 +x 5 = 10.14 6.28x 2 6.28x 3 +0.14x 4 +x 6 = 1.65 6.36x 2 6.36x 3 +x 7 = 0 (x 1, x 2, x 3, x 4 ) L 4 (x 5, x 6, x 7 ) L 3 x 4 Z. The constraints in red represent the disjunctive conic cut. An integer optimal solution is obtained after adding one cut: x = (1.32, 0.93, 0.93, 0.00, 10.06, 10.06, 0.00), with and optimal objective value: z = 0.24.

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work CONVEX HULL OF THE INTERSECTION OF A DISJUNCTION AND A CONVEX SET Consider a closed convex set E and two halfspaces H 1 = {x R n : a x α} and H 2 = {x R n : b x β}, such that they do not intersect inside E, i.e., E H 1 H 2 =. Denote H = 1 = {x Rn : a x = α}, and H = 2 = {x Rn : b x = β}. If a convex cone K s.t. H = 1 E = K H= 1 and H= 2 E = K H= 2 are bounded, then conv(e (H 1 H 2 )) = E K.

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work INTERSECTION OF AN AFFINE SPACE AND A SECOND ORDER CONE Consider an affine subspace H = {x Ax = b} and x 0 H. Let H A be s.t rank([h, A T ]) = n, & columns of H are orthonormal. We can write H = {x x = x 0 + Hz, z R}. Then, there exist a matrix Q R n m n m, q R n m, ρ R, s.t. H L n = {y x = x 0 + Hz with z Qz + 2q z + ρ 0}. Further, Q has at most one negative eigenvalue. Define a quadric as the set Q = {z z Qz + 2q z + ρ 0}, which we also denote as Q = (Q, q, ρ).

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work UNI-PARAMETRIC FAMILY OF QUADRICS Q(τ) Given two hyperplanes H 1 = {z a 1 z = α 1} and H 2 = {z a 2 z = α 2}. Let Q = (Q, q, ρ) be a quadric where Q is positive definite. The family of quadrics having the same intersection with H 1 and H 2 as the quadric Q is parametrized by τ R as Q(τ), where Q(τ) = Q + τ a 1a T 2 + a 2a T 1 ω q(τ) = q τ α 2a 1 + α 1 a 2 ω ρ(τ) = ρ + 2τ α 1α 2 ω, where ω = { 2a T 1 a 2 if a T 1 a 2 0 1 if a T 1 a 2 = 0.

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work UNI-PARAMETRIC FAMILY OF QUADRICS Q(τ) Sequence of quadrics x Q(τ)x + 2q(τ) x + ρ(τ) 0, for 106.863 τ 1617

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work UNI-PARAMETRIC FAMILY OF QUADRICS Q(τ) Range Description τ = 8.9946, τ = 1617 Paraboloid τ = 106.863, τ = 9.581 Cones 8.9946 < τ < 1617 Ellipsoids τ > 1617 Two sheets hyperboloids 106.863 < τ < 8.9946 One sheet hyperboloids τ < 106.863, Two sheets hyperboloids Behavior of the quadrics for different ranges of τ

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work THE DISJUNCTIVE CONIC CUT FOR PARALLEL DISJUNCTIONS Theorem Let A 1 = {x a 1 x = α 1} and A 2 = {x a 2 x = α 2}, be two parallel hyperplanes where a 1 = γa 2. The disjunctive conic cut is the quadric generated by Q(ˆτ) = (Q(ˆτ), q(ˆτ), ρ(ˆτ)), where ˆτ is the larger root of equation q(τ) Q(τ)q(τ) ρ(τ) = 0.

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work y y OUR DISJUNCTIVE CONIC CUT IS NEW Atamtürk and Narayanan designed a conic mixed integer rounding inequality. Our disjunctive conic cut is different, sometimes stronger. Consider the problem: minimize: x y subject to: x +y +2t = α (x 4 3 )2 + (y 1) 2 t x Z, y R Relaxed Optimal Solution 1.5 1 Conic mixed integer 0.5 rounding inequality 0 0.5 Disjunctive conic cut 1 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 x Case with α = 4.66 4 Relaxed Optimal Solution 3 2 1 In this particular example our disjunctive conic cut is stronger. 0 1 Disjunctive conic cut Conic mixed integer rounding inequality 2 1 0 1 2 3 4 x Case with α = 8

Introduction Disjunctive Conic Cuts for MISOCO Formulation of a Disjunctive Conic Cut Conclusions and Future Work CONCLUSIONS We developed a new disjunctive conic cut for MISOCO. It is algebraically simple to find the disjunctive conic cut for MISOCO problems. Next steps Develop disjunctive conic cuts for the case when Q is not positive definite. Develop a prototype branch-and-cut framework for solving MISOCO problems using disjunctive conic cuts. Develop strategies which, and how many disjunctive conic cuts to generate when several cones are in the problem. Develop a comprehensive branch-and-cut framework for solving MISOCO problems using disjunctive conic cuts. Far future work Develop disjunctive conic cuts for Semidefinite Optimization.