Convexification of Mixed-Integer Quadratically Constrained Quadratic Programs

Similar documents
Semidefinite Programming Basics and Applications

Quadratic reformulation techniques for 0-1 quadratic programs

Using quadratic convex reformulation to tighten the convex relaxation of a quadratic program with complementarity constraints

5. Duality. Lagrangian

Convex Optimization M2

A Note on Representations of Linear Inequalities in Non-Convex Mixed-Integer Quadratic Programs

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Convex Optimization Boyd & Vandenberghe. 5. Duality

Lecture: Duality.

Convex Optimization M2

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Nonconvex Quadratic Programming: Return of the Boolean Quadric Polytope

Lecture: Duality of LP, SOCP and SDP

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Lecture 6: Conic Optimization September 8

Lecture 18: Optimization Programming

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

EE/AA 578, Univ of Washington, Fall Duality

Comparing Convex Relaxations for Quadratically Constrained Quadratic Programming

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Convex Optimization & Lagrange Duality

CS711008Z Algorithm Design and Analysis

4y Springer NONLINEAR INTEGER PROGRAMMING

12. Interior-point methods

Lecture 8. Strong Duality Results. September 22, 2008

Cutting Planes for First Level RLT Relaxations of Mixed 0-1 Programs

Linear and non-linear programming

Module 04 Optimization Problems KKT Conditions & Solvers

Lecture: Introduction to LP, SDP and SOCP

Lecture Note 5: Semidefinite Programming for Stability Analysis

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

A Binarisation Approach to Non-Convex Quadratically Constrained Quadratic Programs

subject to (x 2)(x 4) u,

Relaxations and Randomized Methods for Nonconvex QCQPs

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

On Valid Inequalities for Quadratic Programming with Continuous Variables and Binary Indicators

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Lecture: Cone programming. Approximating the Lorentz cone.

ICS-E4030 Kernel Methods in Machine Learning

There are several approaches to solve UBQP, we will now briefly discuss some of them:

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Convex Optimization and Support Vector Machine

Introduction to Semidefinite Programming I: Basic properties a

10 Numerical methods for constrained problems

Support Vector Machines: Maximum Margin Classifiers

Optimality, Duality, Complementarity for Constrained Optimization

Written Examination

Lecture 3: Semidefinite Programming

Lecture: Examples of LP, SOCP and SDP

EE 227A: Convex Optimization and Applications October 14, 2008

Tutorial on Convex Optimization for Engineers Part II

Linear and Combinatorial Optimization

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones

Appendix A Taylor Approximations and Definite Matrices

Decomposition Methods for Quadratic Zero-One Programming

l p -Norm Constrained Quadratic Programming: Conic Approximation Methods

EE364a Review Session 5

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

CS-E4830 Kernel Methods in Machine Learning

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

The min-cut and vertex separator problem

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Algorithms for constrained local optimization

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Optimality Conditions for Constrained Optimization

Duality Theory of Constrained Optimization

Interior Point Methods for Convex Quadratic and Convex Nonlinear Programming

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Copositive Programming and Combinatorial Optimization

Continuous Optimisation, Chpt 9: Semidefinite Problems

CONSTRAINED NONLINEAR PROGRAMMING

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Lecture 7: Convex Optimizations

Convex Optimization and SVM

Part IB Optimisation

IE 521 Convex Optimization

CHAPTER 2: QUADRATIC PROGRAMMING

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method

c 2000 Society for Industrial and Applied Mathematics

Nonlinear Optimization: What s important?

Lecture #21. c T x Ax b. maximize subject to

Relations between Semidefinite, Copositive, Semi-infinite and Integer Programming

Network Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini

Homework Set #6 - Solutions

Solving large Semidefinite Programs - Part 1 and 2

minimize x subject to (x 2)(x 4) u,

An introductory example

Lagrangian-Conic Relaxations, Part I: A Unified Framework and Its Applications to Quadratic Optimization Problems

Interior Point Algorithms for Constrained Convex Optimization

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

12. Interior-point methods

Constrained Optimization and Lagrangian Duality

CSCI : Optimization and Control of Networks. Review on Convex Optimization

Convex relaxation. In example below, we have N = 6, and the cut we are considering

A Brief Review on Convex Optimization

Transcription:

Convexification of Mixed-Integer Quadratically Constrained Quadratic Programs Laura Galli 1 Adam N. Letchford 2 Lancaster, April 2011 1 DEIS, University of Bologna, Italy 2 Department of Management Science, Lancaster University, United Kingdom

Outline Literature Review Motivation Relaxations of QCQPs Quadratic Convex Reformulation (QCR) Convexification of 0-1 QCQP Reformulation with equations only Reformulation with inequalities Convexification of MIQCQP General integer variables Continuous variables Conclusions

Motivation It is well known that Semidefinite Programming (SDP) can be used to form useful relaxations for a variety of N P-hard optimisation problems, such as: zero-one linear programs (0-1 LP) zero-one quadratic programs (0-1 QP) non-convex quadratically constrained quadratic programs (QCQP) general polynomial programs

Motivation (cont.) Recently, Billionnet et al. (2009, 2010) proposed to use SDP to reformulate 0-1 QPs, not just to relax them. The idea is to perturb the objective function in such a way that it is convex and the continuous relaxation of the 0-1 QP is tighter. We examine to what extent this approach can be applied to the (much) more general case of mixed-integer quadratically constrained quadratic programs (MIQCQP).

Semidefinite relaxation of QCQP We start with Shor s (1987) relaxation of non-convex QCQP (which is an N P-hard global optimisation problem). An instance takes the form: inf x T Q 0 x +c 0 x s.t. x T Q j x +c j x b j x R n, (j = 1,...,m) where the Q j are symmetric, square and rational matrices of order n, the c j are rational n-vectors, and the b j are rational scalars.

Semidefinite relaxation of QCQP (cont.) The basic SDP relaxation of the QCQP is derived as follows: define the n n symmetric matrix X = xx T replace the terms x T Q j x with Q j X define the augmented symmetric matrix ( )( ) T ( ) 1 1 1 x T Y = = x x x X note that Y is positive semidefinite (psd).

Semidefinite relaxation of QCQP (cont.) The resulting SDP is: inf Q 0 X +c 0 x s.t. Q j X +c j x b j Y 0. (j = 1,...,m) (The last constraint imposes psd-ness on Y.)

Lagrangian relaxation of QCQP As noted by Fujie & Kojima (1997), Lémaréchal & Oustry (2001) and others, there is a connection between semidefinite relaxation and Lagrangian relaxation. Suppose we relax all of the constraint in Lagrangian fashion, and let λ denote the vector of Lagrangian multipliers. The Lagrangian objective becomes: ( ) m m m x (Q T 0 + λ j Q )x j + c 0 + λ j c j x λ j b j. j=1 j=1 j=1

Lagrangian relaxation of QCQP (cont.) Let the Lagrangian objective be denoted by f(x,λ). The relaxed problem is: inf { f(x,λ) : x R n }, which is an unconstrained quadratic minimisation problem. The Lagrangian dual is: sup inf x R nf(x,λ). λ R m +

Lagrangian relaxation of QCQP (cont.) Suppose that strong duality holds for the SDP. Then one can show that: the value of the Lagrangian dual equals the SDP bound, and the optimal Lagrangian multiplier vector λ is nothing but the optimal dual vector for the SDP.

QCQP summary We have: opt SDP bound dual SDP bound = Lagrangian bound and the second inequality is an equality under suitable constraint qualifications.

0-1 QP Now consider a 0-1 QP of the form: min x T Qx +c T x s.t. Ax = b Dx f x {0,1} n. As observed by many authors, the condition that x be binary is equivalent to requiring x i xi 2 = 0 (i = 1,...,n).

SDP relaxation of 0-1 QP This leads immediately to the following semidefinite relaxation: min Q X +c T x s.t. Ax = b Dx f x = diag(x) Y 0.

Convex reformulation of 0-1 QP Reformulating a 0-1 QP instance means perturbing the objective function in such a way that: the objective function is convex the cost of each feasible solution is unchanged the constraints remain unchanged, so that the feasible set is unchanged.

Minimum eigenvalue reformulation of 0-1 QP A simple reformulation technique was given by Hammer & Rubin (1970). They add 1 i n M(x2 i x i ) to the objective function, where M > 0 is large enough to ensure that the matrix Q +MI of quadratic coefficients is psd. In this way, they ensure that the reformulated instance is convex. (A suitable value for M is the minimum eigenvalue of Q, multiplied by 1.)

Quadratic Convex Reformulation (QCR) Billionnet et al. (2009) develop this idea further, in two ways: they perturb the objective function by addinggeneral terms of the form λ i (x i x 2 i ) and µ jk (a j x b j )x k. instead of merely ensuring that the reformulated instance is convex, they also ensure that its continuous relaxation is as tight as possible.

Quadratic Convex Reformulation (QCR) (cont.) Billionnet et al. (2009) show that the optimal multipliers λ and µ can be found by solving an SDP. They call this approach convex quadratic reformulation or QCR. Once QCR has been applied, one can feed the reformulated instance into any convex MIQP solver. (An extension of QCR to general MIQP was given recently by Billionnet et al., 2010.)

0-1 QCQP Our first goal is to extend the QCR method to 0-1 QCQP. We suppose the instance is written as: min x T Q 0 x +c 0 x s.t. x T Q j x +c j x = b j (j N 1) x T Q j x +c j x b j (j N 2) x {0,1} n. (We assume for simplicity of notation that any linear constraints, along with any additional valid inequalities, are subsumed in the above system.)

SDP relaxation of 0-1 QCQP The semidefinite relaxation is immediate: min Q 0 X +c 0 x s.t. Q j X +c j x = b j (j N 1) Q j X +c j x b j (j N 2) x = diag(x) Y 0. The Lagrangian relaxation is easy to write as well.

Convex reformulation of 0-1 QCQP By analogy with the original QCR method, we seek to transform a 0-1 QCQP instance into another 0-1 QCQP instance, that has the following four properties: the objective and constraint functions are convex the set of feasible solutions is identical to that of the original instance the cost of each feasible solution is the same as it was in the original instance...

Convex reformulation of 0-1 QCQP By analogy with the original QCR method, we seek to transform a 0-1 QCQP instance into another 0-1 QCQP instance, that has the following four properties: the objective and constraint functions are convex the set of feasible solutions is identical to that of the original instance the cost of each feasible solution is the same as it was in the original instance...... and the lower bound from the continuous relaxation is equal to the semidefinite/lagrangian bound for the original instance.

0-1 QCQP with Quadratic Equations For the case in which all constraints are equations (i.e., N 2 = ), we propose the following scheme: 1 Solve the SDP, obtaining optimal dual vectors λ R N 1 and ν R n ; 2 Perturb the objective function by adding terms of the form x T Diag(ν )x (ν ) T x and λ j (xt Q j x +c j x b j ) for j N 1 ; 3 Replace each quadratic equation with two quadratic inequalities; 4 Convexify the quadratic inequalities arbitrarily (e.g., using the minimum eigenvalue method).

0-1 QCQP with Quadratic Equations (cont.) Theorem When the above scheme is applied to a 0-1 QCQP instance with N 2 =, the reformulated instance is convex. Moreover, if the primal SDP satisfies the Slater condition, then the lower bound from the continuous relaxation of the reformulated instance will equal the SDP bound.

Handling Quadratic Inequalities: two difficulties The case of quadratic inequalities is more tricky because of two serious difficulties! The first can be shown considering the following simple instance: max x 1 + +x n s.t. x i x j 0 (1 i < j n) x 2 i x i = 0 (i = 1,...,n). The optimum is 1 and one can check that the SDP bound is also 1.

Handling Quadratic Inequalities: two difficulties (cont.) The quadratic inequalities are non-convex, so we must use the equations x 2 i x i = 0 to convexify them. It turns out that the best we can do is to replace the inequalities with: 1 2 (x2 i x i )+ 1 2 (x2 j x j )+x i x j 0 (1 i < j n). But the optimal solution to the continuous relaxation of the reformulated instance is then (1/2,...,1/2) T, yielding a very poor upper bound of n/2.

Handling Quadratic Inequalities: two difficulties (cont.) In other words, when quadratic inequalities are present: opt SDP bound = dual SDP bound = Lagrangian bound LP bound from reformulated instance and the last inequality is usually strict! An intuitive explanation for this phenomenon is that quadratic inequalities cannot be used to perturb the objective function.

Handling Quadratic Inequalities: two difficulties (cont.) The second difficulty is that finding an optimal reformulation is itself a rather complex optimisation problem. It is possible to formulate the problem of finding the best such reformulation as one huge SDP. Unfortunaley, the size and complexity of this SDP grows rapidly as the number of inequalities increases.

Handling Quadratic Inequalities: possible remedies If we are willing to convert the 0-1 QCQP into a mixed 0-1 QCQP, then we can do better. We can add continuous slack variables to convert every quadratic inequality into a quadratic equation. Once this is done, it is always possible to reformulate the mixed 0-1 QCQP instance in such a way that the bound from the continuous relaxation is equal to the Lagrangian bound. As before, if the primal SDP satisfies the Slater condition, this bound will be equal to the SDP bound.

Handling Quadratic Inequalities: possible remedies (cont.) The above scheme has one disadvantage: we have to add N 2 continuous variables. This is unattractive when N 2 is large. We have another variant in which only one continuous variable needs to be added. (The basic idea is to use the dual SDP solution to construct a surrogate quadratic inequality that contains all of the information we need to get a good bound, and then assign a slack variable only to the surrogate constraint.)

General-Integer Variables Next, we move on to the case of general-integer variables. A first observation is that integer variables present no problem for the semidefinite and Lagrangian relaxations. As for reformulation, note that we cannot use terms of the form x 2 i x i to perturb the objective or constraints coefficient of integer variables. This can prevent the existence of a convex reformulation.

General-Integer Variables (cont.) To get around this, we can use binary expansion. In fact, if such a variable is nonnegative and bounded from above by 2 p for some positive integer p, then it can be replaced by p binary variables. Then, if all variables are bounded in this way, we can reduce the problem to 0-1 QCQP. We show that this transformation cannot make the SDP bound any worse, and in fact in some cases it can make it better!

General-Integer Variables (cont.) Consider the following instance with just one integer variable: } min {x1 2 3x 1 : x 1 3, x 1 Z +. There are two optimal solutions of cost 2, obtained by setting x 1 to either 1 or 2. The basic SDP relaxation is: min{x 11 3x 1 : 0 x 1 3, Y 0}. The optimal solution to this relaxation is to set x 1 to 1.5 and X 11 to 2.25, giving a lower bound of 2.25.

General-Integer Variables (cont.) Now, applying binary expansion to this instance, we use the substitution x 1 = x 1 1 +2 x 2 1, where x 1 1 and x 2 1 are new binary variables. The SDP relaxation for the transformed instance is: min X 1,1 11 +4 X 2,2 11 s.t. x 1 1 = X 1,1 11 x 1 2 2,2 = X 11 Ỹ 0. +4 X 1,2 11 3 x1 1 6 x 2 1 One can check with an SDP solver that all optimal solutions to this SDP have cost 2. So, the transformed SDP yields an improved lower bound.

General-Integer Variables (cont.) If, on the other hand, the general-integer variables are not bounded from above a priori, then we run into a serious problem: Theorem (Jeroslow 1973) Testing feasibility of an integer program with quadratic constraints is undecidable. So we cannot expect an efficient convexification procedure in general. (Of course, in practical applications, variables can usually be bounded quite easily.)

Continuous Variables Finally, we consider the case in which continuous variables are permitted. In the context of the semidefinite and Lagrangian approaches, continuous variables can be handled in exactly the same way as integer variables, and therefore they present no problems. When it comes to reformulation, on the other hand, continuous variables cause serious difficulties. (Indeed, a convex reformulation may not exist in general, and we cannot get around this issue by binarisation, as we do for integer variables.)

Continuous Variables (cont.) Here, three outcomes are possible: The instance can be convexified, and the bound from the continuous relaxation is as good as the SDP bound. The instance can be convexified, but the bound from the continuous relaxation is worse than the SDP bound. The instance cannot be convexified at all. We have some necessary and sufficient conditions for these three cases to hold.

Conclusions Basically, we have three possible situations: When there are no quadratic inequalities and no continuous variables, and all general-integer variables are bounded, we can always obtain a convex reformulation that gives a strong bound. When quadratic inequalities are present, this may not be possible, unless we are prepared to add slack variables. When continuous variables are present, a convex reformulation is not even guaranteed to exist.