E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Similar documents
SF2822 Applied nonlinear optimization, final exam Saturday December

Lecture 7: Convex Optimizations

Lecture 8. Strong Duality Results. September 22, 2008

SF2822 Applied nonlinear optimization, final exam Wednesday June

EE 227A: Convex Optimization and Applications October 14, 2008

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Lecture: Duality of LP, SOCP and SDP

Semidefinite Programming Basics and Applications

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

CS711008Z Algorithm Design and Analysis

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

5. Duality. Lagrangian

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

Constrained Optimization and Lagrangian Duality

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

Convex Optimization Boyd & Vandenberghe. 5. Duality

Module 04 Optimization Problems KKT Conditions & Solvers

Convex Optimization & Lagrange Duality

12. Interior-point methods

Lecture: Duality.

Linear and non-linear programming

Convex Optimization M2

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture 6: Conic Optimization September 8

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

SDP Relaxations for MAXCUT

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Lecture 20: November 1st

SEMIDEFINITE PROGRAM BASICS. Contents

Semidefinite Programming

Additional Homework Problems

15. Conic optimization

10-725/ Optimization Midterm Exam

Duality in Linear Programs. Lecturer: Ryan Tibshirani Convex Optimization /36-725

LECTURE 13 LECTURE OUTLINE

4. Algebra and Duality

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Global Optimization of Polynomials

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Convexification of Mixed-Integer Quadratically Constrained Quadratic Programs

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Duality. Geoff Gordon & Ryan Tibshirani Optimization /

10 Numerical methods for constrained problems

Support vector machines

Convex Optimization and Support Vector Machine

Applications of Linear Programming

Interior Point Algorithms for Constrained Convex Optimization

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

Support Vector Machines

Lecture 3: Semidefinite Programming

Optimization for Machine Learning

Duality Theory of Constrained Optimization

Semidefinite Programming

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Optimality Conditions for Constrained Optimization

Strong duality in Lasserre s hierarchy for polynomial optimization

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Trust Region Problems with Linear Inequality Constraints: Exact SDP Relaxation, Global Optimality and Robust Optimization

Lecture: Cone programming. Approximating the Lorentz cone.

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

Convex Optimization M2

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Assignment 1: From the Definition of Convexity to Helley Theorem

ICS-E4030 Kernel Methods in Machine Learning

Lecture 14: Optimality Conditions for Conic Problems

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Lecture 18: Optimization Programming

LMI Methods in Optimal and Robust Control

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

minimize x subject to (x 2)(x 4) u,

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

Lecture: Introduction to LP, SDP and SOCP

12. Interior-point methods

Quadratic reformulation techniques for 0-1 quadratic programs

A priori bounds on the condition numbers in interior-point methods

CSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods

Homework 4. Convex Optimization /36-725

Optimality, Duality, Complementarity for Constrained Optimization

CSCI : Optimization and Control of Networks. Review on Convex Optimization

EE/AA 578, Univ of Washington, Fall Duality

Constrained optimization: direct methods (cont.)

A Brief Review on Convex Optimization

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

Convex Optimization and SVM

Continuous Optimisation, Chpt 9: Semidefinite Problems

ELE539A: Optimization of Communication Systems Lecture 6: Quadratic Programming, Geometric Programming, and Applications

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Constrained optimization

Convex Optimization and Modeling

Transcription:

E5295/5B5749 Convex optimization with engineering applications Lecture 5 Convex programming and semidefinite programming A. Forsgren, KTH 1 Lecture 5 Convex optimization 2006/2007

Convex quadratic program A quadratic program is an optimization problem on the form (LP ) 1 minimize x IR n 2 xt Hx + c T x subject to Ax = b, x 0. May be written on many (equivalent) forms. We will only consider convex quadratic programs, where H 0. The nonconvex quadratic programming problem is NP-hard. A. Forsgren, KTH 2 Lecture 5 Convex optimization 2006/2007

For a primal quadratic program Primal and dual quadratic programs (P QP ) minimize 1 2 xt Hx + c T x subject to Ax = b, we will associate a dual quadratic program x 0, (DQP ) max 1 2 wt Hw + b T y subject to Hw + A T y + s = c, s 0. We may derive the dual by Lagrangian relaxation. If H 0, we may eliminate w. A. Forsgren, KTH 3 Lecture 5 Convex optimization 2006/2007

Quadratic programming, optimality conditions Primal and dual quadratic programs (P QP ) minimize 1 2 xt Hx + c T x subject to Ax = b, x 0, (DQP ) max 1 2 wt Hw + b T y subject to Hw + A T y + s = c, s 0. Optimality conditions: Ax = b, Hx + A T y + s = c, x j s j = 0, j = 1,..., n, x 0, s 0. A. Forsgren, KTH 4 Lecture 5 Convex optimization 2006/2007

Nonlinearly constrained convex program Consider a convex optimization problem on the form minimize f(x) (CP ) subject to g i (x) 0, i I, a T i x b i = 0, i E, x IR n, I E = {1,..., m}, I E =, where f : IR n IR and g i : IR n IR are convex and twice continuously differentiable on IR n. Let F = {x IR n : g i (x) 0, i I, a T i x b i = 0, i E}. A. Forsgren, KTH 5 Lecture 5 Convex optimization 2006/2007

Nonlinearly constrained problems require some regularity For a nonlinearly constrained problem, the analogous first-order conditions are not always necessary. Let g(x) = (x 1 1) 2 x 2 2 + 1 (x 1 + 1) 2 x 2 2 + 1 with x = x 2 0 0. x* x 1 The linearization of the constraints at x does not describe the feasible region sufficiently well. A. Forsgren, KTH 6 Lecture 5 Convex optimization 2006/2007

Constraint qualification If the constraints satisfy some regularity condition, often referred to as a constraint qualification, the analogous results hold. Definition. Let F = {x IR n : g i (x) 0, i I, a T i x b i = 0, i E}, where g i : IR n IR are convex and twice continuously differentiable on IR n. Then, the Slater constraint qualification holds if there is a point x such that g i ( x) > 0, i I, a T i x b i = 0, i E. With the Slater constraint qualification, strong duality holds. A. Forsgren, KTH 7 Lecture 5 Convex optimization 2006/2007

Optimality conditions for convex program, cont. Proposition. Let (CP ) be a convex program for which the Slater constraint qualification holds. Then, the point x is a minimizer to (CP ) if and only if there is a λ IR m such that (i) g i (x ) 0, i I, and a T i x b i = 0, i E, (ii) f(x ) = i I g i (x )λ i + i E a i λ i, (iii) λ i 0, i I, (iv) λ i g i (x ) = 0, i I. These conditions are often referred to as the KKT conditions. A. Forsgren, KTH 8 Lecture 5 Convex optimization 2006/2007

Metric for semidefinite matrices For x and y in IR n, the inner product is given by x T y = The associated norm is the Euclidean norm. Let S n denote the space of symmetric n n matrices. For X and Y in S n, the inner product is given by n j=1 x j y j. trace(x T Y ) = trace(xy ) = n n x ij y ij. i=1 j=1 The associated norm is the Frobenius norm. Proposition. If A 0 and B 0 belong to S n, then trace(ab) 0. Proof. If A = A T 0, there is an L such that A = LL T. Then trace(ab) = trace(ll T B) = trace(l T BL) 0. A. Forsgren, KTH 9 Lecture 5 Convex optimization 2006/2007

Semidefinite programming In semidefinite programming X is a matrix in S n. Linear programming is the special case when X is diagonal. In LP we have constraints A T i x = b i, i = 1,..., m, where A T i row i of the constraint matrix A. IR n is In SDP A i belongs to S n for i = 1,..., m, and the constraints become trace(a i X) = b i, i = 1,..., m. Analogously, LP has the objective function c T x, where c IR n. In SDP the objective function becomes trace(cx), where C S n. A. Forsgren, KTH 10 Lecture 5 Convex optimization 2006/2007

A semidefinite program on standard form A semidefinite program on standard form may be written as minimize trace(cx) (P SDP ) subject to trace(a i X) = b i, i = 1,..., m, X = X T 0. (If C and A i, i = 1,..., m, are diagonal we may choose X diagonal which gives (P LP ).) As for LP the form is not so interesting. The important feature is that we have a linear problem with equality constraints and matrix inequalities. (P SDP ) is a convex problem. A. Forsgren, KTH 11 Lecture 5 Convex optimization 2006/2007

Lagrangian relaxation of a semidefinite program (P SDP ) minimize X S n trace(cx) subject to trace(a i X) = b i, i = 1,..., m, X = X T 0. For a given y IR m, we obtain ϕ(y) = minimize = X=X T 0 trace(cx) m i=1 b T y if m i=1 A i y i C, otherwise. y i (trace(a i X) b i ) A. Forsgren, KTH 12 Lecture 5 Convex optimization 2006/2007

The dual of a semidefinite program (P SDP ) minimize X S n trace(cx) subject to trace(a i X) = b i, i = 1,..., m, X = X T 0. (DSDP ) maximize y IR m mi=1 b i y i subject to m i=1 A i y i C. If X is feasible to (P SDP ) and y, S are feasible to (DSDP ), then trace(cx) b T y = trace(xs). A. Forsgren, KTH 13 Lecture 5 Convex optimization 2006/2007

Semidefinite programs with no duality gap Some regularity condition has to be enforced in order to ensure that there is no duality gap. For example, the Slater condition, which requires a strictly feasible X, i.e., an X that satisfies trace(a i X) = b i, i = 1,..., m, X = X T 0. Similar requirement on (DSDP ) ensures no duality gap and existence of optimal solutions to both (P SDP ) and (DSDP ). A. Forsgren, KTH 14 Lecture 5 Convex optimization 2006/2007

Semidefinite optimality The set W defined by W = {w IR m : w i = trace(a i X), i = 1,..., m, X = X T 0} is not closed. As an example, let A 1 = 0 1 1 0 and A 2 = 1 0 0 0. For ɛ 0, let w(ɛ) = (2 ɛ) T. Then, w(ɛ) W for ɛ > 0, where the associated X(ɛ) is given by X(ɛ) = However, w(0) W. ɛ 1 1 x 22 (ɛ) for x 22 (ɛ) 1 ɛ. A. Forsgren, KTH 15 Lecture 5 Convex optimization 2006/2007

A semidefinite program may not have an optimal solution Let A 1 = 0 1 1 0, b 1 = 2 and C = 1 0 0 0. The primal problem has optimal value 0 but no optimal solution. The dual problem has optimal value 0 and an optimal solution. (Wolkowicz, Saigal and Vandenberghe: Handbook of Semidefinite Programming, Example 4.1.1.) A. Forsgren, KTH 16 Lecture 5 Convex optimization 2006/2007

Let A 1 = C = A semidefinite program may have duality gap a 0 0 0 0 0 0 0 0 1 0 0 0 0 1 0 1 0, A 2 = 0 0 0 0 1 0 0 0 0, b = 1 0, where a is a given positive number. and The primal problem has optimal value a and an optimal solution. The dual problem has optimal value 0 and an optimal solution. (Wolkowicz, Saigal and Vandenberghe: Handbook of Semidefinite Programming, Example 4.1.2.) A. Forsgren, KTH 17 Lecture 5 Convex optimization 2006/2007

Suggested reading Suggested reading in the textbook: ˆ Section 4.4. ˆ Sections 5.4 5.9. A. Forsgren, KTH 18 Lecture 5 Convex optimization 2006/2007