EE 227A: Convex Optimization and Applications October 14, 2008

Similar documents
Lecture 14: Optimality Conditions for Conic Problems

Convex Optimization M2

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

5. Duality. Lagrangian

Lecture 7: Weak Duality

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Lecture Note 5: Semidefinite Programming for Stability Analysis

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lecture: Duality.

Convex Optimization Boyd & Vandenberghe. 5. Duality

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Semidefinite Programming Basics and Applications

4. Algebra and Duality

Convex Optimization & Lagrange Duality

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

Additional Homework Problems

Lecture: Duality of LP, SOCP and SDP

Global Optimization of Polynomials

Lagrangian Duality and Convex Optimization

Lecture 6: Conic Optimization September 8

SEMIDEFINITE PROGRAM BASICS. Contents

Lecture 8. Strong Duality Results. September 22, 2008

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Duality Uses and Correspondences. Ryan Tibshirani Convex Optimization

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Learning the Kernel Matrix with Semi-Definite Programming

EE/AA 578, Univ of Washington, Fall Duality

Optimization for Machine Learning

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

Learning the Kernel Matrix with Semidefinite Programming

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

ICS-E4030 Kernel Methods in Machine Learning

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

CS-E4830 Kernel Methods in Machine Learning

Lecture 3: Semidefinite Programming

Semidefinite Programming

Tutorial on Convex Optimization: Part II

12. Interior-point methods

Nonlinear Optimization: What s important?

ORF 523 Lecture 9 Spring 2016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Thursday, March 10, 2016

Convex Optimization M2

subject to (x 2)(x 4) u,

Homework Set #6 - Solutions

Lecture 7: Convex Optimizations

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

Convex Optimization and SVM

Semidefinite Programming

Agenda. 1 Duality for LP. 2 Theorem of alternatives. 3 Conic Duality. 4 Dual cones. 5 Geometric view of cone programs. 6 Conic duality theorem

Solving large Semidefinite Programs - Part 1 and 2

LECTURE 13 LECTURE OUTLINE

Lagrangian Duality Theory

Duality Theory of Constrained Optimization

Relaxations and Randomized Methods for Nonconvex QCQPs

Lecture 7: Positive Semidefinite Matrices

Convex Optimization. Lecture 12 - Equality Constrained Optimization. Instructor: Yuanzhang Xiao. Fall University of Hawaii at Manoa

EE364a Review Session 5

Constrained Optimization and Lagrangian Duality

Convex Optimization in Classification Problems

16.1 L.P. Duality Applied to the Minimax Theorem

Lecture: Introduction to LP, SDP and SOCP

Constrained optimization

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 6 Fall 2009

Trust Region Problems with Linear Inequality Constraints: Exact SDP Relaxation, Global Optimality and Robust Optimization

Linear and non-linear programming

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Applications of Linear Programming

On the Method of Lagrange Multipliers

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Optimality Conditions for Constrained Optimization

c 2005 Society for Industrial and Applied Mathematics

Lecture 6. Foundations of LMIs in System and Control Theory

1 Review of last lecture and introduction

15. Conic optimization

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Quadratic reformulation techniques for 0-1 quadratic programs

Continuous Optimisation, Chpt 9: Semidefinite Optimisation

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Summer School: Semidefinite Optimization

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Continuous Optimisation, Chpt 9: Semidefinite Problems

Conic Linear Optimization and its Dual. yyye

Module 04 Optimization Problems KKT Conditions & Solvers

Interior Point Algorithms for Constrained Convex Optimization

Lagrange Relaxation and Duality

Convex Optimization and Modeling

The Lagrangian L : R d R m R r R is an (easier to optimize) lower bound on the original problem:

Lecture 5. The Dual Cone and Dual Problem

Introduction to Semidefinite Programming I: Basic properties a

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Practice Exam 1: Continuous Optimisation

Advances in Convex Optimization: Theory, Algorithms, and Applications

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

Rank-one LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the well-known Lyapunov's matrix inequality about

Gauge optimization and duality

Perturbation Analysis of Optimization Problems

12. Interior-point methods

Transcription:

EE 227A: Convex Optimization and Applications October 14, 2008 Lecture 13: SDP Duality Lecturer: Laurent El Ghaoui Reading assignment: Chapter 5 of BV. 13.1 Direct approach 13.1.1 Primal problem Consider the SDP in standard form: p : C, : A i, = b i, i =1,..., m, 0, (13.1) where C, A i are given symmetric matrices, A, B = Tr AB denotes the scalar product between two symmetric matrices, and b R m is given. 13.1.2 Dual problem At first glance, the problem (13.1) is not amenable to the duality theory developed so far, since the constraint 0 is not a scalar one. Minimum eigenvalue representation. We develop a dual based on a representation of the problem via the minimum eigenvalue, as p C, : A i, = b i, i =1,..., m, λ min () 0, (13.2) where we have used the minimum eigenvalue function of a symmetric matrix A, given by λ min (A) := min Y Y, A : Y 0, Tr Y = 1 (13.3) to represent the positive semi-definiteness condition 0 in the SDP. The proof of the above representation of the minimum eigenvalue can be obtained by first showing that we can without loss of generality assume that A is diagonal, and noticing that we can then restrict Y to be diagonal as well. Note that the above representation proves that λ min is concave, so problem (13.2) is indeed convex as written. 13-1

Lagrangian and dual function. The Lagrangian for the maximization problem (13.2) is L(, λ, ν) = C, + ν i (b i A i, )+λ λ min () = ν T b + C ν i A i, + λ λ min (), where nu R m and λ 0 are the dual variables. The corresponding dual function g(λ, ν) : L(, λ, ν). involves the following subproblem, in which Z = C m ν ia i and λ 0 are given: G(λ, Z) : Z, + λλ min(). (13.4) For fixed λ 0, the function G(,λ) is the conjugate of the convex function λλ min. We have ( ) G(λ, Z) Z, + λ min Y, where, : Tr Y =1 min Z + λy, [eqn. (13.3)], Tr Y =1 Z + Y, [replace Y by λy ] min, Tr Y =λ min max, Tr Y =λ = min, Tr Y =λ = G(λ, Z), Z + Y, [minimax inequality] { 0 if Z + Y =0 + otherwise { 0 if Tr Z = λ 0, Z 0, G(λ, Z) := + otherwise. We now show that G(λ, Z) =G(λ, Z). To prove this, first note that G(λ, Z) 0 (since = 0 is feasible). This shows that G(λ, Z) itself is 0 if (λ, Z) is in the domain of G. Conversely, if Tr Z + λ 0, choosing = ɛti with ɛ = sign(tr Z + λ) and t + implies G(λ, Z) =+. Likewise, if Tr Z + λ = 0 and λ< 0, we choose = ti with t +, with the same result. Finally, if Tr Z = λ 0 but λ max (Z) > 0, choose = tuu T, with u a unit-norm eigenvector corresponding to the largest eigenvalue of Z, and t +. Here, we have Z, + λλ min () =t(λ max (Z)+λ) +, where we have exploited the fact that λ max (Z)+λ>λ 0. 13-2

Dual problem. Coming back to the Lagrangian, we need to apply our result to Z = C m ν ia i. The dual function is { 0 if Z := C m g(λ, ν) = ν ia i 0, Tr Z = λ 0, + otherwise. We obtain the dual problem λ,ν,z νt b : Z = C ν i A i 0, Tr Z = λ 0, or, after elimination of λ, and noticing that Z 0 implies Tr Z 0: ν ν T b : The dual problem is also an SDP, in standard inequality form. 13.2 Conic approach 13.2.1 Conic Lagrangian The same dual can be obtained with the conic Lagrangian L(, ν, Y ) := C, + ν i A i C (13.5) ν i (b i A i, )+ Y,, where now we associate a matrix dual variable Y to the constraint 0. Let us check that the Lagrangian above works, in the sense that we can represent the constrained maximization problem (13.1) as an unconstrained, maximin problem: p min L(, ν, Y ). We need to check that, for an arbitrary matrix Z, we have min Y, = { 0 if 0 otherwise. (13.6) This is an immediate consequence of the following: min Y, = min t 0 min Y, = min, Tr Y =t t 0 tλ min(), where we have exploited the representation of the minimum eigenvalue given in (13.3). The geometric interpretation is that the cone of positive-semidefinite matrices has a 90 o angle at the origin. 13-3

13.2.2 Dual problem The minimax inequality then implies The corresponding dual function is p d := min max ν, L(, ν, Y ). g(y, ν) { ν L(, ν, Y )= T b if C m ν ia i + Y =0 otherwise. The dual problem then writes ν, g(y, ν) = min ν, νt b : C ν i A i = Y 0. After elimination of the variable Y, we find the same problem as before, namely (13.5). 13.3 Weak and strong duality 13.3.1 Weak duality For the maximization problem (13.2), weak duality states that p d. Note that the fact that weak duality inequality ν T b C, holds for any primal-dual feasible pair (, ν), is a direct consequence of (13.6). 13.3.2 Strong duality From Slater s theorem, strong duality will hold if the primal problem is strictly feasible, that is, if there exist 0 such that A i, = b i, i =1,..., m. Using the same approach as above, one can show that the dual of problem (13.5) is precisely the primal problem (13.2). Hence, if the dual problem is strictly feasible, then strong duality also holds.recall that we say that a problem is attained if its optimal set is not empty. It turns out that if both problems are strictly feasible, then both problems are attained. A strong duality theorem. The following theorem summarizes our results. Theorem 1. Consider the SDP p : C, : A i, = b i, i =1,..., m, 0 13-4

and its dual The following holds: ν ν T b : ν i A i C. Duality is symmetric, in the sense that the dual of the dual is the primal. Weak duality always holds: p d, so that, for any primal-dual feasible pair (, ν), we have ν T b C,. If the primal (resp. dual) problem is bounded above (resp. below), and strictly feasible, then p = d and the dual (resp. primal) is attained. If both problems are strictly feasible, then p = d and both problems are attained. 13.4 Examples 13.4.1 An SDP where strong duality fails Contrarily to linear optimization problems, SDPs can fail to have a zero duality gap, even when they are feasible. Consider the example: x 2 + 1 0 0 p = min x 2 : 0 x 1 x 2 0. x 0 x 2 0 Any primal feasible x satisfies x 2 = 0. Indeed, positive-semidefiniteness of the lower-right 2 2 block in the LMI of the above problem writes, using Schur complements, as x 1 0, x 2 2 0. Hence, we have p = 0. The dual is d Y S 3 Y 11 : Y 0, Y 22 =0, 1 Y 11 2Y 23 =0. Any dual feasible Y satisfies Y 23 = 0 (since Y 22 = 0), thus Y 11 = 1 =d. 13.4.2 An eigenvalue problem For a matrix A S n +, we consider the SDP p The associated Lagrangian, using the conic approach, is A, : Tr =1, 0. (13.7) L(, Y, ν) = A, + ν(1 Tr )+ Y,, 13-5

with the matrix dual variable Y 0, while ν R is free. The dual function is g(y, ν) We obtain the dual problem p ν,y L(, Y, ν) = { ν if νi = Y + A + otherwise. ν : Y + A = λi, Y 0. Eliminating Y leads to {ν : νi A} = λ max (A). ν Both the primal and dual problems are strictly feasible, so p = d, and both values are attained. This proves the representation (13.7) for the largest eigenvalue of A. 13.4.3 SDP relaxations for a non-convex quadratic problem In lectures 7 and 11, we have seen two kinds of relaxation for the non-convex problem p : x x T Wx : x 2 i 1, i =1,..., n, where the symmetric matrix W S n is given. One relaxation is based on a standard relaxation of the constraints, and leads to (see lecture 11) p d lag := min Tr D : D W, D diagonal. (13.8) D Another relaxation (lecture 7) involved expressing the problem as an SDP with rank constraints on the a matrix = xx T : d rank : W, : 0, ii =1, i =1,..., m. Let us examine the dual of the first relaxation (13.8). We note that the problem is strictly feasible, so strong duality holds. Using the conic approach, we have d lag := min D Y = d rank. max min D Y, W : Tr D + Y, W D Tr D + Y, W D 0, Y ii =1, i =1,..., m This shows that both Lagrange and rank relaxations give the same value, and are dual of each other. In general, for arbitrary non-convex quadratic problems, the rank relaxation can be shown to be always better than the Lagrange relaxation, as the former is the (conic) dual to the latter. If either is strictly feasible, then they have the same optimal value. 13-6