Lagrangian Duality Theory

Similar documents
Convex Optimization M2

Conic Linear Optimization and its Dual. yyye

5. Duality. Lagrangian

Convex Optimization & Lagrange Duality

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lecture: Duality.

Convex Optimization Boyd & Vandenberghe. 5. Duality

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Lecture: Duality of LP, SOCP and SDP

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

10 Numerical methods for constrained problems

More First-Order Optimization Algorithms

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

EE/AA 578, Univ of Washington, Fall Duality

Lecture 3. Optimization Problems and Iterative Algorithms

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Linear and Combinatorial Optimization

Convex Optimization and SVM

4TE3/6TE3. Algorithms for. Continuous Optimization

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

subject to (x 2)(x 4) u,

Constrained Optimization and Lagrangian Duality

Convex Optimization and Modeling

The Lagrangian L : R d R m R r R is an (easier to optimize) lower bound on the original problem:

On the Method of Lagrange Multipliers

Lagrangian Duality and Convex Optimization

LECTURE 10 LECTURE OUTLINE

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

Lecture 6: Conic Optimization September 8

Homework Set #6 - Solutions

Lagrange Relaxation and Duality

Convex Optimization and Support Vector Machine

12. Interior-point methods

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Lecture 8. Strong Duality Results. September 22, 2008

4. Algebra and Duality

ICS-E4030 Kernel Methods in Machine Learning

Semidefinite Programming Basics and Applications

Lagrangian Duality. Richard Lusby. Department of Management Engineering Technical University of Denmark

Penalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way.

Additional Homework Problems

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Lecture: Introduction to LP, SDP and SOCP

12. Interior-point methods

Agenda. 1 Duality for LP. 2 Theorem of alternatives. 3 Conic Duality. 4 Dual cones. 5 Geometric view of cone programs. 6 Conic duality theorem

Nonlinear Programming

EE364a Review Session 5

Semidefinite Programming

Linear and non-linear programming

Tutorial on Convex Optimization: Part II

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Solving Dual Problems

Optimality, Duality, Complementarity for Constrained Optimization

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Support Vector Machines

Lagrangian Duality. Evelien van der Hurk. DTU Management Engineering

CS-E4830 Kernel Methods in Machine Learning

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

Applications of Linear Programming

Lecture 18: Optimization Programming

Support Vector Machines: Maximum Margin Classifiers

Date: July 5, Contents

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Optimization for Machine Learning

Duality. Geoff Gordon & Ryan Tibshirani Optimization /

Summer School: Semidefinite Optimization

EE 227A: Convex Optimization and Applications October 14, 2008

Lecture 7: Convex Optimizations

Lecture 5. The Dual Cone and Dual Problem

Part IB Optimisation

LECTURE 13 LECTURE OUTLINE

Lecture Notes on Support Vector Machine

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Generalization to inequality constrained problem. Maximize

CSCI : Optimization and Control of Networks. Review on Convex Optimization

Math 273a: Optimization Subgradients of convex functions

A Brief Review on Convex Optimization

A ten page introduction to conic optimization

Interior Point Algorithms for Constrained Convex Optimization

Nonlinear Programming and the Kuhn-Tucker Conditions

Machine Learning And Applications: Supervised Learning-SVM

Constrained Optimization Theory

Optimality Conditions for Constrained Optimization

Interior Point Methods for Mathematical Programming

1 Review of last lecture and introduction

15. Conic optimization

Duality Theory of Constrained Optimization

Primal/Dual Decomposition Methods

Interior Point Methods in Mathematical Programming

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

3.10 Lagrangian relaxation

Transcription:

Lagrangian Duality Theory Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapter 14.1-4 1

Recall Primal and Dual of Conic LP (CLP ) minimize c x subject to a i x = b i, i = 1, 2,..., m, x K; and it dual problem (CLD) maximize b T y m subject to i y i a i + s = c, s K, where y R m, s is called the dual slack vector/matrix, and K is the dual cone of K. In general, K can be decomposed to K = K 1 K 2... K p, that is, x = (x 1 ; x 2 ;...; x p ), x i K i, i. Note that K = K1 K2... Kp, or s = (s 1 ; s 2 ;...; s p ), s i Ki, i. This is a powerful but very structured duality form. 2

Lagrangian Function We now consider the general constrained optimization: (GCO) min f(x) s.t. c i x) (, =, ) 0, i = 1,..., m, For Lagrange Multipliers. Y := {y i (, free, ) 0, i = 1,..., m}, the Lagrangian Function is again given by L(x, y) = f(x) y T c(x) = f(x) m y i c i (x), y Y. i=1 We now develop the Lagrangian Duality theory as an alternative to Conic Duality theory. For general nonlinear constraints, the Lagrangian Duality theory is more applicable. 3

Toy Example Again minimize (x 1 1) 2 + (x 2 1) 2 subject to x 1 + 2x 2 1 0, 2x 1 + x 2 1 0. L(x, y) = f(x) y T c(x) = f(x) 2 y i c i (x) = i=1 = (x 1 1) 2 + (x 2 1) 2 y 1 (x 1 + 2x 2 1) y 2 (2x 1 + x 2 1), (y 1 ; y 2 ) 0 where L x (x, y) = 2(x 1 1) y 1 2y 2 2(x 2 1) 2y 1 y 2 4

Lagrangian Relaxation Problem For given multipliers y Y, consider problem (LRP ) inf L(x, y) = f(x) y T c(x) s.t. x R n. Again, y i can be viewed as a penalty parameter to penalize constraint violation c i (x), i = 1,..., m. In the toy example, for given (y 1 ; y 2 ) 0, the LRP is: inf (x 1 1) 2 + (x 2 1) 2 y 1 (x 1 + 2x 2 1) y 2 (2x 1 + x 2 1) s.t. (x 1 ; x 2 ) R 2, and it has a close form solution x: x 1 = y 1 + 2y 2 2 + 1 and x 2 = 2y 1 + y 2 2 + 1 with the minimal value function = 1.25y 2 1 1.25y 2 2 2y 1 y 2 2y 1 2y 2. 5

Inf-Value Function as the Dual Objective For any y Y, the the minimal value function (including unbounded from below or infeasible cases) and the Lagrangian Dual Problem (LDP) are given by: ϕ(y) := inf x L(x, y), s.t. x R n. (LDP ) sup y ϕ(y), s.t. y Y. Theorem 1 The Lagrangian dual objective ϕ(y) is a concave function. Proof: For any given two multiply vectors y 1 Y and y 2 Y, ϕ(αy 1 + (1 α)y 2 ) = inf x L(x, αy 1 + (1 α)y 2 ) = inf x [f(x) (αy 1 + (1 α)y 2 ) T c(x)] = inf x [αf(x) + (1 α)f(x) α(y 1 ) T c(x) (1 α)(y 2 ) T c(x)] = inf x [αl(x, y 1 ) + (1 α)l(x, y 2 )] α[inf x L(x, y 1 )] + (1 α)[inf x L(x, y 2 )] = αϕ(y 1 ) + (1 α)ϕ(y 2 ), 6

Dual Objective Establishes a Lower Bound Theorem 2 (Weak duality theorem) For every y Y, the Lagrangian dual function ϕ(y) is less or equal to the infimum value of the original GCO problem. Proof: ϕ(y) = inf x {f(x) y T c(x)} inf x {f(x) y T c(x) s.t. c(x)(, =, )0 } inf x {f(x) : s.t. c(x)(, =, )0 }. The first inequality is from the fact that the unconstrained inf-value is no greater than the constrained one. The second inequality is from c(x)(, =, )0 and y(, free, )0 imply y T c(x) 0. 7

The Lagrangian Dual Problem for the Toy Example where x = ( 1 3 ; 1 3). minimize (x 1 1) 2 + (x 2 1) 2 subject to x 1 + 2x 2 1 0, 2x 1 + x 2 1 0; ϕ(y) = 1.25y 2 1 1.25y 2 2 2y 1 y 2 2y 1 2y 2, y 0. where y = ( 4 9 ; ) 4 9. max 1.25y 2 1 1.25y 2 2 2y 1 y 2 2y 1 2y 2 s.t. (y 1 ; y 2 ) 0. 8

The Lagrangian Dual of LP I Consider LP problem and its conic dual problem is given by (LP ) minimize c T x subject to Ax = b, x 0; (LD) maximize b T y subject to A T y + s = c, s 0. We now derive the Lagrangian Dual of (LP). Let the Lagrangian multipliers be y( free ) for equalities and s 0 for constraints x 0. Then the Lagrangian function would be L(x, y, s) = c T x y T (Ax b) s T x = (c A T y s) T x + b T y; where x is free. 9

The Lagrangian Dual of LP II Now consider the Lagrangian dual objective [ ϕ(y, s) = inf L(x, y, s) = inf x Rn (c A T y s) T x + b T y ]. x R n If (c A T y s) 0, then ϕ(y, s) =. Thus, in order to maximize ϕ(y, s), the dual must choose its variables (y, s 0) such that (c A T y s) = 0. This constraint, together with the sign constraint s 0, establish the Lagrangian dual problem: which is identical to the conic dual of LP. (LDP ) maximize b T y subject to A T y + s = c, s 0. 10

The Lagrangian Dual of LP with the Log-Barrier I For a fixed µ > 0, consider the problem min c T x µ n j=1 log(x j) s.t. Ax = b, x 0 Again, the non-negativity constraints can be ignored if the feasible region has an interior, that is, any minimizer must have x(µ) > 0. Thus, the Lagrangian function would be simply given by n n L(x, y) = c T x µ log(x j ) y T (Ax b) = (c A T y) T x µ log(x j ) + b T y. j=1 Then, the Lagrangian dual objective (we implicitly need x > 0 for the function to be defined) n ϕ(y) := inf L(x, y) = inf (c A T y) T x µ log(x j ) + b T y. x x j=1 j=1 11

The Lagrangian Dual of LP with the Log-Barrier II First, from the view point of the dual, the dual needs to choose y such that c A T y > 0, since otherwise the primal can choose x > 0 to make ϕ(y) go to. Now for any given y such that c A T y > 0, the inf problem has a unique finite close-form minimizer x Thus, x j = ϕ(y) = b T y + µ µ (c A T y) j, j = 1,..., n. n log(c A T y) j + nµ(1 log(µ)). j=1 Therefore, the dual problem, for any fixed µ, can be written as max y ϕ(y) = nµ(1 log(µ)) + max y [bt y + µ n log(c A T y) j ]. j=1 This is actually the LP dual with the Log-Barrier on dual inequality constraints c A T y 0. 12

Lagrangian Strong Duality Theorem Theorem 3 Let (GCO) be a convex minimization problem and the infimum f of (GCO) be finite, and the suprermum of (LDP) be ϕ. In addition, let (GCO) have an interior-point feasible solution with respect to inequality constraints, that is, there is ˆx such that all inequality constraints are strictly held. Then, f = ϕ, and (LDP) admits a maximizer y such that ϕ(y ) = f. Furthermore, if (GCO) admits a minimizer x, then y i c i (x ) = 0, i = 1,..., m. The assumption of interior-point feasible solution is called Constraint Qualification condition, which was also needed as a condition to prove the strong duality theorem for general Conic Linear Optimization. Note that the problem would be a convex minimization problem if all equality constraints are hyperplane or affine functions c i (x) = a i x b i, all other level sets are convex. 13

An Example where the Constraint Qualification Failed Consider the problem (x = (0; 0)): min x 1 s.t. x 2 1 + (x 2 1) 2 1 0, (y 1 0) x 2 1 + (x 2 + 1) 2 1 0, (y 2 0) L(x, y) = x 1 y 1 (x 2 1 + (x 2 1) 2 1) y 2 (x 2 1 + (x 2 + 1) 2 1). ϕ(y) = 1 + (y 1 y 2 ) 2 y 1 + y 2, (y 1, y 2 ) 0 Although there is no duality gap, but the dual does not admit a (finite) maximizer... In summary: The primal and Lagrangian dual may 1) have a duality gap; 2) have zero duality gap but the optimal solutions are not attainable; or 3) have zero duality gap and the optimal solutions are attainable. 14

Proof of Lagrangian Strong Duality Theorem where the Constraints c(x) 0 Consider the convex (?) set C := {(κ; s) : x s.t. f(x) κ, c(x) s}. Then, (f ; 0) is on the closure of C. From the supporting hyperplane theorem, there exists (y 0; y ) 0 such that y 0f inf (κ;s) C (y 0κ + (y ) T s). First, we show y 0, since otherwise one can choose some (0; s 0) such that the inequality is violated. Secondly, we show y 0 > 0, since otherwise one can choose (κ ; 0) if y < 0, or (0; s = c(ˆx) < 0) if y = 0 (then y 0), such that the above inequality is violated. Now let us divide both sides by y0 and let y := y /y0, we have f inf (κ + (κ;s) C (y ) T s) = inf (f(x) x (y ) T c(x)) = ϕ(y ) ϕ. Then, from the weak duality theorem, we must have f = ϕ. 15

If (GCO) admits a minimizer x, then f(x ) = f so that f(x ) inf x [ f(x) (y ) T c(x) ] f(x ) (y ) T c(x ) = f(x ) m yi c i (x ), i which implies that m yi c i (x ) 0. Since y i 0 and c i(x ) 0 for all i, it must be true y i c i(x ) = 0 for all i. i 16

More on Lagrangian Duality Consider the constrained problem with additional constraints Typically, Ω has a simple form such as the cone (GCO) inf f(x) s.t. c i (x) (, =, ) 0, i = 1,..., m, x Ω R n. Ω = R n + = {x : x 0} or the box Ω := {x : e x e.} Then, when derive the Lagrangian dual, there is not need to introduce multipliers for Ω constraints. 17

Lagrangian Relaxation Problem Consider again the (partial) Lagrangian Function: and define the dual objective function of y be L(x, y, s) = f(x) y T c(x), y Y ; ϕ(y) := inf x L(x, y) s.t. x Ω. Theorem 4 The Lagrangian dual function ϕ(y) is a concave function. Theorem 5 (Weak duality theorem) For every y Y, the Lagrangian dual function value ϕ(y) is less or equal to the infimum value of the original GCO problem. 18

The Lagrangian Dual Problem (LDP ) sup ϕ(y) s.t. y Y. would called the Lagrangian dual of the original GCO problem: Theorem 6 (Strong duality theorem) Let (GCO) be a convex minimization problem, the infimum f of (GCO) be finite, and the suprermum of (LDP) be ϕ. In addition, let (GCO) have an interior-point feasible solution with respect to inequality constraints, that is, there is ˆx such that all inequality constraints are strictly held. Then, f = ϕ, and (LDP) admits a maximizer y such that ϕ(y ) = f. Furthermore, if (GCO) admits a minimizer x, then y i c i (x ) = 0, i = 1,..., m. 19

The Lagrangian Dual of LP with Bound Constraints Consider (LP ) minimize c T x subject to Ax = b, e x e ( x 1); Let the Lagrangian multipliers be y for equality constraints. Then the Lagrangian dual objective would be ϕ(y) = inf L(x, y) = inf e x e e x e where if (c A T y) j 0, x j = 1; and otherwise, x j = 1. Therefore, the Lagrangian dual is [ (c A T y) T x + b T y ] ; (LDP ) maximize b T y c A T y 1 subject to y R m. 20

The Conic Duality vs. Lagrangian Duality I Consider SOCP problem and it conic dual problem (SOCP ) minimize c T x subject to Ax = b, x 1 x 1 2 0; (SOCD) maximize b T y subject to A T y + s = c, s 1 s 1 2 0. Let the Lagrangian multipliers be y for equalities and scalar s 0 for the single constraint x 1 x 1 2. Then the Lagrangian function would be L(x, y, s) = c T x y T (Ax b) s(x 1 x 1 2 ) = (c A T y) T x s(x 1 x 1 2 ) + b T y. 21

The Conic Duality vs. Lagrangian Duality II Now consider the Lagrangian dual objective [ ϕ(y, s) = inf L(x, y, s) = inf x Rn (c A T y) T x s(x 1 x 1 2 ) + b T y ]. x R n The objective function of the problem may not be differentiable so that the classical optimal condition theory do not apply. Consequently, it is difficult to write a clean form of the Lagrangian dual problem. On the other hand, many nonlinear optimization problems, even they are convex, are difficult to transform them into structured CLP problems (especially to construct the dual cones). Therefore, each of the duality form, Conic or Lagrangian, has its own pros and cons. 22