Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Similar documents
Convex Optimization Boyd & Vandenberghe. 5. Duality

EE/AA 578, Univ of Washington, Fall Duality

5. Duality. Lagrangian

Convex Optimization M2

Lecture: Duality.

Lecture: Duality of LP, SOCP and SDP

On the Method of Lagrange Multipliers

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

CSCI : Optimization and Control of Networks. Review on Convex Optimization

A Brief Review on Convex Optimization

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Convex Optimization & Lagrange Duality

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Tutorial on Convex Optimization for Engineers Part II

Convex Optimization and Modeling

Lagrangian Duality and Convex Optimization

Lecture 18: Optimization Programming

The Lagrangian L : R d R m R r R is an (easier to optimize) lower bound on the original problem:

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

ICS-E4030 Kernel Methods in Machine Learning

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

CS-E4830 Kernel Methods in Machine Learning

EE364a Review Session 5

Lecture 7: Convex Optimizations

Constrained Optimization and Lagrangian Duality

Lecture 8. Strong Duality Results. September 22, 2008

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

12. Interior-point methods

Tutorial on Convex Optimization: Part II

Convex Optimization and SVM

Optimization for Machine Learning

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

12. Interior-point methods

Interior Point Algorithms for Constrained Convex Optimization

Optimality, Duality, Complementarity for Constrained Optimization

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

4TE3/6TE3. Algorithms for. Continuous Optimization

EE 227A: Convex Optimization and Applications October 14, 2008

Lecture 6: Conic Optimization September 8

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

The Karush-Kuhn-Tucker conditions

Part IB Optimisation

Machine Learning. Lecture 6: Support Vector Machine. Feng Li.

Convex Optimization and Support Vector Machine

Lecture Notes on Support Vector Machine

CS711008Z Algorithm Design and Analysis

Support Vector Machines: Maximum Margin Classifiers

subject to (x 2)(x 4) u,

Support vector machines

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Lagrangian Duality Theory

COM S 578X: Optimization for Machine Learning

Primal/Dual Decomposition Methods

Optimality Conditions for Constrained Optimization

Numerical Optimization

Chap 2. Optimality conditions

Lecture Note 5: Semidefinite Programming for Stability Analysis

Applications of Linear Programming

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

CONSTRAINT QUALIFICATIONS, LAGRANGIAN DUALITY & SADDLE POINT OPTIMALITY CONDITIONS

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 6 Fall 2009

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Convex Optimization M2

Lecture: Introduction to LP, SDP and SOCP

Support Vector Machines for Regression

Linear and non-linear programming

Homework Set #6 - Solutions

Convexification of Mixed-Integer Quadratically Constrained Quadratic Programs

10 Numerical methods for constrained problems

Lecture 14: Optimality Conditions for Conic Problems

Rate Control in Communication Networks

Nonlinear Programming and the Kuhn-Tucker Conditions

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Optimisation convexe: performance, complexité et applications.

Lagrangian Duality. Richard Lusby. Department of Management Engineering Technical University of Denmark

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

Constrained Optimization

Constrained optimization

ECE Optimization for wireless networks Final. minimize f o (x) s.t. Ax = b,

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

1. f(β) 0 (that is, β is a feasible point for the constraints)

Duality Uses and Correspondences. Ryan Tibshirani Convex Optimization

Math 273a: Optimization Subgradients of convex functions

Lecture 3. Optimization Problems and Iterative Algorithms

The Karush-Kuhn-Tucker (KKT) conditions

A Tutorial on Convex Optimization II: Duality and Interior Point Methods

CS711008Z Algorithm Design and Analysis

Support Vector Machines

4. Algebra and Duality

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

Lecture 2: Linear SVM in the Dual

9. Dual decomposition and dual algorithms

Generalization to inequality constrained problem. Maximize

Additional Homework Problems

Support Vector Machines

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Transcription:

Duality Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Lagrangian Consider the optimization problem in standard form min f 0 (x) s.t. f i (x) 0, i = 1,,m h i (x) = 0, i = 1,,p variable x R n, domain D = m i=0 domf i p i=1 domh i, optimal value p Lagrangian: L : R n R m R p R, with dom L = D R m R p m p L(x,λ,ν) = f 0 (x)+ λ i f i (x)+ ν i h i (x) i=1 i=1 λ i : Lagrange multiplier associated with f i (x) 0 ν i : Lagrange multiplier associated with h i (x) = 0

Lagrangian dual function Lagrangian dual function: g : R m R p R, g(λ,ν) = inf x D L(x,λ,ν) = inf x D f 0(x)+ m λ i f i (x)+ i=1 g is concave, can be for some λ,ν p ν i h i (x) i=1

Lagrangian dual function: lower bound property Theorem (lower bounds on optimal value) For any λ ર 0 and any ν, we have g(λ,ν) p Proof. Suppose x is a feasible point. Since λ ર 0, m λ i f i ( x)+ i=1 Therefore L( x,λ,ν) f 0 ( x). Hence p ν i h i ( x) 0 i=1 g(λ,ν) = inf x D L(x,λ,ν) L( x,λ,ν) f 0( x) which holds for every feasible x. Thus the lower bound follows.

The dual problem Lagrange dual problem max g(λ,ν) s.t. λ ર 0 find the best lower bound on p a convex optimization problem; optimal value denoted d λ,ν are dual feasible if λ ર 0,(λ,ν) domg (means g(λ,ν) > ) (λ,ν ): dual optimal multipliers p d is called the optimal duality gap

Weak and strong duality Weak duality: d p always true (for both convex and nonconvex problems) Strong duality: d = p does not hold in general (usually) holds for convex problems conditions that guarantee strong duality in convex problems are called constraint qualifications

Slater s constraint qualification Consider the standard convex optimization problem min f 0 (x) s.t. f i (x) 0, i = 1,,m Ax = b with variable x R n, domain D = m i=0 domf i. Slater s condition: exists a point that is strictly feasible, i.e., x relintd such that f i (x) < 0, i = 1,,m, Ax = b (interior relative to affine hull) can be relaxed: affine inequalities do not need to hold with strict inequalities Slater s theorem: The strong duality holds if the Slater s condition holds and the problem is convex.

Complementary slackness Suppose strong duality holds; x is primal optimal; (λ,ν ) is dual optimal f 0 (x ) = g(λ,ν ) = inf x D f 0(x)+ f 0 (x )+ f 0 (x ) m λ i f i(x)+ i=1 m λ i f i (x )+ i=1 Hence the inequalities must hold with equality x minimizes L(x,λ,ν ) λ i f i(x ) = 0 for all i = 1,,m: p νi h i(x) i=1 p νi h i (x ) i=1 λ i = 0 = f i (x ) = 0, f i (x ) < 0 = λ i = 0 known as complementary slackness

Karush-Kuhn-Tucker (KKT) conditions If strong duality holds, x is primal optimal, (λ,ν) is dual optimal, and f i,h i are differentiable, then the following four conditions (called KKT conditions) must hold 1. primal constraints: f i (x) 0,i = 1,,m,h i (x) = 0,i = 1,,p 2. dual constraints: λ ર 0 3. complementary slackness: λ i f i(x ) = 0 4. gradient of Lagrangian w.r.t. x vanishes: f 0 (x)+ m λ i f i (x)+ i=1 p ν i h i (x) = 0 i=1

KKT conditions for convex problem If x, λ, ν satisfy KKT for a convex problem, then they are optimal: from complementary slackness: f 0 ( x) = L( x, λ, ν) from the 4-th condition and convexity: g( λ, ν) = L( x, λ, ν) So f 0 ( x) = g( λ, ν) If the Slater s condition is satisfied and f i is differentiable, then x is optimal iff λ,ν that satisfy KKT

Minimax interpretation Given Lagrangian The primal problem: L(x,λ,ν) = f 0 (x)+λ T f(x)+ν T h(x) (P) inf sup x P λર0,ν L(x,λ,ν) The dual problem: (D) sup λર0,ν inf L(x,λ,ν) x P Weak duality: sup λર0,ν inf x P L(x,λ,ν) inf x P sup λર0,ν L(x,λ,ν)

Saddle point implies strong duality Strong duality: sup λર0,ν inf x P L(x,λ,ν) = inf x P sup λર0,ν L(x,λ,ν) Saddle-point interpretation: (x,λ,ν ) is a saddle point of L if for all λ ર 0,ν,x P. L(x,λ,ν) L(x,λ,ν ) L(x,λ,ν ) The strong duality holds if (x,λ,ν ) a saddle point of L

Examples

Standard form LP (P) min c T x s.t. Ax = b, x ર 0 (D) max b T ν A T ν +c ર 0

Quadratic program (P) min x T Px s.t. Ax b (assume P S n ++ ) (D) max 1 4 λt AP 1 A T λ b T λ s.t. λ ર 0

Equality constrained norm minimization (P) min x s.t. Ax = b (D) max b T ν A T ν 1 Note: y = sup{x T y x 1} is the dual norm of. inf x ( x y T x) = 0 if y 1 and otherwise.

Two-way partitioning (P) min x T Wx s.t. x 2 i = 1, i = 1,,n nonconvex; feasible set: 2 n discrete points partition {1,,n} into two sets; W ij cost of associating i,j to the same set; W ij cost of assigning to different sets (D) max 1 T ν s.t. W +diag(ν) ર 0 lower bound example: ν = λ min (W)1 gives p nλ min (W)

Perturbation and sensitivity anaysis perturbed optimization problem min f 0 (x) s.t. f i (x) u i, i = 1,,m h i (x) = v i, i = 1,,p its dual min g(λ,ν) u T λ v T ν s.t. λ ર 0 x primal variable; u,v are parameters p (u,v) is optimal value as a function of u,v

Global sensitivity anaysis assume strong duality holds for unperturbed problem (u = 0,v = 0), and λ,ν are dual optimal for unperturbed problem By weak duality on the perturbed problem: sensitivity interpretation: p (u,v) g(λ,ν ) u T λ v T ν p (0,0) u T λ v T ν λ i large: small u i = large change in p ν i large and positive: v i < 0 = large increase in p

Local sensitivity analysis: LP Consider LP (P) min c T x s.t. Ax y (D) max y T λ A T λ+c = 0, λ ર 0 The optimal value: p = y T λ. Thus p y i yi=0= λ i

Local sensitivity analysis Suppose strong duality holds for unperturbed problem (u = 0,v = 0), and λ,ν are dual optimal for unperturbed problem. If p (u,v) is differentiable at (0, 0), then λ i = p (0,0) u i ν i = p (0,0) v i If p (u,v) is not differentiable at (0,0), then ( λ, ν ) p (0,0). Proof. By the weak duality on the perturbed problem: p (0,0) u i p (0,0) u i Thus equality holds. Similar proof for ν i p (te i,0) p (0,0) = lim λ i t 0 t p (te i,0) p (0,0) = lim λ i t 0 t

Problems with generalized inequalities Lagrangian min f 0 (x) s.t. f i (x) ર Ki 0, i = 1,,m h i (x) = v i, i = 1,,p λ i : Lagrange multiplier for f i (x) Ki 0 Lagrangian L: L(x,λ 1,,λ m,ν) = f 0 (x)+ Dual function g: m λ T i f i (x)+ i=1 p ν i h i (x) i=1 g(λ 1,,λ m,ν) = inf x D L(x,λ 1,,λ m,ν,x)

Problems with generalized inequalities: dual problem Theorem (lower bound property) λ i ર K i 0, i = 1,,m = g(λ 1,,λ m,ν) p Lagrange dual problem max g(λ 1,,λ m,ν) s.t. λ i ર K i 0, i = 1,,m weak duality: p d p d : optimal duality gap strong duality: p = d for convex problem with constraint qualification. Slater s: primal problem is strictly feasible

Semidefinite program (SDP) Primal SDP where F i,g S k Dual SDP where Z S k. min c T x s.t. x 1 F 1 + +x n F n G max tr(gz) s.t. tr(f i Z)+c i = 0, i = 1,,n Z ર 0 Strong duality if primal SDP is strictly feasible, i.e. x with x 1 F 1 + +x n F n G