Lagrange Relaxation: Introduction and Applications

Similar documents
Lagrange duality. The Lagrangian. We consider an optimization program of the form

Lecture: Duality of LP, SOCP and SDP

5. Duality. Lagrangian

Convex Optimization Boyd & Vandenberghe. 5. Duality

Lecture: Duality.

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Convex Optimization M2

EE/AA 578, Univ of Washington, Fall Duality

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Solution Methods for Stochastic Programs

Primal/Dual Decomposition Methods

Duality in Linear Programs. Lecturer: Ryan Tibshirani Convex Optimization /36-725

subject to (x 2)(x 4) u,

Stochastic Programming: From statistical data to optimal decisions

The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 44

Convex Optimization & Lagrange Duality

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Part IB Optimisation

Lagrangian Duality and Convex Optimization

Duality. Geoff Gordon & Ryan Tibshirani Optimization /

Linear and Combinatorial Optimization

Math 273a: Optimization Subgradients of convex functions

Lecture 7: Lagrangian Relaxation and Duality Theory

Airline Network Revenue Management by Multistage Stochastic Programming

EE364a Review Session 5

Optimization for Machine Learning

Convex Optimization and SVM

Lagrangian Duality. Richard Lusby. Department of Management Engineering Technical University of Denmark

4. Algebra and Duality

Lecture 24 November 27

The Lagrangian L : R d R m R r R is an (easier to optimize) lower bound on the original problem:

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

3.10 Lagrangian relaxation

Decomposition Techniques in Mathematical Programming

MATH 829: Introduction to Data Mining and Analysis Support vector machines and kernels

Example Problem. Linear Program (standard form) CSCI5654 (Linear Programming, Fall 2013) Lecture-7. Duality

EE 227A: Convex Optimization and Applications October 14, 2008

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Lecture 7: Convex Optimizations

Lagrange Relaxation and Duality

ICS-E4030 Kernel Methods in Machine Learning

Lecture Note 5: Semidefinite Programming for Stability Analysis

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Lagrangian Relaxation in MIP

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Solving large Semidefinite Programs - Part 1 and 2

TMA947/MAN280 APPLIED OPTIMIZATION

9. Dual decomposition and dual algorithms

Outline. Relaxation. Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING. 1. Lagrangian Relaxation. Lecture 12 Single Machine Models, Column Generation

Lecture 10: Duality in Linear Programs

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

CS-E4830 Kernel Methods in Machine Learning

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

Homework Set #6 - Solutions

Convex Optimization Conjugate, Subdifferential, Proximation

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

Lagrangian Duality. Evelien van der Hurk. DTU Management Engineering

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Duality (Continued) min f ( x), X R R. Recall, the general primal problem is. The Lagrangian is a function. defined by

Stability of optimization problems with stochastic dominance constraints

Integer Programming ISE 418. Lecture 16. Dr. Ted Ralphs

BBM402-Lecture 20: LP Duality

Conjugate duality in stochastic optimization

The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 38

Linear Programming. Operations Research. Anthony Papavasiliou 1 / 21

Constrained Optimization and Lagrangian Duality

MVE165/MMG631 Linear and integer optimization with applications Lecture 8 Discrete optimization: theory and algorithms

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Benders Decomposition Methods for Structured Optimization, including Stochastic Optimization

Can Li a, Ignacio E. Grossmann a,

Can Li a, Ignacio E. Grossmann a,

Dual Decomposition.

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Lecture 2: Linear SVM in the Dual

Constrained optimization

Convex Optimization and Support Vector Machine

On a class of minimax stochastic programs

Lecture Notes on Support Vector Machine

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

On the Method of Lagrange Multipliers

Recita,on: Loss, Regulariza,on, and Dual*

4y Springer NONLINEAR INTEGER PROGRAMMING

Introduction to integer programming II

Stochastic Integer Programming

Pattern Classification, and Quadratic Problems

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Lecture 18: Optimization Programming

Convex Optimization and Modeling

Lagrangian Duality Theory

and to estimate the quality of feasible solutions I A new way to derive dual bounds:

Applications of Linear Programming

Dual and primal-dual methods

Dual Proximal Gradient Method

Convex Optimization. Ofer Meshi. Lecture 6: Lower Bounds Constrained Optimization

Solving Dual Problems

Transcription:

1 / 23 Lagrange Relaxation: Introduction and Applications Operations Research Anthony Papavasiliou

2 / 23 Contents 1 Context 2 Applications Application in Stochastic Programming Unit Commitment

3 / 23 Table of Contents 1 Context 2 Applications Application in Stochastic Programming Unit Commitment

4 / 23 When to Use Lagrange Relaxation Consider the following optimization problem: p = max f 0 (x) f (x) 0 h(x) = 0 with x D R n, f : R n R m, h : R n R l Context for Lagrange relaxation: 1 Complicating constraints f (x) 0 and h(x) = 0 make the problem difficult 2 Dual function is relatively easy to evaluate g(u, v) = sup(f 0 (x) u T f (x) v T h(x)) (1) x D

5 / 23 Idea of Dual Decomposition Dual function g(u, v) is convex regardless of primal problem Computation of g(u, v), π g(u, v) is relatively easy But... g(u, v) may be non-differentiable Idea: minimize g(u, v) using algorithms that rely on linear approximation of g(u, v) 1 Subgradient method 2 Cutting plane methods 3 Bundle methods

6 / 23 Why Minimize the Dual Function? We want to minimize the dual function for two reasons: Get as tight bounds to p as possible If we want to find primal optimal x by using the dual function, we have to minimize L(x, u, v) for (u, v) dual optimal

Why Minimize the Dual Function: Bounding p Proposition: If u 0 then g(u, v) p Proof: If x is feasible and u 0 then f 0 ( x) L( x, u, v) sup L(x, u, v) = g(u, v). x D Minimizing over all feasible x gives p g(u, v) Conclusion: minimizing g(u, v) gives tightest possible bound to p 7 / 23

Dual Function Properties Proposition: g(u, v) is convex lower-semicontinous [ 1. If (u, ] v) is such that (1) has optimal solution x u,v, then f (x u,v ) h(x u,v ) is a subgradient of g 1 A function is lower-semicontinuous when its epigraph is a closed subset of R m R l R. 8 / 23

Graphical Interpretation 9 / 23 Think of each x D as an index k K, then g(u, v) = max k K (f 0k u T f k v T h k )

10 / 23 Table of Contents 1 Context 2 Applications Application in Stochastic Programming Unit Commitment

11 / 23 Lagrange Relaxation of Stochastic Programs Consider 2-stage stochastic program: min f 1(x) + E ω [f 2(y(ω), ω)] s.t. h1 i (x) 0, i = 1,..., m 1, h2 i (x, y(ω), ω) 0, i = 1,..., m 2 Introduce non-anticipativity constraint x(ω) = x and reformulate problem as min E ω [f 1(x) + f 2(y(ω), ω)] s.t. h1 i (x) 0, i = 1,..., m 1, h2 i (x(ω), y(ω), ω) 0, i = 1,..., m 2 (ν(ω)) : x(ω) = x, ω Ω

Dual Function of Stochastic Program 12 / 23 Denote ν = (ν(ω), ω Ω), then g(ν) = g1(ν) + E ω g2(ν(ω), ω) where g1(ν) = inf x f 1(x) + ( ν(ω)) T x ω Ω s.t. h1 i (x) 0, i = 1,..., m 1, and g2(ν(ω), ω) = inf x(ω),y(ω) f 2(y(ω), ω) ν(ω) T x(ω) s.t. h2 i (x(ω), y(ω), ω) 0, i = 1,..., m 2

13 / 23 Duality and Problem Reformulations Equivalent formulations of a problem can lead to very different duals Reformulating the primal problem can be useful when the dual is difficult to derive, or uninteresting Common reformulations Introduce new variables and equality constraints (we have seen this already) Rearrange constraints in subproblems

14 / 23 Rearranging Constraints Note alternative relaxation to the previous stochastic program: min E ω [f 1(x(ω)) + f 2(y(ω), ω)] s.t. h1 i (x(ω)) 0, i = 1,..., m 1, h2 i (x(ω), y(ω), ω) 0, i = 1,..., m 2 x(ω) = x This relaxation is probably useless (because subproblem involving x has no constraints)

15 / 23 A Two-Stage Stochastic Integer Program [Sen, 2000] Scenario Constraints Binary Solutions ω = 1 2x 1 + y 1 2 and D 1 = {(0, 0), (1, 0)} 2x 1 y 1 0 ω = 2 x 2 y 2 0 D 2 = {(0, 0), (1, 0), (1, 1)} ω = 3 x 3 + y 3 1 D 3 = {(0, 0), (0, 1), (1, 0)} 3 equally likely scenarios Define x as first-stage decision, x ω is first-stage decision for scenario ω Non-anticipativity constraint: x 1 = 1 3 (x 1 + x 2 + x 3 )

Formulation of Problem and Dual Function max(1/3)y 1 + (1/3)y 2 + (1/3)y 3 s.t. 2x 1 + y 2 2 2x 1 y 1 0 x 2 y 2 0 x 3 + y 3 1 2 3 x 1 1 3 x 2 1 3 x 3 = 0, (λ) x ω, y ω {0, 1}, ω Ω = {1, 2, 3} g(λ) = max { 2λ (x ω,y ω) D ω 3 x 1 + 1 3 y 1 λ 3 x 2 + 1 3 y 2 λ 3 x 3 + 1 3 y 3} = max(0, 2λ/3) + max(0, ( λ + 1)/3) + max(1/3, λ/3) 16 / 23

Duality Gap 17 / 23 Dual function can be expressed equivalently 1 3 2 3 λ λ 1 2 g(λ) = 3 1 3 λ 1 λ 0 2 3 + 1 3 λ 0 λ 1 1 3 + 2 3 λ 1 λ Primal optimal value p = 1/3 with x 1 = x 2 = x 3 = 1, y 1 = 0, y 2 = 1, y 3 = 0 Dual optimal value d = 2/3 at λ = 0 Conclusion: We have a duality gap of 1/3

18 / 23 Alternative Relaxation Add new explicit first-stage decision variable x, with the following non-anticipativity constraints: x 1 = x, (λ 1 ) x 2 = x, (λ 2 ) x 3 = x, (λ 3 ) Dual function: g(λ) = max (x 1,y 1 ) D 1 { 1 3 y 1 λ 1 x 1 } + max (x 2,y 2 ) D 2 { 1 3 y 2 λ 2 x 2 } + max (x 3,y 3 ) D 3 { 1 3 y 3 λ 3 x 3 } + max x {0,1} {(λ 1 + λ 2 + λ 3 )x} = max(0, λ 1 ) + max(0, 1 3 λ 2) + max( 1 3, λ 3) + max(0, λ 1 + λ 2 + λ 3 )

Closing the Duality Gap 19 / 23 Choosing λ 1 = 0, λ 2 = 1 3 and λ 3 = 1 3, we have the following dual function value: g(0, 1/3, 1/3) = 0 + 0 + 1 3 + 0 = 1 3

20 / 23 Conclusions of Example Different relaxations can result in different duality gaps Computational trade-off: introducing more Lagrange multipliers results in better bounds but larger search space

Unit Commitment 21 / 23 Consider the following notation Schedule I power plants over horizon T Denote xi t as control of unit i in period t, denote x i [resp. x t ] as vector (xi t)t t=1 [resp. (x t i ) i I] Denote D i as feasible set of unit i Denote C i (x i ) as cost of producing x i by unit i Denote c t (x t ) 0 as a complicating constraint that needs to be satisfied collectively by units, and suppose that it is additive: c t (x t ) = i I ct i (x t i )

22 / 23 Unit commitment problem: min i I C i (x i ) (u t ) : x i D i, i I ci t (x i t ) 0, t = 1,..., T i I Relax complicating constraints to obtain the following Lagrangian: L(x, u) = i I (C i (x i ) + T t=1 u t c t i (x t i )) What have we gained? We can solve one problem per plant: min (C i (x i ) + x i D i T t=1 u t c t i (x t i ))

23 / 23 References [1] S. Boyd, "Subgradient methods", EE364b lecture slides, http://stanford.edu/class/ee364b/lectures/ [2] C. Lemaréchal, "Lagrangian Relaxation", Computational combinatorial optimization. Springer Berlin Heidelberg, pp. 112-156, 2001.