The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 44
|
|
- Grant Mathews
- 5 years ago
- Views:
Transcription
1 1 / 44 The L-Shaped Method Operations Research Anthony Papavasiliou
2 Contents 2 / 44 1 The L-Shaped Method [ 5.1 of BL] 2 Optimality Cuts [ 5.1a of BL] 3 Feasibility Cuts [ 5.1b of BL] 4 Proof of Convergence [ 5.1c of BL] 5 Example: Capacity Expansion Planning
3 Table of Contents 3 / 44 1 The L-Shaped Method [ 5.1 of BL] 2 Optimality Cuts [ 5.1a of BL] 3 Feasibility Cuts [ 5.1b of BL] 4 Proof of Convergence [ 5.1c of BL] 5 Example: Capacity Expansion Planning
4 4 / 44 Extensive Form 2-Stage Stochastic Linear Program K (EF ) : min c T x + p k qk T y k s.t. Ax = b k=1 T k x + Wy k = h k, k = 1,..., K x 0, y k 0, k = 1,..., K K realizations of random vector ξ, with probabilities p k, k = 1,..., K Randomness: q k, h k, T k y k : second-stage decision given realization k
5 Block Structure 5 / 44 Multi-stage decision making process A T 1 W T 2 W. T K W Idea: ignore constraints of future stages
6 6 / 44 Master Problem We know that K V (x) = {min p k qk T y k Wy k = h k T k x, y k 0} k=1 is a piecewise linear function of x Define master problem as (M) : min z = c T x + θ (1) s.t. Ax = b D l x d l, l = 1,..., r (2) E l x + θ e l, l = 1,..., s (3) x 0, θ R Feasibility cuts: equation 2 Optimality cuts: equation 3
7 Optimality Cuts 7 / 44 Consider a trial first-stage decision x v Let πk v be simplex multipliers of second-stage problem: min w = qk T y s.t. Wy = h k T k x v y 0 e s+1 E s+1 x supports V (x) at x v, where E s+1 = e s+1 = K p k (πk v )T T k (4) k=1 K p k (πk v )T h k (5) k=1
8 8 / 44 V (x) e s+1 E s+1 x x v x
9 9 / 44 The L-Shaped Algorithm Step 0. Set r = s = v = 0 Step 1. Solve master problem (M). Let (x v, θ v ) be an optimal solution. If s = 0 (no optimality cuts), remove θ from (M) and set θ 0 = Step 2. If x / K 2, add feasibility cut (equation 2) and return to Step 1. Otherwise, go to Step 3. K 2 = {x y : Wy = h k T k x, y 0, k = 1,..., K } Step 3. Compute E s+1, e s+1. Let w v = e s+1 E s+1 x v. If θ v w v, stop with x v an optimal solution. Otherwise, set s = s + 1, add optimality cut to equation 3 and return to Step 1.
10 Table of Contents 10 / 44 1 The L-Shaped Method [ 5.1 of BL] 2 Optimality Cuts [ 5.1a of BL] 3 Feasibility Cuts [ 5.1b of BL] 4 Proof of Convergence [ 5.1c of BL] 5 Example: Capacity Expansion Planning
11 11 / 44 Example 1 z = min 100x x 2 + E ξ (q 1 y 1 + q 2 y 2 ) s.t. x 1 + x y y 2 60x 1 8y 1 + 5y 2 80x 2 y 1 d 1, y 2 d 2 x 1 40, x 2 20, y 1, y 2 0 ξ = (d 1, d 2, q 1, q 2 ) = { (500, 100, 24, 28), p 1 = 0.4 (300, 300, 28, 32), p 2 = 0.6
12 Iteration 1 Step 1. min{100x x 2 x 1 + x 2 120, x 1 40, x 2 20} x 1 = (40, 20) T, θ 1 = Step 3. For ξ = ξ 1 solve min{ 24y 1 28y 2 6y y , 8y 1 + 5y y 1 500, 0 y 2 100} w 1 = 6100, y T = (137.5, 100), π T 1 = (0, 3, 0, 13) For ξ = ξ 2 solve min{ 28y 1 32y 2 6y y , 8y 1 + 5y y 1 300, 0 y 2 300} w 2 = 8384, y T = (80, 192), π2 T = ( 2.32, 1.76, 0, 0) 12 / 44
13 Iteration 1: Optimality Cut h 1 = (0, 0, 500, 100) T, h 2 = (0, 0, 300, 300) T T,1 = ( 60, 0, 0, 0) T, T,2 = (0, 80, 0, 0) T e 1 = 0.4 π1 T h π2 T h 2 = 0.4 ( 1300)+0.6 (0) = 520 E 1 = 0.4 π1 T T πt 2 T = 0.4(0, 240) + 0.6(139.2, 140.8) = (83.52, ) w 1 = 520 (83.52, ) x 1 = w 1 = > θ 1 =, therefore add the cut 83.52x x 2 + θ / 44
14 Iteration 2 14 / 44 Step 1. Solve master min{100x x 2 + θ x 1 + x 2 120, x 1 40, x 2 20, 83.52x x 2 + θ 520} z = , x 2 = (40, 80) T, θ 2 = Step 3. Add the cut 211.2x 1 + θ 1584
15 Iteration 3 15 / 44 Step 1. Solve master. z = , x 3 = (66.828, ) T, θ 3 = Step 3. Add the cut 115.2x x 2 + θ 2104
16 16 / 44 Iteration 4 Step 1. Solve master. z = 889.5, x 4 = (40, 33.75) T, θ 4 = 9952 Step 3. There are multiple solutions for ξ = ξ 2. Select one, add the cut x x 2 + θ 0
17 17 / 44 Iteration 5 Step 1. Solve master min{100x x 2 + θ x 1 + x 2 120, x 1 40, x 2 20, 83.52x x 2 + θ 520, 211.2x 1 + θ x x 2 + θ 2104, x x 2 + θ 0} z = , x 5 = (46.667, 36.25) T, θ 5 = Step 3. w 5 = 520 (83.52, ) x 5 = = θ 5, stop. x = (46.667, 36.25) T is the optimal solution.
18 18 / 44 Example 2 z = min E ξ (y 1 + y 2 ) s.t. 0 x 10 y 1 y 2 = ξ x y 1, y p 1 = 1/3 ξ = 2 p 2 = 1/3 4 p 3 = 1/3 K 2 = R
19 L-Shaped Method in Example 2 19 / 44 Iteration 1, Step 1: x 1 = 0 Iteration 1, Step 3: x 1 not optimal, add cut: θ 7/3 x Iteration 2, Step 1: x 2 = 10 Iteration 2, Step 3: x 2 not optimal, add cut: θ x 7/3 Iteration 3, Step 1: x 3 = 7/3 Iteration 3, Step 3: x 3 not optimal, add cut: θ (x + 1)/3 Iteration 4, Step 1: x 4 = 1.5 Iteration 4, Step 3: x 4 not optimal, add cut: θ (5 x)/3 Iteration 3, Step 1: x 5 = 2 Iteration 3, Step 3: x 5 is optimal
20 20 / 44 V (x) 1 x 1 1 x 3 4 V (x 1 ) = 7/3 and V (x) = 7/3 x around x 1 (7 x)/3 is the optimality cut at x 1
21 Geometric Interpretation of Optimality Cuts Q(x, ξ) = min y {q(ω) T y W (ω)y = h(ω) T (ω)x, y 0} yields dual optimal multipliers π T = q B (ω) T B(ω) 1 From linear programming duality Q(x, ξ) = q B (ω) T B(ω) 1 (h(ω) T (ω)x) B(ω) is optimal basis as long as y B = B(ω) 1 (h(ω) T (ω)x) 0, y N = 0 and q B (ω) T B(ω) 1 W q(ω) T Taking expectation, V (x) = E ξ {q B (ω) T B(ω) 1 (h(ω) T (ω)x)} in the set ξ Ξ {x B(ω) 1 (h(ω) T (ω)x) 0} This is exactly the expression for optimality cuts at x v 21 / 44
22 Table of Contents 22 / 44 1 The L-Shaped Method [ 5.1 of BL] 2 Optimality Cuts [ 5.1a of BL] 3 Feasibility Cuts [ 5.1b of BL] 4 Proof of Convergence [ 5.1c of BL] 5 Example: Capacity Expansion Planning
23 Feasibility Cuts 23 / 44 Consider the following problem: (F) : min w = e T v + + e T v s.t. Wy + Iv + Iv = h k T k x v y 0, v + 0, v 0 with dual multipliers σ v. Define D r+1 = (σ v ) T T k d r+1 = (σ v ) T h k Step 2 of L-shaped method: For k = 1,..., K solve (F). If w = 0 for all k, go to Step 3. Else, add D r+1 x d r+1, set r = r + 1 and go to Step 1.
24 24 / 44 Example min 3x 1 + 2x 2 E ξ (15y y 2 ) s.t. 3y 1 + 2y 2 x 1, 2y 1 + 5y 2 x 2 0.8ξ 1 y 1 ξ 1, 0.8ξ 2 y 2 ξ 2 x, y 0 ξ = (4, 4), p 1 = 0.25 (4, 8), p 2 = 0.25 (6, 4), p 3 = 0.25 (6, 8), p 4 = 0.25
25 Generating a Feasibility Cut For x 1 = (0, 0) T, ξ = (6, 8) T, solve min v + v +,v 1 + v 1 + v v 2 + v v 3 +,y v v 4 + v v 5 + v v 6 s.t. v + 1 v 1 + 3y 1 + 2y 2 0, v + 2 v 2 + 2y 1 + 5y 2 0 v + 3 v 3 + y 1 4.8, v + 4 v 4 + y v + 5 v 5 + y 1 6, v + 6 v 6 + y 2 8 We get w = 11.2, σ 1 = ( 3/11, 1/11, 1, 1, 0, 0) h = (0, 0, 4.8, 6.4, 6, 8) T, T,1 = ( 1, 0, 0, 0, 0, 0) T, T,2 = (0, 1, 0, 0, 0, 0) T D 1 = ( 3/11, 1/11, 1, 1, 0, 0) T = (3/11, 1/11), d 1 = ( 3/11, 1/11, 1, 1, 0, 0) h = /11x 1 + 1/11x / 44
26 26 / 44 Induced Constraints Going by the book: Iteration 2 master problem: x 2 = (41.067, 0) T Iteration 2 feasibility cut: x Iteration 3 master problem: x 3 = (33.6, 22.4) T Iteration 3 feasibility cut: x Iteration 4 master problem: x 4 = (27.2, 41.6) T is feasible Induced Constraints: Observe that for ξ = (6, 8) T, y 1 4.8, y This implies x , x , which should be added directly to the master Personal experience: feasibility cuts are impractical
27 27 / 44 Complete Recourse and Relative Complete Recourse pos W = {z z = Wy, y 0} K 1 = {x Ax = b}, K 2 = {x y : Wy = h k T k x, k = 1,..., K } Feasibility cuts are not necessary when we have: Complete recourse: pos W = R m 2 Relative complete recourse: K 2 K 1
28 Table of Contents 28 / 44 1 The L-Shaped Method [ 5.1 of BL] 2 Optimality Cuts [ 5.1a of BL] 3 Feasibility Cuts [ 5.1b of BL] 4 Proof of Convergence [ 5.1c of BL] 5 Example: Capacity Expansion Planning
29 29 / 44 Proof of Convergence Key observation: when we solve the master problem (M) subject to x K 1 K 2 either θ v V (x v ) or θ v < V (x v ) The first case is our termination criterion and implies optimality. The second case can occur finitely many times. K 1 = {x Ax = b, x 0}, K 2 = {x y : Wy = h k T k x, y 0, k = 1,..., K }
30 30 / 44 θ v V (x v ) Is Equivalent to the Termination Criterion Q(x, ξ(ω)) = min y {q(ω) T y Wy = h(ω) T (ω)x, y 0} From linear programming duality, Q(x v, ξ k ) = (π v k )T (h k T k x v ) Taking expectation, V (x) = E ξ Q(x v, ξ) = K k=1 p k (π v k )T (h k T k x v ) Our termination criterion is θ v w v, where K K w v = e s+1 E s+1 x v = ( p k (πk v )T h k ) ( p k (πk v )T T k )x v k=1 k=1
31 31 / 44 θ v V (x v ) Implies Optimality Fact: Q(x, ξ) is convex in x, and π v k is a subgradient at x v : Q(x, ξ k ) (π v k )T h k (π v k )T T k x Taking expectation: V (x) K k=1 p k (π v k )T (h k T k )x Full problem can be written as min c T x + θ s.t. V (x) θ x K 1 K 2 Therefore master (M) is a relaxation and gives lower bound If there were x K 1 K 2 with c T x + V ( x) < c T x v + V (x v ) then x v would not be optimal for (M) since (M) would be at most c T x + V ( x)
32 θ v < V (x v ) Occurs Finitely Many Times 32 / 44 We have seen that each optimality cut is equivalent to θ V (x l ) + V (x l )(x x l ), l = 1,..., s Every time θ v < V (x v ) occurs, the set of new multipliers (πk v ), k = 1,..., K must be different from those generated previously Optimal multipliers πk v correspond to optimal bases of min y {qk T y Wy = h k T k x, y 0} There are finitely many bases, therefore finitely many combinations of optimal multipliers (πk v ), k = 1,..., K
33 33 / 44 Finite Termination of Feasibility Cuts Key observations: 1 Only a finite number of feasibility cuts can be generated 2 The feasibility cuts of equation 2 are valid Valid feasibility cuts: remove current candidate x v from consideration do not remove any other candidates x K 2 from consideration
34 Finite Number of Possible Cuts 34 / 44 Recall feasibility cuts: D l x d l, l = 1,..., r + 1, where D r+1 = (σ v ) T T k d r+1 = (σ v ) T h k and σ v the dual multipliers of (F) : min w = e T v + + e T v s.t. Wy + Iv + Iv = h k T k x v y 0, v + 0, v 0 σ v corresponds to a basis B of W : σ v = c T B B 1 W has a finite number of bases
35 35 / 44 Valid Feasibility Cuts: x v Is Cut Off If w > 0 then by linear programming duality (σ v ) T (h k T k x v ) > 0, therefore D r+1 x d r+1 removes x v from consideration
36 Valid Feasibility Cuts: No Other x K 2 Is Cut Off Consider the following subset of R m 2: {z z = h k T k x, x K 2 } This set is contained in pos W = {z z = Wy, y 0} by definition of K 2 The half-space defined by (σ v ) T z 0 contains pos W To see this, note that the reduced cost of y must be 0: c T N c T B B 1 N = (σ v ) T W 0 where c T N = 0, (σv ) T = c T B B 1, N = W Any element of pos W can be expressed as z = Wy, y 0, hence from the previous inequality (σ v ) T z 0 Therefore, (σ v ) T (h Tx) 0 will be a valid feasibility cut 36 / 44
37 Feasibility Cuts 37 / 44 h k T k x v R m 2 σ r+1 pos W
38 Table of Contents 38 / 44 1 The L-Shaped Method [ 5.1 of BL] 2 Optimality Cuts [ 5.1a of BL] 3 Feasibility Cuts [ 5.1b of BL] 4 Proof of Convergence [ 5.1c of BL] 5 Example: Capacity Expansion Planning
39 Example: Capacity Expansion Planning min x,y 0 s.t. n m (I i x i + E ξ C i T j y ij (ω)) i=1 j=1 n y ij (ω) = D j (ω), j = 1,..., m i=1 m y ij (ω) x i, i = 1,... n j=1 I i, C i : fixed/variable cost of technology i D j (ω), T j : height/width of load block j y ij (ω): capacity of i allocated to j x i : capacity of i Note: D j is not dependent on ω 39 / 44
40 Problem Data Two possible realizations of load duration curve: Reference scenario: 10% 10x wind scenario: 90% Duration (hours) Level (MW) Level (MW) Reference scenario 10x wind scenario Base load Medium load Peak load / 44
41 Slave Problem 41 / 44 (S ω ) : min y 0 (λ j (ω)) : (ρ i (ω)) : n i=1 j=1 m C i T j y ij n y ij = D j (ω), j = 1,..., m i=1 m y ij x i, i = 1,... n j=1 where x has been fixed from the master problem
42 Sequence of Investment Decisions Iteration Coal (MW) Gas (MW) Nuclear (MW) Oil (MW) / 44
43 43 / 44 Observations Investment candidate in each iteration necessarily different from all past iterations Greedy behavior: low capital cost in early iterations
The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 38
1 / 38 The L-Shaped Method Operations Research Anthony Papavasiliou Contents 2 / 38 1 The L-Shaped Method 2 Example: Capacity Expansion Planning 3 Examples with Optimality Cuts [ 5.1a of BL] 4 Examples
More informationStochastic Integer Programming
IE 495 Lecture 20 Stochastic Integer Programming Prof. Jeff Linderoth April 14, 2003 April 14, 2002 Stochastic Programming Lecture 20 Slide 1 Outline Stochastic Integer Programming Integer LShaped Method
More informationSolution Methods for Stochastic Programs
Solution Methods for Stochastic Programs Huseyin Topaloglu School of Operations Research and Information Engineering Cornell University ht88@cornell.edu August 14, 2010 1 Outline Cutting plane methods
More informationLinear Programming. Operations Research. Anthony Papavasiliou 1 / 21
1 / 21 Linear Programming Operations Research Anthony Papavasiliou Contents 2 / 21 1 Primal Linear Program 2 Dual Linear Program Table of Contents 3 / 21 1 Primal Linear Program 2 Dual Linear Program Linear
More informationLagrange Relaxation: Introduction and Applications
1 / 23 Lagrange Relaxation: Introduction and Applications Operations Research Anthony Papavasiliou 2 / 23 Contents 1 Context 2 Applications Application in Stochastic Programming Unit Commitment 3 / 23
More informationLinear programming. Saad Mneimneh. maximize x 1 + x 2 subject to 4x 1 x 2 8 2x 1 + x x 1 2x 2 2
Linear programming Saad Mneimneh 1 Introduction Consider the following problem: x 1 + x x 1 x 8 x 1 + x 10 5x 1 x x 1, x 0 The feasible solution is a point (x 1, x ) that lies within the region defined
More informationA Benders Algorithm for Two-Stage Stochastic Optimization Problems With Mixed Integer Recourse
A Benders Algorithm for Two-Stage Stochastic Optimization Problems With Mixed Integer Recourse Ted Ralphs 1 Joint work with Menal Güzelsoy 2 and Anahita Hassanzadeh 1 1 COR@L Lab, Department of Industrial
More informationBenders Decomposition Methods for Structured Optimization, including Stochastic Optimization
Benders Decomposition Methods for Structured Optimization, including Stochastic Optimization Robert M. Freund April 29, 2004 c 2004 Massachusetts Institute of echnology. 1 1 Block Ladder Structure We consider
More informationThe Primal-Dual Algorithm P&S Chapter 5 Last Revised October 30, 2006
The Primal-Dual Algorithm P&S Chapter 5 Last Revised October 30, 2006 1 Simplex solves LP by starting at a Basic Feasible Solution (BFS) and moving from BFS to BFS, always improving the objective function,
More informationCOT 6936: Topics in Algorithms! Giri Narasimhan. ECS 254A / EC 2443; Phone: x3748
COT 6936: Topics in Algorithms! Giri Narasimhan ECS 254A / EC 2443; Phone: x3748 giri@cs.fiu.edu https://moodle.cis.fiu.edu/v2.1/course/view.php?id=612 Gaussian Elimination! Solving a system of simultaneous
More informationStochastic Programming: Basic Theory
Stochastic Programming: Basic Theory John R. Birge John.Birge@ChicagoBooth.edu www.chicagobooth.edu/fac/john.birge JRBirge ICSP, Bergamo, July 2013 1 Goals Provide background on the basics of stochastic
More informationDecomposition Algorithms for Two-Stage Distributionally Robust Mixed Binary Programs
Decomposition Algorithms for Two-Stage Distributionally Robust Mixed Binary Programs Manish Bansal Grado Department of Industrial and Systems Engineering, Virginia Tech Email: bansal@vt.edu Kuo-Ling Huang
More informationStochastic Decomposition
IE 495 Lecture 18 Stochastic Decomposition Prof. Jeff Linderoth March 26, 2003 March 19, 2003 Stochastic Programming Lecture 17 Slide 1 Outline Review Monte Carlo Methods Interior Sampling Methods Stochastic
More informationGeneration and Representation of Piecewise Polyhedral Value Functions
Generation and Representation of Piecewise Polyhedral Value Functions Ted Ralphs 1 Joint work with Menal Güzelsoy 2 and Anahita Hassanzadeh 1 1 COR@L Lab, Department of Industrial and Systems Engineering,
More informationBenders Decomposition Methods for Structured Optimization, including Stochastic Optimization
Benders Decomposition Methods for Structured Optimization, including Stochastic Optimization Robert M. Freund May 2, 2001 Block Ladder Structure Basic Model minimize x;y c T x + f T y s:t: Ax = b Bx +
More informationModeling Uncertainty in Linear Programs: Stochastic and Robust Linear Programming
Modeling Uncertainty in Linear Programs: Stochastic and Robust Programming DGA PhD Student - PhD Thesis EDF-INRIA 10 November 2011 and motivations In real life, Linear Programs are uncertain for several
More informationFinancial Optimization ISE 347/447. Lecture 21. Dr. Ted Ralphs
Financial Optimization ISE 347/447 Lecture 21 Dr. Ted Ralphs ISE 347/447 Lecture 21 1 Reading for This Lecture C&T Chapter 16 ISE 347/447 Lecture 21 2 Formalizing: Random Linear Optimization Consider the
More informationLesson 27 Linear Programming; The Simplex Method
Lesson Linear Programming; The Simplex Method Math 0 April 9, 006 Setup A standard linear programming problem is to maximize the quantity c x + c x +... c n x n = c T x subject to constraints a x + a x
More informationDecomposition Algorithms with Parametric Gomory Cuts for Two-Stage Stochastic Integer Programs
Decomposition Algorithms with Parametric Gomory Cuts for Two-Stage Stochastic Integer Programs Dinakar Gade, Simge Küçükyavuz, Suvrajeet Sen Integrated Systems Engineering 210 Baker Systems, 1971 Neil
More informationReformulation and Sampling to Solve a Stochastic Network Interdiction Problem
Network Interdiction Stochastic Network Interdiction and to Solve a Stochastic Network Interdiction Problem Udom Janjarassuk Jeff Linderoth ISE Department COR@L Lab Lehigh University jtl3@lehigh.edu informs
More informationScenario-Free Stochastic Programming
Scenario-Free Stochastic Programming Wolfram Wiesemann, Angelos Georghiou, and Daniel Kuhn Department of Computing Imperial College London London SW7 2AZ, United Kingdom December 17, 2010 Outline 1 Deterministic
More informationSelected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A.
. Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A. Nemirovski Arkadi.Nemirovski@isye.gatech.edu Linear Optimization Problem,
More informationThe Sample Average Approximation Method Applied to Stochastic Routing Problems: A Computational Study
Computational Optimization and Applications, 24, 289 333, 2003 c 2003 Kluwer Academic Publishers. Manufactured in The Netherlands. The Sample Average Approximation Method Applied to Stochastic Routing
More informationPart 4. Decomposition Algorithms
In the name of God Part 4. 4.4. Column Generation for the Constrained Shortest Path Problem Spring 2010 Instructor: Dr. Masoud Yaghini Constrained Shortest Path Problem Constrained Shortest Path Problem
More informationFenchel Decomposition for Stochastic Mixed-Integer Programming
Fenchel Decomposition for Stochastic Mixed-Integer Programming Lewis Ntaimo Department of Industrial and Systems Engineering, Texas A&M University, 3131 TAMU, College Station, TX 77843, USA, ntaimo@tamu.edu
More informationSMPS and the Recourse Function
IE 495 Lecture 8 SMPS and the Recourse Function Prof. Jeff Linderoth February 5, 2003 February 5, 2003 Stochastic Programming Lecture 8 Slide 1 Outline Formulating Stochastic Program(s). SMPS format Jacob
More informationPart 1. The Review of Linear Programming
In the name of God Part 1. The Review of Linear Programming 1.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline Introduction Formulation of the Dual Problem Primal-Dual Relationship Economic Interpretation
More informationLecture 8: Column Generation
Lecture 8: Column Generation (3 units) Outline Cutting stock problem Classical IP formulation Set covering formulation Column generation A dual perspective Vehicle routing problem 1 / 33 Cutting stock
More informationThe Strong Duality Theorem 1
1/39 The Strong Duality Theorem 1 Adrian Vetta 1 This presentation is based upon the book Linear Programming by Vasek Chvatal 2/39 Part I Weak Duality 3/39 Primal and Dual Recall we have a primal linear
More informationPrimal/Dual Decomposition Methods
Primal/Dual Decomposition Methods Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2018-19, HKUST, Hong Kong Outline of Lecture Subgradients
More informationMath 164-1: Optimization Instructor: Alpár R. Mészáros
Math 164-1: Optimization Instructor: Alpár R. Mészáros Final Exam, June 9, 2016 Name (use a pen): Student ID (use a pen): Signature (use a pen): Rules: Duration of the exam: 180 minutes. By writing your
More informationBenders Decomposition
Benders Decomposition Yuping Huang, Dr. Qipeng Phil Zheng Department of Industrial and Management Systems Engineering West Virginia University IENG 593G Nonlinear Programg, Spring 2012 Yuping Huang (IMSE@WVU)
More informationLinear and Integer Programming - ideas
Linear and Integer Programming - ideas Paweł Zieliński Institute of Mathematics and Computer Science, Wrocław University of Technology, Poland http://www.im.pwr.wroc.pl/ pziel/ Toulouse, France 2012 Literature
More informationDisjunctive Decomposition for Two-Stage Stochastic Mixed-Binary Programs with Random Recourse
Disjunctive Decomposition for Two-Stage Stochastic Mixed-Binary Programs with Random Recourse Lewis Ntaimo Department of Industrial and Systems Engineering, Texas A&M University, 3131 TAMU, College Station,
More informationConvex Optimization. Ofer Meshi. Lecture 6: Lower Bounds Constrained Optimization
Convex Optimization Ofer Meshi Lecture 6: Lower Bounds Constrained Optimization Lower Bounds Some upper bounds: #iter μ 2 M #iter 2 M #iter L L μ 2 Oracle/ops GD κ log 1/ε M x # ε L # x # L # ε # με f
More informationDecomposition methods in optimization
Decomposition methods in optimization I Approach I: I Partition problem constraints into two groups: explicit and implicit I Approach II: I Partition decision variables into two groups: primary and secondary
More informationSection Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018
Section Notes 9 Midterm 2 Review Applied Math / Engineering Sciences 121 Week of December 3, 2018 The following list of topics is an overview of the material that was covered in the lectures and sections
More informationLecture 8: Column Generation
Lecture 8: Column Generation (3 units) Outline Cutting stock problem Classical IP formulation Set covering formulation Column generation A dual perspective 1 / 24 Cutting stock problem 2 / 24 Problem description
More informationMulti-Area Stochastic Unit Commitment for High Wind Penetration in a Transmission Constrained Network
Multi-Area Stochastic Unit Commitment for High Wind Penetration in a Transmission Constrained Network Anthony Papavasiliou Center for Operations Research and Econometrics Université catholique de Louvain,
More information3.10 Lagrangian relaxation
3.10 Lagrangian relaxation Consider a generic ILP problem min {c t x : Ax b, Dx d, x Z n } with integer coefficients. Suppose Dx d are the complicating constraints. Often the linear relaxation and the
More informationLagrangian Relaxation in MIP
Lagrangian Relaxation in MIP Bernard Gendron May 28, 2016 Master Class on Decomposition, CPAIOR2016, Banff, Canada CIRRELT and Département d informatique et de recherche opérationnelle, Université de Montréal,
More informationApplications of Linear Programming
Applications of Linear Programming lecturer: András London University of Szeged Institute of Informatics Department of Computational Optimization Lecture 9 Non-linear programming In case of LP, the goal
More informationUses of duality. Geoff Gordon & Ryan Tibshirani Optimization /
Uses of duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Remember conjugate functions Given f : R n R, the function is called its conjugate f (y) = max x R n yt x f(x) Conjugates appear
More informationComputations with Disjunctive Cuts for Two-Stage Stochastic Mixed 0-1 Integer Programs
Computations with Disjunctive Cuts for Two-Stage Stochastic Mixed 0-1 Integer Programs Lewis Ntaimo and Matthew W. Tanner Department of Industrial and Systems Engineering, Texas A&M University, 3131 TAMU,
More informationIn the original knapsack problem, the value of the contents of the knapsack is maximized subject to a single capacity constraint, for example weight.
In the original knapsack problem, the value of the contents of the knapsack is maximized subject to a single capacity constraint, for example weight. In the multi-dimensional knapsack problem, additional
More informationScenario decomposition of risk-averse two stage stochastic programming problems
R u t c o r Research R e p o r t Scenario decomposition of risk-averse two stage stochastic programming problems Ricardo A Collado a Dávid Papp b Andrzej Ruszczyński c RRR 2-2012, January 2012 RUTCOR Rutgers
More informationNetwork Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini
In the name of God Network Flows 6. Lagrangian Relaxation 6.3 Lagrangian Relaxation and Integer Programming Fall 2010 Instructor: Dr. Masoud Yaghini Integer Programming Outline Branch-and-Bound Technique
More informationLinear Programming Methods
Chapter 11 Linear Programming Methods 1 In this chapter we consider the linear programming approach to dynamic programming. First, Bellman s equation can be reformulated as a linear program whose solution
More informationBilevel Integer Optimization: Theory and Algorithms
: Theory and Algorithms Ted Ralphs 1 Joint work with Sahar Tahernajad 1, Scott DeNegre 3, Menal Güzelsoy 2, Anahita Hassanzadeh 4 1 COR@L Lab, Department of Industrial and Systems Engineering, Lehigh University
More informationSlice Sampling Approaches for Stochastic Optimization
Slice Sampling Approaches for Stochastic Optimization John Birge, Nick Polson Chicago Booth INFORMS Philadelphia, October 2015 John Birge, Nick Polson (Chicago Booth) Split Sampling 05/01 1 / 28 Overview
More informationAn Adaptive Partition-based Approach for Solving Two-stage Stochastic Programs with Fixed Recourse
An Adaptive Partition-based Approach for Solving Two-stage Stochastic Programs with Fixed Recourse Yongjia Song, James Luedtke Virginia Commonwealth University, Richmond, VA, ysong3@vcu.edu University
More informationand to estimate the quality of feasible solutions I A new way to derive dual bounds:
Lagrangian Relaxations and Duality I Recall: I Relaxations provide dual bounds for the problem I So do feasible solutions of dual problems I Having tight dual bounds is important in algorithms (B&B), and
More informationThe DC Optimal Power Flow
1 / 20 The DC Optimal Power Flow Quantitative Energy Economics Anthony Papavasiliou The DC Optimal Power Flow 2 / 20 1 The OPF Using PTDFs 2 The OPF Using Reactance 3 / 20 Transmission Constraints Lines
More informationCapacity Planning with uncertainty in Industrial Gas Markets
Capacity Planning with uncertainty in Industrial Gas Markets A. Kandiraju, P. Garcia Herreros, E. Arslan, P. Misra, S. Mehta & I.E. Grossmann EWO meeting September, 2015 1 Motivation Industrial gas markets
More informationMonte Carlo Methods for Stochastic Programming
IE 495 Lecture 16 Monte Carlo Methods for Stochastic Programming Prof. Jeff Linderoth March 17, 2003 March 17, 2003 Stochastic Programming Lecture 16 Slide 1 Outline Review Jensen's Inequality Edmundson-Madansky
More informationTwo-stage integer programs with stochastic right-hand sides: a superadditive dual approach
Math. Program., Ser. B 108, 275 296 (2006) Digital Object Identifier (DOI) 10.1007/s10107-006-0711-y Nan Kong Andrew J. Schaefer Brady Hunsaker Two-stage integer programs with stochastic right-hand sides:
More informationSection Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010
Section Notes 9 IP: Cutting Planes Applied Math 121 Week of April 12, 2010 Goals for the week understand what a strong formulations is. be familiar with the cutting planes algorithm and the types of cuts
More informationChapter 3, Operations Research (OR)
Chapter 3, Operations Research (OR) Kent Andersen February 7, 2007 1 Linear Programs (continued) In the last chapter, we introduced the general form of a linear program, which we denote (P) Minimize Z
More informationCONVERGENCE ANALYSIS OF SAMPLING-BASED DECOMPOSITION METHODS FOR RISK-AVERSE MULTISTAGE STOCHASTIC CONVEX PROGRAMS
CONVERGENCE ANALYSIS OF SAMPLING-BASED DECOMPOSITION METHODS FOR RISK-AVERSE MULTISTAGE STOCHASTIC CONVEX PROGRAMS VINCENT GUIGUES Abstract. We consider a class of sampling-based decomposition methods
More informationSupplementary lecture notes on linear programming. We will present an algorithm to solve linear programs of the form. maximize.
Cornell University, Fall 2016 Supplementary lecture notes on linear programming CS 6820: Algorithms 26 Sep 28 Sep 1 The Simplex Method We will present an algorithm to solve linear programs of the form
More informationMachine Learning. Support Vector Machines. Fabio Vandin November 20, 2017
Machine Learning Support Vector Machines Fabio Vandin November 20, 2017 1 Classification and Margin Consider a classification problem with two classes: instance set X = R d label set Y = { 1, 1}. Training
More informationPenalty and Barrier Methods General classical constrained minimization problem minimize f(x) subject to g(x) 0 h(x) =0 Penalty methods are motivated by the desire to use unconstrained optimization techniques
More informationOutline. Relaxation. Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING. 1. Lagrangian Relaxation. Lecture 12 Single Machine Models, Column Generation
Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING 1. Lagrangian Relaxation Lecture 12 Single Machine Models, Column Generation 2. Dantzig-Wolfe Decomposition Dantzig-Wolfe Decomposition Delayed Column
More informationThe dual simplex method with bounds
The dual simplex method with bounds Linear programming basis. Let a linear programming problem be given by min s.t. c T x Ax = b x R n, (P) where we assume A R m n to be full row rank (we will see in the
More information4. Duality Duality 4.1 Duality of LPs and the duality theorem. min c T x x R n, c R n. s.t. ai Tx = b i i M a i R n
2 4. Duality of LPs and the duality theorem... 22 4.2 Complementary slackness... 23 4.3 The shortest path problem and its dual... 24 4.4 Farkas' Lemma... 25 4.5 Dual information in the tableau... 26 4.6
More informationYou should be able to...
Lecture Outline Gradient Projection Algorithm Constant Step Length, Varying Step Length, Diminishing Step Length Complexity Issues Gradient Projection With Exploration Projection Solving QPs: active set
More informationΩ R n is called the constraint set or feasible set. x 1
1 Chapter 5 Linear Programming (LP) General constrained optimization problem: minimize subject to f(x) x Ω Ω R n is called the constraint set or feasible set. any point x Ω is called a feasible point We
More informationLagrangian Duality and Convex Optimization
Lagrangian Duality and Convex Optimization David Rosenberg New York University February 11, 2015 David Rosenberg (New York University) DS-GA 1003 February 11, 2015 1 / 24 Introduction Why Convex Optimization?
More informationEE364a Review Session 5
EE364a Review Session 5 EE364a Review announcements: homeworks 1 and 2 graded homework 4 solutions (check solution to additional problem 1) scpd phone-in office hours: tuesdays 6-7pm (650-723-1156) 1 Complementary
More informationStochastic Dual Dynamic Integer Programming
Stochastic Dual Dynamic Integer Programming Jikai Zou Shabbir Ahmed Xu Andy Sun December 26, 2017 Abstract Multistage stochastic integer programming (MSIP) combines the difficulty of uncertainty, dynamics,
More informationRelation of Pure Minimum Cost Flow Model to Linear Programming
Appendix A Page 1 Relation of Pure Minimum Cost Flow Model to Linear Programming The Network Model The network pure minimum cost flow model has m nodes. The external flows given by the vector b with m
More informationAdditional Exercises for Introduction to Nonlinear Optimization Amir Beck March 16, 2017
Additional Exercises for Introduction to Nonlinear Optimization Amir Beck March 16, 2017 Chapter 1 - Mathematical Preliminaries 1.1 Let S R n. (a) Suppose that T is an open set satisfying T S. Prove that
More informationDisjunctive Decomposition for Two-Stage Stochastic Mixed-Binary Programs with GUB Constraints
Disjunctive Decomposition for Two-Stage Stochastic Mixed-Binary Programs with GUB Constraints Brian Keller Booz Allen Hamilton, 134 National Business Parkway, Annapolis Junction, MD 20701, USA, keller
More informationImprovements to Benders' decomposition: systematic classification and performance comparison in a Transmission Expansion Planning problem
Improvements to Benders' decomposition: systematic classification and performance comparison in a Transmission Expansion Planning problem Sara Lumbreras & Andrés Ramos July 2013 Agenda Motivation improvement
More informationLinear Programming: Chapter 5 Duality
Linear Programming: Chapter 5 Duality Robert J. Vanderbei September 30, 2010 Slides last edited on October 5, 2010 Operations Research and Financial Engineering Princeton University Princeton, NJ 08544
More informationMulti-Area Stochastic Unit Commitment for High Wind Penetration
Multi-Area Stochastic Unit Commitment for High Wind Penetration Workshop on Optimization in an Uncertain Environment Anthony Papavasiliou, UC Berkeley Shmuel S. Oren, UC Berkeley March 25th, 2011 Outline
More information14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.
CS/ECE/ISyE 524 Introduction to Optimization Spring 2016 17 14. Duality ˆ Upper and lower bounds ˆ General duality ˆ Constraint qualifications ˆ Counterexample ˆ Complementary slackness ˆ Examples ˆ Sensitivity
More informationA Tighter Variant of Jensen s Lower Bound for Stochastic Programs and Separable Approximations to Recourse Functions
A Tighter Variant of Jensen s Lower Bound for Stochastic Programs and Separable Approximations to Recourse Functions Huseyin Topaloglu School of Operations Research and Information Engineering, Cornell
More informationTRINITY COLLEGE DUBLIN THE UNIVERSITY OF DUBLIN. School of Mathematics
JS and SS Mathematics JS and SS TSM Mathematics TRINITY COLLEGE DUBLIN THE UNIVERSITY OF DUBLIN School of Mathematics MA3484 Methods of Mathematical Economics Trinity Term 2015 Saturday GOLDHALL 09.30
More informationConvex Optimization and Support Vector Machine
Convex Optimization and Support Vector Machine Problem 0. Consider a two-class classification problem. The training data is L n = {(x 1, t 1 ),..., (x n, t n )}, where each t i { 1, 1} and x i R p. We
More informationCopositive Programming and Combinatorial Optimization
Copositive Programming and Combinatorial Optimization Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with I.M. Bomze (Wien) and F. Jarre (Düsseldorf) IMA
More informationOn the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems
MATHEMATICS OF OPERATIONS RESEARCH Vol. xx, No. x, Xxxxxxx 00x, pp. xxx xxx ISSN 0364-765X EISSN 156-5471 0x xx0x 0xxx informs DOI 10.187/moor.xxxx.xxxx c 00x INFORMS On the Power of Robust Solutions in
More informationOblique derivative problems for elliptic and parabolic equations, Lecture II
of the for elliptic and parabolic equations, Lecture II Iowa State University July 22, 2011 of the 1 2 of the of the As a preliminary step in our further we now look at a special situation for elliptic.
More informationA New Robust Scenario Approach to Supply. Chain Optimization under Bounded Uncertainty
A New Robust Scenario Approach to Supply Chain Optimization under Bounded Uncertainty by Niaz Bahar Chowdhury A thesis submitted to the Department of Chemical Engineering in conformity with the requirements
More informationThe Dual Simplex Algorithm
p. 1 The Dual Simplex Algorithm Primal optimal (dual feasible) and primal feasible (dual optimal) bases The dual simplex tableau, dual optimality and the dual pivot rules Classical applications of linear
More informationLinear Programming Redux
Linear Programming Redux Jim Bremer May 12, 2008 The purpose of these notes is to review the basics of linear programming and the simplex method in a clear, concise, and comprehensive way. The book contains
More informationDecomposition Algorithms for Two-Stage Chance-Constrained Programs
Mathematical Programming manuscript No. (will be inserted by the editor) Decomposition Algorithms for Two-Stage Chance-Constrained Programs Xiao Liu Simge Küçükyavuz Luedtke James Received: date / Accepted:
More informationDistributed power allocation for D2D communications underlaying/overlaying OFDMA cellular networks
Distributed power allocation for D2D communications underlaying/overlaying OFDMA cellular networks Marco Moretti, Andrea Abrardo Dipartimento di Ingegneria dell Informazione, University of Pisa, Italy
More informationCombinatorial Data Mining Method for Multi-Portfolio Stochastic Asset Allocation
Combinatorial for Stochastic Asset Allocation Ran Ji, M.A. Lejeune Department of Decision Sciences July 8, 2013 Content Class of Models with Downside Risk Measure Class of Models with of multiple portfolios
More information3.7 Cutting plane methods
3.7 Cutting plane methods Generic ILP problem min{ c t x : x X = {x Z n + : Ax b} } with m n matrix A and n 1 vector b of rationals. According to Meyer s theorem: There exists an ideal formulation: conv(x
More informationA Progressive Hedging Approach to Multistage Stochastic Generation and Transmission Investment Planning
A Progressive Hedging Approach to Multistage Stochastic Generation and Transmission Investment Planning Yixian Liu Ramteen Sioshansi Integrated Systems Engineering Department The Ohio State University
More informationOptimal Placement & sizing of Distributed Generator (DG)
Chapter - 5 Optimal Placement & sizing of Distributed Generator (DG) - A Single Objective Approach CHAPTER - 5 Distributed Generation (DG) for Power Loss Minimization 5. Introduction Distributed generators
More informationGestion de la production. Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA
Gestion de la production Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA 1 Contents 1 Integer Linear Programming 3 1.1 Definitions and notations......................................
More informationDuality. Geoff Gordon & Ryan Tibshirani Optimization /
Duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Duality in linear programs Suppose we want to find lower bound on the optimal value in our convex problem, B min x C f(x) E.g., consider
More informationDecomposition with Branch-and-Cut Approaches for Two Stage Stochastic Mixed-Integer Programming
Decomposition with Branch-and-Cut Approaches for Two Stage Stochastic Mixed-Integer Programming by Suvrajeet Sen SIE Department, University of Arizona, Tucson, AZ 85721 and Hanif D. Sherali ISE Department,
More information"SYMMETRIC" PRIMAL-DUAL PAIR
"SYMMETRIC" PRIMAL-DUAL PAIR PRIMAL Minimize cx DUAL Maximize y T b st Ax b st A T y c T x y Here c 1 n, x n 1, b m 1, A m n, y m 1, WITH THE PRIMAL IN STANDARD FORM... Minimize cx Maximize y T b st Ax
More informationLearning Algorithms for Separable Approximations of Discrete Stochastic Optimization Problems
Learning Algorithms for Separable Approximations of Discrete Stochastic Optimization Problems Warren B. Powell Department of Operations Research and Financial Engineering, Princeton University, Princeton,
More informationApplications. Stephen J. Stoyan, Maged M. Dessouky*, and Xiaoqing Wang
Introduction to Large-Scale Linear Programming and Applications Stephen J. Stoyan, Maged M. Dessouky*, and Xiaoqing Wang Daniel J. Epstein Department of Industrial and Systems Engineering, University of
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57
More informationLecture 5. x 1,x 2,x 3 0 (1)
Computational Intractability Revised 2011/6/6 Lecture 5 Professor: David Avis Scribe:Ma Jiangbo, Atsuki Nagao 1 Duality The purpose of this lecture is to introduce duality, which is an important concept
More information