Solution Methods for Stochastic Programs

Similar documents
Stochastic Integer Programming

The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 44

Stochastic Decomposition

A Tighter Variant of Jensen s Lower Bound for Stochastic Programs and Separable Approximations to Recourse Functions

Introduction to Nonlinear Stochastic Programming

Lagrange Relaxation: Introduction and Applications

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Duality. Geoff Gordon & Ryan Tibshirani Optimization /

The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 38

Stochastic Programming: From statistical data to optimal decisions

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Benders Decomposition

Decomposition with Branch-and-Cut Approaches for Two Stage Stochastic Mixed-Integer Programming

Duality in Linear Programs. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Lecture 3. Optimization Problems and Iterative Algorithms

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Stochastic Integer Programming An Algorithmic Perspective

Lecture 2: Convex Sets and Functions

CONVERGENCE ANALYSIS OF SAMPLING-BASED DECOMPOSITION METHODS FOR RISK-AVERSE MULTISTAGE STOCHASTIC CONVEX PROGRAMS

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Constrained Optimization

Primal Solutions and Rate Analysis for Subgradient Methods

On a class of minimax stochastic programs

Learning Algorithms for Separable Approximations of Discrete Stochastic Optimization Problems

Outline. Relaxation. Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING. 1. Lagrangian Relaxation. Lecture 12 Single Machine Models, Column Generation

Solving Dual Problems

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

Dual methods and ADMM. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725

Lagrangian Relaxation in MIP

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Financial Optimization ISE 347/447. Lecture 21. Dr. Ted Ralphs

Fenchel Decomposition for Stochastic Mixed-Integer Programming

c 2004 Society for Industrial and Applied Mathematics

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

IE 495 Final Exam. Due Date: May 1, 2003

Network Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini

10 Numerical methods for constrained problems

Constrained Optimization and Lagrangian Duality

INEXACT CUTS IN BENDERS' DECOMPOSITION GOLBON ZAKERI, ANDREW B. PHILPOTT AND DAVID M. RYAN y Abstract. Benders' decomposition is a well-known techniqu

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

You should be able to...

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Math 273a: Optimization Subgradients of convex functions

Primal/Dual Decomposition Methods

Complexity of two and multi-stage stochastic programming problems

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

Diana Barro and Elio Canestrelli. Combining stochastic programming and optimal control to solve multistage stochastic optimization problems

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

A New Dynamic Programming Decomposition Method for the Network Revenue Management Problem with Customer Choice Behavior

A SIMPLE PARALLEL ALGORITHM WITH AN O(1/T ) CONVERGENCE RATE FOR GENERAL CONVEX PROGRAMS

Approximate Primal Solutions and Rate Analysis for Dual Subgradient Methods

Algorithms for nonlinear programming problems II

Support Vector Machines

Lagrangean Decomposition for Mean-Variance Combinatorial Optimization

Sum-Power Iterative Watefilling Algorithm

Applications of Linear Programming

4. Algebra and Duality

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

SMPS and the Recourse Function

On Robust Optimization of Two-Stage Systems

Convex approximations for a class of mixed-integer recourse models

Math 273a: Optimization Subgradients of convex functions

Reformulation of chance constrained problems using penalty functions

Inverse Stochastic Dominance Constraints Duality and Methods

3.10 Lagrangian relaxation

Decomposition Algorithms for Stochastic Programming on a Computational Grid

Upper bound for optimal value of risk averse multistage problems

Citation for published version (APA): van der Vlerk, M. H. (2002). Convex approximations for complete integer recourse models. s.n.

An Adaptive Partition-based Approach for Solving Two-stage Stochastic Programs with Fixed Recourse

Lecture 1. Stochastic Optimization: Introduction. January 8, 2018

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Lecture 8. Strong Duality Results. September 22, 2008

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Lagrangian Duality Theory

Stochastic Programming in Enterprise-Wide Optimization

SOLVING A MINIMIZATION PROBLEM FOR A CLASS OF CONSTRAINED MAXIMUM EIGENVALUE FUNCTION

Lecture 10: Duality in Linear Programs

Sample Average Approximation (SAA) for Stochastic Programs

arxiv: v2 [math.oc] 18 Nov 2017

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

Scenario decomposition of risk-averse two stage stochastic programming problems

Computational Stochastic Programming

Lecture 7: Lagrangian Relaxation and Duality Theory

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Bounding in Multi-Stage. Stochastic Programming. Problems. Olga Fiedler a Andras Prekopa b

Stochastic Programming Math Review and MultiPeriod Models

A Duality-Based Relaxation and Decomposition Approach for Inventory Distribution Systems

Benders Decomposition Methods for Structured Optimization, including Stochastic Optimization

Subgradient Methods in Network Resource Allocation: Rate Analysis

Conditional Gradient (Frank-Wolfe) Method

A. Shapiro Introduction

Monte Carlo Methods for Stochastic Programming

Lecture Note 5: Semidefinite Programming for Stability Analysis

Part 1. The Review of Linear Programming

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

A Cutting Surface Method for Uncertain Linear Programs with Polyhedral Stochastic Dominance Constraints

Transcription:

Solution Methods for Stochastic Programs Huseyin Topaloglu School of Operations Research and Information Engineering Cornell University ht88@cornell.edu August 14, 2010 1

Outline Cutting plane methods for convex optimization Decomposition methods for two-stage stochastic programs Dual methods for two-stage stochastic programs 2

Cutting Plane Methods for Convex Optimization 3

A Simple Method for Convex Optimization Consider the optimization problem z = min x X f(x), where f is convex and X is a compact and convex set Necessary background is a first course on optimization that includes linear programming duality Recall the subgradient inequality, which constructs a cutting plane approximation to f at point x 0 by f(x 0 ) + f(x 0 ) [x x 0 ] f(x) 4

f (x) f (x) x0 x x 0 x f ( x0) + f ( x0)[ x x0] f ( x0) + g( x0)[ x x0] 5

A Simple Method for Convex Optimization Consider the optimization problem z = min x [ 1,2] (x)2 Start with an initial guess for the optimal solution x 1 = 1 Use the subgradient inequality to construct a cutting plane approximation to the objective function at x 1 (x 1 ) 2 + 2 x 1 (x x 1 ) = 1 2 (x + 1) = 2 x 1 6

4 3 2 1 0-2 -1 0 1 2 3-1 -2-3 -4-5 4 3 2 1 0-2 -1 0 1 2 3-1 -2-3 -4-5 7

A Simple Method for Convex Optimization Minimize the lower bound approximation over the feasible set by solving the problem min v subject to x [ 1, 2] v 2 x 1 Obtain a guess for the optimal solution x 2 = 2 and a guess for the optimal objective value v 2 = 5 Use the subgradient inequality to construct a cutting plane approximation to the objective function at x 2 (x 2 ) 2 + 2 x 2 (x x 2 ) = 4 + 4 (x 2) = 4 x 4 8

4 3 2 1 0-2 -1 0 1 2 3-1 -2-3 -4-5 4 3 2 1 0-2 -1 0 1 2 3-1 -2-3 -4-5 9

A Simple Method for Convex Optimization Minimize the lower bound approximation over the feasible set by solving the problem min subject to x [ 1, 2] v v 2 x 1, v 4 x 4 Obtain a guess for the optimal solution x 3 = 1 2 optimal objective value v 3 = 2 and a guess for the Use the subgradient inequality to construct a cutting plane approximation to the objective function at x 3 (x 3 ) 2 + 2 x 3 (x x 3 ) = 1 4 + (x 1 2 ) = x 1 4 10

4 3 2 1 0-2 -1 0 1 2 3-1 -2-3 -4-5 4 3 2 1 0-2 -1 0 1 2 3-1 -2-3 -4-5 11

A Simple Method for Convex Optimization Minimize the lower bound approximation over the feasible set by solving the problem min v subject to x [ 1, 2] v 2 x 1, v 4 x 4, v x 1/4 Obtain a guess for the optimal solution x 4 = 1 4 optimal objective value v 4 = 1 2... and a guess for the 12

Important Observations At each iteration, we minimize a lower bound approximation to the objective function f Using f n to denote the lower bound approximation to the objective function f at iteration n, f n (x) f n+1 (x) f(x) for all x X 13

Important Observations At iteration n, x n minimizes the lower bound approximation f n and Note that v n = min x X f n (x) = f n (x n ) v n = min x X f n (x) min x X f(x) = z f(x n ) If v n = f(x n ), then f(x n ) = z and x n must be optimal 14

Cutting Plane Method for Convex Optimization 1. Start with an initial guess for the optimal solution x 1 X and set n = 1 2. Construct a cutting plane approximation to f at x n f(x n ) + f(x n )[x x n ] f(x) 3. Minimize the lower bound approximation by solving min subject to v x X v f(x i ) + f(x i ) [x x i ] i = 1,..., n 4. Letting (v n+1, x n+1 ) be an optimal solution, if v n+1 = f(x n+1 ), then stop, else increase n by 1 and go to Step 2 15

Convergence of the Cutting Plane Method Theorem Letting the sequence of points {x n } n be generated by the cutting plane method, lim n f(x n ) = z f(x 4 ) f(x 1 ) f(x 3 )f(x 7 ) f(x 2 ) f(x 6 ) f(x 5 ) z* Take the first two points that provide an objective function value that lies to the right of z + These points are x 2 and x 5 16

Convergence of the Cutting Plane Method Cutting plane approximation that we generate at iteration 2 is f(x 2 ) + f(x 2 ) [x x 2 ] and the constraint v f(x 2 ) + f(x 2 ) [x x 2 ] remains in our lower bound approximation after iteration 2 At iteration 5, since this constraint is still in our lower bound approximation, it must to be satisfied by the solution (v 5, x 5 ) obtained at iteration 5 and we have v 5 f(x 2 ) + f(x 2 ) [x 5 x 5 ] Since f(x 5 ) z +, we have f(x 5 ) z v 5 17

Convergence of the Cutting Plane Method Thus, f(x 5 ) f(x 2 ) f(x 2 ) [x 5 x 2 ] Letting C = max x X f(x), f(x) f(y) C x y We obtain f(x 5 ) f(x 2 ) f(x 2 ) [x 5 x 2 ] C x 5 x 2 + C x 5 x 2 so that /2 C x 5 x 2 18

Decomposition Methods for Two-Stage Stochastic Programs 19

Decomposition Methods for Two-Stage Stochastic Programs Assume that there are K < scenarios and the only random component is the right side of the constraints in the second stage Use p k to denote the probability of scenario k and h k to denote the right side of the constraints in the second stage under scenario k min subject to c t x + Q(x) Ax = b x 0, where the recourse function Q(x) = K k=1 p k Q k (x) is defined as Q k (x) = min subject to q t y W y = h k T x y 0 20

Decomposition Methods for Two-Stage Stochastic Programs We can write the two-stage stochastic problem in its deterministic equivalent form K min c t x + p k q t y k k=1 subject to Ax = b T x + W y k = h k x 0, y k 0 k = 1,..., K k = 1,..., K This problem has both large number of constraints and large number of decision variables 21

Convexity of the Recourse Function If the recourse function Q is convex and it is tractable to compute its subgradients, then we can use the cutting plane method to solve the two-stage stochastic program min subject to c t x + Q(x) Ax = b x 0 22

Convexity of the Recourse Function The second stage problem is Q k (x) = min subject to q t y W y = h k T x y 0 Let π k (x) be an optimal solution to the dual of the second stage problem Q k (x) = max subject to [h k T x] t π W t π q By optimality of π k (x 0 ), Q k (x 0 ) = [h k T x 0 ] t π k (x 0 ) and by feasibility, Q k (x) [h k T x] t π k (x 0 ) 23

Convexity of the Recourse Function The second stage problem is Q k (x) = min subject to q t y W y = h k T x y 0 Let π k (x) be an optimal solution to the dual of the second stage problem Q k (x 0 ) = max subject to [h k T x 0 ] t π W t π q By optimality of π k (x 0 ), Q k (x 0 ) = [h k T x 0 ] t π k (x 0 ) and by feasibility, Q k (x) [h k T x] t π k (x 0 ) 24

Convexity of the Recourse Function We obtain the subgradient inequality for Q k Taking expectations, Q k (x) Q k (x 0 ) π t k(x 0 ) T [x x 0 ] Q(x) Q(x 0 ) K p k πk(x t 0 ) T [x x 0 ] k=1 Thus, a subgradient of Q at point x 0 is given by K g(x 0 ) = p k πk(x t 0 ) T k=1 25

Decomposition Method for Two-Stage Stochastic Programs 1. Start with an initial guess for the optimal solution x 1 and set n = 1 2. Construct a cutting plane approximation to Q at x n Q(x n ) + g(x n ) [x x n ] 3. Minimize the lower bound approximation by solving min c t x + v subject to A x = b, x 0 v Q(x i ) + g(x i ) [x x i ] i = 1,..., n 4. Letting (v n+1, x n+1 ) be an optimal solution, if v n+1 = Q(x n+1 ), then stop, else increase n by 1 and go to Step 2 26

Dealing with Infeasibility Assume that we are at iteration n with the solution x n and there exists a scenario k under which the second stage problem is infeasible Q k (x n ) = min subject to q t y W y = h k T x n y 0 We can detect this infeasibility by solving U k (x n ) = min subject to e t z + + e t z W y + z + z = h k T x n y 0, z + 0, z 0 and observing that U k (x n ) > 0 27

Dealing with Infeasibility Dual of the infeasibility detection problem is U k (x n ) = max [h k T x n ] t η subject to W t η 0 1 η 1 Let η k (x n ) be the solution to the dual of the infeasibility detection problem and consider the constraint [h k T x] t η k (x n ) 0 This constraint is not satisfied by the problematic point x n since [h k T x n ] t η k (x n ) = U k (x n ) > 0 This constraint is satisfied by any nonproblematic point x since [h k T x] t η k (x n ) U k (x) = 0 28

Dealing with Infeasibility Dual of the infeasibility detection problem is U k (x) = max [h k T x] t η subject to W t η 0 1 η 1 Let η k (x n ) be the solution to the dual of the infeasibility detection problem and consider the constraint [h k T x] t η k (x n ) 0 This constraint is not satisfied by the problematic point x n since [h k T x n ] t η k (x n ) = U k (x n ) > 0 This constraint is satisfied by any nonproblematic point x since [h k T x] t η k (x n ) U k (x) = 0 29

Decomposition Method for Two-Stage Stochastic Programs 1. Start with an initial guess for the optimal solution x 1, set n = 1, N O = and Nk F = for all k = 1,..., K 2. With the solution x n, solve the second stage problem Q k (x n ) = min subject to q t y W y = h k T x n y 0 for all k = 1,..., K 3. If there is a scenario k under which the second stage problem is infeasible, then construct the constraint and set N F k N F k {n} [h k T x] t η k (x n ) 0 30

4. If the second stage problem is feasible for all scenarios, then construct a cutting plane approximation to Q at x n and set N O N O {n} Q(x n ) + g(x n ) [x x n ] 5. Minimize the lower bound approximation by solving min c t x + v subject to A x = b, x 0 v Q(x i ) + g(x i ) [x x i ] i N O 0 [h k T x] t η k (x i ) i N F k, k = 1,..., K 6. Letting (v n+1, x n+1 ) be an optimal solution, if v n+1 = Q(x n+1 ), then stop, else increase n by 1 and go to Step 2 31

Dual Methods for Two-Stage Stochastic Programs 32

Dual Methods for Two-Stage Stochastic Programs Consider the deterministic equivalent form K z = min c t x + p k q t y k k=1 subject to Ax = b T x + W y k = h k x 0, y k 0 k = 1,..., K k = 1,..., K Write the deterministic equivalent form as K z = min c t x 0 + p k q t y k k=1 subject to Ax k = b k = 1,..., K T x k + W y k = h k k = 1,..., K x k x 0 = 0 k = 1,..., K (λ k ) x k 0, y k 0 k = 1,..., K 33

Dual Methods for Two-Stage Stochastic Programs Relax the constraints that link the scenarios by associating the Lagrange multipliers λ = (λ 1,..., λ K ) with them L(λ) = min [ c t K k=1 λ t k ] x 0 + K λ t k x k + k=1 subject to Ax k = b k = 1,... K T x k + W y k = h k x k 0, y k 0 K p k q t y k k=1 k = 1,..., K k = 1,..., K Relaxed problem decomposes by the scenarios and it can be solved in a tractable fashion by solving one subproblem for each scenario 34

Dual Methods for Two-Stage Stochastic Programs For any λ, L(λ) z To obtain the tightest possible lower bound on z, solve max λ L(λ) z The tightest possible lower bound satisfies max λ L(λ) = z When viewed as a function of the Lagrange multipliers, the optimal nonobjective of the relaxed problem is concave in λ We can use the cutting plane method to solve the problem max λ L(λ) 35

Limitations and Extensions Obtaining a subgradient of the recourse function requires solving the second stage problem for all scenarios Stochastic decomposition and cutting plane and partial sampling methods allow solving the second stage problem for only a subset of the scenarios Cutting plane methods do not take advantage of a good initial solution Regularized decomposition and trust region methods try to make use of a good initial solution by limiting how much we move towards a promising point Cutting plane methods get within the vicinity of the optimal solution quite fast, but can take a lot of iterations to get to the optimal solution 36

Further Reading Introduction to Stochastic Programming, J. R. Birge, F. Louveaux Numerical Techniques for Stochastic Optimization, edited by Yu. Ermoliev, R. J-B Wets Stochastic Decomposition: A Statistical Method for Large Scale Stochastic Linear Programming, J. L. Higle, S. Sen Stochastic Linear Programming: Models, Theory, and Computation, P. Kall, J. Mayer Stochastic Programming, P. Kall and S. W. Wallace Handbooks in Operations Research and Management Science: Stochastic Programming, edited by A. Ruszczynski, A. Shapiro Lectures on Stochastic Programming: Modeling and Theory, A. Shapiro, D. Dentcheva, A. Ruszczynski Applications of Stochastic Programming, edited by S. W. Wallace, W. T. Ziemba 37

Further Reading Nonlinear Programming, D. P. Bertsekas Convex Analysis and Optimization, D. Bertsekas, A. Nedic, A. Ozdaglar Convex Analysis, R. T. Rockafellar Nonlinear Optimization, A. Ruszczynski (Please excuse any involuntary omissions) 38