MIDAS: A Mixed Integer Dynamic Approximation Scheme

Size: px
Start display at page:

Download "MIDAS: A Mixed Integer Dynamic Approximation Scheme"

Transcription

1 MIDAS: A Mixed Integer Dynamic Approximation Scheme Andy Philpott, Faisal Wahid, Frédéric Bonnans May 7, 2016 Abstract Mixed Integer Dynamic Approximation Scheme (MIDAS) is a new sampling-based algorithm for solving finite-horizon stochastic dynamic programs with monotonic Bellman functions. MIDAS approximates these value functions using step functions, leading to stage problems that are mixed integer programs. We provide a general description of MIDAS, and prove its almost-sure convergence to an ε-optimal policy when the Bellman functions are known to be continuous, and the sampling process satisfies standard assumptions. 1 Introduction In general, multistage stochastic programming problems are extremely difficult to solve. If one discretizes the random variable using a finite set of outcomes in each stage and represents the process as a scenario tree, then the problem size grows exponentially with the number of stages and outcomes [9]. On the other hand, if the random variables are stagewise independent then the problem can be formulated as a dynamic programming recursion, and attacked using an approximate dynamic programming algorithm. When the Bellman functions are known to be convex (if minimizing) or concave (if maximizing) then they can be approximated by cutting planes. This is the basis of the popular Stochastic Dual Dynamic Programming (SDDP) algorithm originally proposed by Pereira and Pinto [7]. This creates a sequence of cutting-plane outer approximations to the Bellman function at each stage by evaluating stage problems on independently sampled sequences of random outcomes. A comprehensive recent summary of this algorithm and its properties is provided by [10]. A number of algorithms that are variations of this idea have been proposed (see e.g. [3, 4, 6]). The first formal proof of almost-sure convergence for SDDP-based algorithms was provided by Chen and Powell in [3] for their CUPPS algorithm. Their proof however relied on a key unstated assumption identified by Philpott and Guan in [8], who provided a new proof of almost-sure convergence for problems University of Auckland University of Auckland Inria & École Polytechnique 1

2 with polyhedral Bellman functions without requiring this assumption. Almost-sure convergence for general convex Bellman functions was recently proved using a different approach by Girardeau et al. [5]. In order to guarantee the validity of the outer approximation of the Bellman functions, SDDP-based methods require these functions to be convex (if minimizing) or concave (if maximizing). However, there are many instances of problems when the optimal value of the stage problem is not a convex (or concave) function of the state variables. For example, if the stage problem involves state variables with components appearing in both the objective function and the right-hand side of a convex optimization problem then the optimal value function might have a saddle point. More generally, if the stage problems are not convex then we cannot guarantee convexity of the optimal value function. This will happen when stage problems are mixed-integer programs (MIPS), or when they incorporate nonlinear effects, such as modeling head effects in hydropower production functions [2]. Our interest in problems of this type is motivated by models of hydroelectricity systems in which we seek to maximize revenue from releasing water through generating stations on a river chain. This can be modeled as a system with discrete-time stochastic dynamics: x t+1 = f t (x t, u t, ξ t ), x 1 = x, t = 1, 2,..., T. In this problem, u t U(x t ) is a control and ξ t is a random noise term. We assume that f t is a continuous function of its arguments. The control u t generates a reward r t (x t, u t ) in each stage and a terminal reward Q(x T +1 ), where r t and Q are continuous and monotonic increasing in their arguments. Given initial state x, we seek an optimal policy yielding V 1 ( x), where [ ] V t (x) = E ξt max r t(x, u, ξ t ) + V t+1 (f t (x, u, ξ t ))} u U(x) V T +1 (x) = Q(x). Here V t (x) denotes the maximum expected reward from the beginning of stage t onwards, given the state is x, and we take action u t after observing the random disturbance ξ t. In the hydro-scheduling context, x = (s, p) represents both a reservoir stock variable s and a price variable p, and u = (v, l) represents the reservoir release through generators and reservoir spill. In this case the dynamics might be represented by [ ] [ ] st+1 st v = t l t + ω t, p t+1 α t p t + b t + η t where ω t is (random) reservoir inflow, and η t is the error term for an autoregressive model of price, so ξ t = [ ω t η t ]. Here we might define r t (s, p, v, l, ξ t ) = (α t p + b t + η t )g(v), giving the revenue earned by released energy g(v) sold at price α t p + b t + η t. For such a model, it is easy to show (under the assumption that we can spill energy without penalty) that V t (s, p) is continuous and monotonic increasing in s and p, but in general V t (s, p) is not concave so representing it within the SDDP framework is not trivial. 2

3 A number of authors have looked to extend SDDP methods to deal with non-convex stage problems. The first approach replaces the non-convex components of the problem with convex approximations, for example using McCormick envelopes to approximate the production function of hydro plants as in [2]. The second approach convexifies the value function, e.g. using Lagrangian relaxation techniques. A recent paper [12] by Thomé et al. proposes a Lagrangian relaxation of of the SDDP subproblem, and then uses the Lagrange multipliers to produce valid Benders cuts. A similar approach is adopted by Steeger and Rebennack in [11]. Abgottspon et al [1] introduce a heuristic to add locally valid cuts that enhance a convexified approximation of the value function. In this paper we propose a new extension of SDDP called Mixed Integer Dynamic Approximation Scheme (MIDAS). MIDAS uses the same algorithmic framework as SDDP but, instead of using cutting planes, MIDAS uses step functions to approximate the value function. The approximation requires an assumption that the value function is monotonic in the state variables. In this case each stage problem can be solved to global optimality as a MIP. We show that if the true Bellman function is continuous then MIDAS converges almost surely to a 2T ε-optimal solution for a problem with T stages. The rest of the paper is laid out as follows. We begin with some preliminary results on ε-outer approximation of monotonic functions. In section 3 we prove the convergence of MIDAS to a 2T ε-optimal solution to a multistage deterministic optimization problem. Lastly, in section 4, we extend our proof to a sampling-based algorithm applied to the multistage stochastic optimization problem, and demonstrate its almost-sure convergence. We conclude the paper in section 5 by discussing the potential value of using MIDAS to solve stochastic multistage integer programs. 2 Mixed Integer Dynamic Approximation Scheme Before describing MIDAS, we prove some preliminary lemmas relating to global optimization. Consider the problem P: max Q(x) s.t. x X where Q(x) is an arbitrary upper semicontinuous function and X is an arbitrary compact subset of R n. It follows that P has a global maximum attained at x, say. Given ε > 0, we seek an ε-optimal solution to P. In other words we will find ˆx X with Q(ˆx) Q(x ) ε. Suppose that there is some upper semicontinuous function Q so that for every x, Q(x) Q(x). (1) Lemma 1. Any solution x to P : max Q(x) s.t. x X. with Q( x) Q( x) + ε is an ε-optimal solution to P. 3

4 Proof. If x solves P then (1) implies so Q(x ) Q(x ) Q( x) Q( x) + ε Q( x) + ε Q(x ), yielding the result. MIDAS updates the value function at points x h that are the most recently visited values of the state variables computed using a forward pass of the algorithm. We can illustrate the basic step of MIDAS using a deterministic two-stage program P : max u,x r(u) + Q(x) s.t. x = f(u), u U, x X. We assume that Q(x) is monotonic increasing and continuous on X. Since X is compact Q(x) will be uniformly continuous. Thus we have Proposition 1. There exists ɛ > 0 and δ > 0 such that for every x,y with x y < δ, Q(x) Q(y) < ε. Suppose that we have computed a finite set of values q h, h = 1, 2,..., H, that approximate Q(x h ) from below, so for some ε > 0 we have q h Q(x h ) q h + ε, h = 1, 2,..., H. Let } Q H (x) = min q h : x x h + δ1, h = 1, 2,..., H. (2) Then Q H (x) is piecewise constant with Q H (x h + δ1) = q h, and Q H (x) Q(x) ε ε. (3) An illustration of the approximation of Q(x) = x sin(10x) using points x h = 0.2, 0.4, 0.6, 0.8, is provided by Figure 1, where δ = 0.5, ε = 0 and Q H (x), shown in black is an ε-upper bound on Q(x), shown in grey. 4

5 Q(x) x Figure 1: Approximation of Q(x) = x + 0.1sin(10x) using points x h = 0.2, 0.4, 0.6, 0.8. A contour plot of Q H (x) when H = 4 for an example in two dimensions is illustrated in Figure 2. x 2 q 4 q 2 q 3 q 1 x 1 Figure 2: Contour plot of Q H (x) when H = 4. In Figure 2, the circled points are x h, h = 1, 2, 3, 4. The darker shading indicates increasing values of Q H (x) that equals Q(x h ) in each region containing x h. In general, it is easy to see that Q H (x) as defined by (2) is an upper semicontinuous function, and gives 5

6 an approximation to P which is P H : max u,x r(u) + Q H (x) s.t. x = f(u), u U, x X. One possible MIP formulation of P H is proposed in the Appendix. Lemma 2. Suppose for some H, that (u, x) is a feasible solution to P H then (u, x) is a feasible solution to P and r(u) + Q H (x) r(u) + Q(x) ε ε. Proof. The result follows from (3). Lemma 3. Any solution (ũ, f(ũ)) to P H with Q H (f(ũ)) Q(f(ũ)) + ε is a 2ε + ε-optimal solution to P. Proof. If (ũ, f(ũ)) solves P H then any optimal solution (u, f(u )) to P satisfies r(ũ) + Q H (f(ũ)) r(u ) + Q H (f(u )). Thus r(ũ) + Q(f(ũ)) + ε r(ũ) + Q H (f(ũ)) r(u ) + Q H (f(u )) r(u ) + Q(f(u )) ε ε, where the last inequality follows from (3). Now consider the following algorithm (Algorithm 1) to solve P. Algorithm 1 MIDAS for two stage problem 1. Set H = 1, x 1 = x X δ, q 1 = M. 2. Solve P H to give (u H+1, x H+1 ) where x H+1 = f(u H+1 ). 3. If x H+1 x h < δ for some h H then stop. 4. Otherwise let q H+1 = Q(x H+1 ). 5. Increase H by 1 and go to step 2. 6

7 Algorithm 1 stops the first time that x H+1 x : x x h < δ} for some h H. This means that the algorithm continues as long as x H+1 x h δ for all h H. Since X δ is compact and δ is fixed, the algorithm will terminate after a finite number of steps H with x H+1 x h < δ, for some h H. Lemma 4. Suppose for some H that (u H+1, x H+1 ) solves P H h H. Then (u H+1, x H+1 ) is a 2ε + ε solution to P. and x H+1 x h < δ, for some Proof. We have (u H+1, x H+1 ) solves P H where x H+1 = x(u H+1 ) and by Lemma 2, for every feasible (u, x) with x X δ r(u) + Q H (x) r(u) + Q(x) ε. Let N δ (x h ) = x : x x h < δ}. Since Q(x) is uniformly continuous, by Proposition 1, for every x N δ (x h ) Q(x) Q(x h ) < ε. Since x H+1 is the first iterate to be in N δ (x h ) we have, by construction, that Q H (x) = Q H (x h ) for every x N δ (x h ). Thus Q H (x H+1 ) = Q H (x h ) = q h = Q(x h ). It follows that r(u H+1 ) + Q H (x H+1 ) = r(u H+1 ) + Q(x h ) < r(u H+1 ) + Q(x H+1 ) + ε. Now Lemma 3 implies that a solution (u H+1, x H+1 ) to P H, with Q H (x H+1 ) Q(x H+1 ) + ε is a 2ε + ε-optimal solution to P. 3 Multistage optimization problems In Section 2 we discussed how MIDAS approximates the value function, and proved that this is an outer approximation which is 2ε + ε optimal. In this section we prove the convergence of the MIDAS algorithm for a deterministic multistage optimization problem. Given an initial state x, we consider a multistage deterministic optimization problem of the form: MP: max x,u T r t (x t, u t ) + Q(x T +1 ) t=1 s.t. x t+1 = f t (x t, u t ), t = 1, 2,..., T x 1 = x, u t U t (x t ), x t X t, t = 1, 2,..., T 7

8 This gives a dynamic programming formulation V t (x) = max u U(x) r t(x, u) + V t+1 (f t (x, u))}, V T +1 (x) = Q(x), where we seek a policy that maximizes V 1 ( x). We assume that r t (x, u) is continuous and increasing and for every t, V t (x) is a monotonic increasing continuous function of x. Observe that we now allow the function r to depend on the state x as well as the control. The deterministic MIDAS algorithm (Algorithm 2) applied to MP is as follows. Algorithm 2 Deterministic MIDAS 1. Set H = 1, Q H t (x) = M, t = 1, 2,..., T Forward pass: Set x H 1 = x. For t = 1 to T, (a) Solve max rt (x H u U(x H t ) t, u) + Q H t+1 (f t(x H t, u)) } to give u H t ; (b) If f t (x H t, u H t ) x h t+1 < δ for h < H then set x H t+1 = xh t+1, else set xh t+1 = f t(x H t, u H t ). 3. If for every t = 1 to T + 1 there is some h < H with x H t = x h t then stop. 4. Backward pass: Update Q H T +1 (x) to QH+1 T +1 (x) by adding qh T +1 = Q(xH T +1 ) at point xh T +1. Then for every t = T down to 1 do } (a) Solve ϕ = max r t (x H u U(x H t ) t, u) + Q H+1 t+1 (f t(x H t, u)) ; (b) Update Q H t (x) to Q H+1 t (x) by adding q H t = ϕ at point x H t. 5. Increase H by 1 and go to step 1. We now show that this algorithm will converge in step 3 after a finite number of iterations to a solution (u H 1, uh 2,..., uh T ) that is a 2T ε optimal solution to the dynamic program starting in state x. First observe that the sequence (x H 1, xh 2,..., xh T )} lies in a compact set, so it has some convergent subsequence. It follows that we can find H and h H with x H+1 t x h t < δ for every t = 1, 2,..., T + 1. We demonstrate this using the following two lemmas. Lemma 5. Let N δ (S) = x : x s < δ, s S}. Then any infinite set S lying in a compact set contains a finite subset Z with S N δ (Z). Proof. Consider an infinite set S lying in a compact set X. Since S is a subset of the closed set X, its closure S lies in X, and so S is compact. Now for some fixed δ > 0, it is easy to see that N δ (S) is a open covering of S, since any limit point of S must be in N δ (s}) for some s S. Since S is compact, there exists a finite subcover, i.e. a finite set Z = z 1, z 2,..., z N } S with S S N k=1 N δ(z k ). 8

9 Lemma 6. Suppose for every t there is some h H with x H+1 t x h t < δ. Then the sequence of actions (u H+1 1, u H+1 2,..., u H+1 ) is a 2T ε optimal solution to the dynamic program starting in state x. T Proof. If for every t = 1, 2,..., T + 1 there is some h H with x H+1 t x h t < δ, then the deterministic MIDAS algorithm stops in Step 3. We can then apply Lemma 4, to each stage problem, where we substitute V t+1 (x) for Q(x). Denote by u h t an optimal solution at state x h t to be the (true) t th stage problem of MP, so } u h t arg max r t (x h u U(x h t ) t, u) + V t+1 (f t (x h t, u)). Now apply induction on t to show that in Step 4b, we have for all t = 1, 2,..., T, and all h = 1, 2,..., H. q h t V t (x h t ) 2(t T 1)ε If t = T, then we have qt h +1 = Q(xh T +1 ) = V T +1(x h T +1 ), h = 1, 2,..., H, so ε = 0. It follows that for all h = 1, 2,..., H, } qt h = max r T (x h u U(x h T ) T, u) + Q h+1 T +1 (f T (x h T, u)) r T (x h T, u h T ) + V T +1 (f T (x h T, u h T )) 2ε by Lemma 4 suppose the result is true for t 2, 3,..., T }. Then choosing ε = 2(t T 1)ε, we have for all h = 1, 2,..., H, V t (x h t ) q h t ε. Thus Lemma 4 gives for all h = 1, 2,..., H, } qt 1 h = max r t 1 (x h t 1, u) + Q h u U(x h t 1 ) t (f t 1 (x h t 1, u)) yielding the result for t 1. r t 1 (x h t 1, u h t 1) + V t (f t 1 (x h t 1, u h t 1)) 2ε ε = V t 1 (x h t 1) 2ε + 2(t T 1)ε = V t 1 (x h t 1) + 2(t 1 T 1)ε Now when t = 1, x h 1 = x, so we have for all h = 1, 2,..., H, q h 1 = max u U( x) r 1( x, u) + Q h 2(f 1 ( x, u))} r 1 ( x, u h 1) + V 2 (f 1 ( x, u h 1)) 2T ε. showing that q h 1 is the value of a 2T ε-optimal solution to MP. We can now demonstrate the convergence of MIDAS in a finite number of steps. Theorem 1. MIDAS terminates in a finite number of iterations to a 2T ε optimal solution to the dynamic program starting in state x. 9

10 Proof. The sequence S = (x H 1, xh 2,..., xh T )} H=1,2,... generated by MIDAS lies in a compact set, so by Lemma 5 it contains a finite subsequence Z with S N δ (Z). Let H be the largest index in Z. It follows that for every t = 1, 2,..., T + 1, there exists some h H with x H +1 t x h t < δ. The result follows by Lemma 6. 4 Multistage stochastic optimization problems We extend model MP, described in Section 3, to include random noise ξ t on the state transition function f t (x t, u t, ξ t ). This creates the following multistage stochastic program MSP. [ T ] MSP: max E ξt r t (x t, u t ) + V T +1 (x T +1 ) t=1 s.t. x t+1 = f t (x t, u t, ξ t ), t = 1, 2..., T, x 1 = x u t U(x t ), t = 1, 2..., T, x t X t, t = 1, 2..., T. In MSP, we seek an optimal policy in terms of actions u t that is measurable with respect to the filtration (Ω t, F t, P t ) of ξ t for t = 1, 2,..., T. In this paper we restrict attention to a specific form of MSP that has a dynamic programming formulation. Let V t (x) denote the expected reward at the beginning of stage t onwards, given the state x, and we take action u after observing the random disturbance ξ t. Here V T +1 (x) is a given monotonic continuous function. The dynamic programming recursion is V t (x) = [ ] E ξt max t(x, u) + V t+1 (f t (x, u, ξ t ))} u U(x) V T +1 (x) = Q(x), where we seek a policy that maximizes V 1 (x). We make the following assumptions: 1. ξ t is stagewise independent. 2. The set Ω t of random outcomes in each stage t = 1, 2,..., T is discrete and finite (Ω t = ξ ti i = 1,..., N < } with probabilities P ti > 0). 3. X t is a compact set. The particular form of ξ t we assume yields a finite set of possible sample paths or scenarios ω(j) : j = 1, 2,..., T t=1 N t}. The MIDAS algorithm selects one of these randomly in each forward pass. This gives Algorithm 3 described below: 10

11 Algorithm 3 Sampled MIDAS Set H = 1, Q H t (x) = M, t = 1, 2,..., T + 1. While H < H max do 1. Forward pass: Set x H 1 = x. For t = 1 to T : (a) Sample ξ t to give ξ H t Ω t ; (b) Solve max rt (x H u U(x H t ) t, u) + Q H t+1(f t (x H t, u, ξt H )) } to give u H t ; (c) If f t (x H t, u H t, ξt H ) x h t+1 < δ for h < H then set x H t+1 = xh t+1, else set xh t+1 = f t (x H t, u H t, ξt H ). 2. Backward pass: Update Q H T +1 (x) to QH+1 T +1 (x) by adding qh T +1 = V T +1(x H T +1 ) at point xh T +1. Then for every t = T down to 1 do (a) Solve ϕ = [ N i=1 P ti max r t (x H u U(x H t ) t, u) + Q H+1 t+1 (f t(x H t, u, ξ ti ))} ] ; (b) Update Q H t (x) to Q H+1 t (x) by adding q H t = ϕ at point x H t. 3. Increase H by 1 and go to step 1. Observe that Sampled MIDAS has no stopping criterion. However we have the following result. Lemma 7. There is some H such that for every H > H, f t (x H t, u H t, ξt H ) x h t+1 < δ for some h < H. Proof. Consider the event tree defined by the random variable ξ t, consisting of nodes n N. Any policy when implemented defines a state variable x(n) for each node n N of this tree. Let x = x(n), n N }. Each forward pass H updates values of x(n) (from x H (n) to x H+1 (n)) for nodes n corresponding to ξt H, t = 1, 2,..., T. For nodes n not on this sample path we let x H+1 (n) = x H (n), thus defining x H+1. This defines an infinite sequence S = x H } H= of points, all of which lie in a compact set. By Lemma 5 S contains a finite subsequence Z with S N δ (Z). Let H be the largest index in Z. Suppose it is not true that for every t = 1, 2,..., T, and every ξ t Ω t and every H > H there is some h H with f t (x H t, u H t, ξ t ) x h t+1 < δ. Then for some t and ξ t (corresponding to node n N), and some H > H, f t (x H t, u H t, ξ t ) x h t+1 δ for every h H, so x H+1 (n) x h t+1 δ, giving x H+1 t+1 / N δ (Z), a contradiction. By Lemma 7, after H iterations every sample path produced in step 1 yields state values that have been visited before. It is clear that no additional points will be added in subsequent backward passes and so the MIDAS algorithm will simply replicate points that it has already computed with no improvement in its approximation of Q t+1 (x) at each t. If this situation could be identified then we might terminate the algorithm at this point. This then raises the question of whether the solution obtained at this point 11

12 is a 2T ε-optimal solution to MSP. The answer to this latter question depends on the sampling strategy used in step 1a. Following [8] we assume that this satisfies the Forward Pass Sampling Property. Forward Pass Sampling Property (FPSP): For each j = 1, 2,..., T 1 t=2 N t, with probability 1 H : ξt H t = 1, 2, 3,..., T } = ω(j) } =. FPSP states that each scenario ω(j) is traversed infinitely many times with probability 1 in the forward pass. There are many sampling methods satisfying this property. For example, one sampling method would be independently sampling a single outcome in each stage with a positive probability for each ξ ti in the forward pass, which meets FPSP by the Borel-Cantelli lemma. Another sampling method that satisfies FPSP is to repeat an exhaustive enumeration of each scenario ω(j), j = 1, 2,..., T t=1 N t in the forward pass. Theorem 2. If step 1a satisfies FPSP then sampled MIDAS converges almost surely to a 2T ε-optimal policy of MSP. Proof. We prove the result by induction. Consider the sequence S = x H } H>H defined by Lemma 7. Each x H (n) in S is some point that has been visited in a previous iteration of sampled MIDAS, and no new points are added. Consider any leaf node m in N, corresponding to t = T + 1. By the FPSP this is visited an infinite number of times, so at least once by some iteration H > H with probability 1. The collection x H (m) : H = 1, 2,...} = x h T +1, h = 1, 2,..., H }. For all such h, the terminal value function V T +1 (x h T +1 ) is known exactly, so ε = 0 in the conditions of Lemma 4. It follows that for each x h T, h = 1, 2,..., H and each ξ T,i ( )} max r T (x h u U(x h T ) T, u) + Q H T +1 f T (x h T, u, ξ T,i ) ( ) max r T (x h u U(x h T ) T, u) + V T +1 f T (x h T, u, ξ T,i ) 2ε. so [ P T i i i = i max r T (x h u U(x h T ) T, u) + Q H T +1 ( )} ] f T (x h T, u, ξ T,i ) [ ] )} P T i max r T (x h u U(x h T ) T, u) + V T +1 (f T (x h T, u, ξ T,i ) 2ε [ P T i max r T (x h u U(x h T ) T, u) + V T +1 (f T (x h T, u, ξ T,i ))} ] 2ε. Now suppose that for some t 2, 3,..., T } and each x h t, h = 1, 2,..., H [ ( )} ] P ti max r t (x h u U(x i h t ) t, u) + Q H t+1 f t (x h t, u, ξ t,i ) [ ( P ti max r t (x h u U(x h t ) t, u) + Q H t+1 f t (x h t, u, ξ t,i ))} ] ε. i 12

13 where ε = 2(t T 1)ε. Thus step 2 of Sampled MIDAS and Lemma 4 give for all h = 1, 2,..., H, qt 1 h = [ } ] P t 1,i max r t 1 (x h t 1, u) + Q h u U(x i h t 1 ) t (f t 1 (x h t 1, u, ξ t,i )) [ ] } P t 1,i max r t 1 (x h t 1, u) + Q h u U(x i h t 1 ) t (f t 1 (x h t 1, u, ξ t,i )) 2ε ε = [ P t 1,i max r t 1 (x h t 1, u) + Q h u U(x i h t 1 ) t (f t 1 (x h t 1, u, ξ t,i ))} ] 2ε ε = V t 1 (x h t 1) 2ε + 2(t T 1)ε = V t 1 (x h t 1) + 2(t 1 T 1)ε yielding the result for t 1. Then by induction we have Q H 1 ( x) V 1 ( x) + 2(1 T 1)ε = V 1 ( x) 2T ε yielding the result. 5 Conclusion In this paper, we have proposed a method called MIDAS for solving multistage stochastic dynamic programming problems with monotonic Bellman functions. We have demonstrated the almost-sure convergence of MIDAS to a 2T ε-optimal policy in a finite number of steps. The quality of the optimal policy will depend on the value of δ, since a smaller δ will give a lower ε. We may expect that the number of iterations will increase as δ decreases. Furthermore, since the current formulation treats each stage problem as a MIP, then with each iteration, the stage problems will increase in size. So small δ values will give more iterations, each of which involves problem instances of increasing size. There remains considerable scope for tuning the algorithm by choosing δ adaptively as the algorithm proceeds. MIDAS can be extended in several ways. Currently, we approximate the stage problems using a MIP formulation that treats the Bellman functions as piecewise constant. One formulation of such a MIP is given in the Appendix. A more accurate approximation might be achieved using an alternative formulation that for example uses specially ordered sets to yield a piecewise affine approximation. As long as this provides an ε-outer approximation of the Bellman function, we can incorporate it into a MIDAS scheme, which will converge almost surely by the same arguments above. Although our convergence result applies to problems with continuous Bellman functions, MIDAS can be applied to multistage stochastic integer programming (MSIP) problems. Solving a deterministic equivalent of a MSIP can be difficult, as the scenario tree grows exponentially with the number of stages and the number of outcomes, and MIP algorithms generally scale poorly. However, since solving many small MIPs will be faster than solving a single large MIP, we might expect MIDAS to produce good candidate solutions to MSIP problems with much less computational effort. Our hope is that MIDAS 13

14 will provide a practical computational approach to solving these difficult multistage stochastic integer programming problems. Appendix: A MIP representation of Q H (x) Consider the two-stage problem P : max u r(u) + Q(x) s.t. x = f(u), u U, x X, where Q(x) is continuous and monotonic increasing, and X = x : 0 x i K i, i = 1, 2,..., n}. Since X is compact, we have an upper bound M on Q(x). Suppose we have computed a finite set of values q h, h = 1, 2,..., H, that approximate Q(x h ), so Q(x h ) q h ε, h = 1, 2,..., H. (For example we could choose ε = 0, if Q(x h ) = q h, h = 1, 2,..., H.). For any x X δ = x : δ x i K i, i = 1, 2,..., n}, define Q H (x) to be the optimal value of the mixed integer program Q H (x) = max ϕ s.t. ϕ q h + (M q h )(1 w h ), h = 1, 2,..., H, x k x h k zh k + δ, k = 1, 2,..., n, n k=1 zh k = 1 w h, h = 1, 2,..., H, w h 0, 1}, h = 1, 2,..., H, zk h 0, 1}, k = 1, 2,..., n, h = 1, 2,..., H. and for x X \ X δ set Q H (x) = minq H (y) : y x, y X δ }. The problem P is approximated by P H : max u r(u) + Q H (x) s.t. x = f(u), u U, x X. Since Q is uniformly continuous, there exists ε and δ so that for every x, y with x y < δ we have Q(x) Q(y) < ε. We now show that Q H (x) gives an upper bound on Q(x), to a tolerance level of ε + ε. Proposition 2. For every x X, Q H (x) Q(x) ε ε. 14

15 Proof. Consider first a given point x X δ, and any w h, z h k that is feasible for P H. If w h = 0, h = 1, 2,..., H, then ϕ M is the only constraint on ϕ and so at optimality yielding the result. ϕ = M Q(x), Otherwise assume that the optimal solution to P H (x) has w h = 1 for some h. We define H(x) = h : x k < x h k + δ for every k}, and show that Q H (x) = minq h : h H(x)}. (4) First if h H(x) then w h = 1. This is because choosing w h = 0 implies zk h = 1 for some k, so for at least one k x k x h k + δ, so h / H(x). Now if h / H(x) then any feasible solution to P H (x) can have either w h = 0 or w h = 1. Observe however that if q h < minq h : h H(x)} then choosing w h = 1 would yield a value of ϕ strictly lower than the value obtained by choosing w h = 0, and so choosing w h = 1 cannot maximize ϕ. It follows that Thus the optimal value minq h : h H(x)} minq h : h / H(x), w h = 1}. Q H (x) = minq h : w h = 1} = minq h : h H(x)}. Now by monotonicity of Q, for every h H(x), Q(x) Q(x h + δe), so Q(x) q h Q(x h + δe) q h Q(x h + δe) Q(x h ) + ε < ε + ε by continuity of Q. Since Q H (x) = minq h : h H(x)}, we thus have as required. Q H (x) Q(x) ε ε, To complete the proof, suppose finally that x X \ X δ, so Q H (x) = minq H (y) : y x, y X δ }. Then by the argument above applied to each y X δ, minq H (y) : y x, y X δ } minq(y) ε ε : y x, y X δ } Q(x) ε ε since Q is monotonic increasing. 15

16 References [1] H. Abgottspon, K. Njalsson, M.A. Bucher, and G. Andersson. Risk-averse medium-term hydro optimization considering provision of spinning reserves. In 2014 International Conference on Probabilistic Methods Applied to Power Systems (PMAPS), pages 1 6. IEEE, [2] S. Cerisola, J.M. Latorre, and A. Ramos. Stochastic dual dynamic programming applied to nonconvex hydrothermal models. European Journal of Operational Research, 218(3): , [3] Z. Chen and W.B. Powell. Convergent cutting-plane and partial-sampling algorithm for multistage stochastic linear programs with recourse. Journal of Optimization Theory and Applications, 102(3): , [4] C.J. Donohue and J.R. Birge. The abridged nested decomposition method for multistage stochastic linear programs with relatively complete recourse. Algorithmic Operations Research, 1(1), [5] P. Girardeau, V. Leclere, and A.B. Philpott. On the convergence of decomposition methods for multistage stochastic convex programs. Mathematics of Operations Research, 40(1): , [6] M. Hindsberger and A.B. Philpott. Resa: A method for solving multistage stochastic linear programs. Journal of Applied Operational Research, 6(1):2 15, [7] M.V.F. Pereira and L.M.V.G. Pinto. Multistage stochastic optimization applied to energy planning. Mathematical Programming, 52(1-3): , [8] A.B. Philpott and Z. Guan. On the convergence of stochastic dual dynamic programming and related methods. Operations Research Letters, 36(4): , [9] A. Shapiro. On complexity of multistage stochastic programs. Operations Research Letters, 34(1): 1 8, [10] A. Shapiro. Analysis of stochastic dual dynamic programming method. European Journal of Operational Research, 209(1):63 72, [11] G. Steeger and S. Rebennack. Strategic bidding for multiple price-maker hydroelectric producers. IIE Transactions, 47(9): , [12] F. Thome, M.V.F. Pereira, S. Granville, and M. Fampa. Non-convexities representation on hydrothermal operation planning using SDDP. Technical report, working paper, available: psr-inc. com. br/portal/psr/publicacoes,

MIDAS: A Mixed Integer Dynamic Approximation Scheme

MIDAS: A Mixed Integer Dynamic Approximation Scheme MIDAS: A Mixed Integer Dynamic Approximation Scheme Andy Philpott, Faisal Wahid, Frédéric Bonnans June 16, 2016 Abstract Mixed Integer Dynamic Approximation Scheme (MIDAS) is a new sampling-based algorithm

More information

Decomposition in multistage stochastic integer programming

Decomposition in multistage stochastic integer programming Decomposition in multistage stochastic integer programming Andy Philpott Electric Power Optimization Centre University of Auckland. www.epoc.org.nz NTNU Winter School, Passo Tonale, January 15, 2017. Plans

More information

The MIDAS touch: EPOC Winter Workshop 9th September solving hydro bidding problems using mixed integer programming

The MIDAS touch: EPOC Winter Workshop 9th September solving hydro bidding problems using mixed integer programming The MIDAS touch: solving hydro bidding problems using mixed integer programming EPOC Winter Workshop 9th September 2015 Faisal Wahid, Cédric Gouvernet, Prof. Andy Philpott, Prof. Frédéric Bonnans 0 / 30

More information

arxiv: v2 [math.oc] 18 Nov 2017

arxiv: v2 [math.oc] 18 Nov 2017 DASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS arxiv:1711.04650v2 [math.oc] 18 Nov 2017 Vincent Guigues School of Applied Mathematics, FGV Praia

More information

Stochastic Dual Dynamic Integer Programming

Stochastic Dual Dynamic Integer Programming Stochastic Dual Dynamic Integer Programming Jikai Zou Shabbir Ahmed Xu Andy Sun December 26, 2017 Abstract Multistage stochastic integer programming (MSIP) combines the difficulty of uncertainty, dynamics,

More information

Stochastic dual dynamic programming with stagewise dependent objective uncertainty

Stochastic dual dynamic programming with stagewise dependent objective uncertainty Stochastic dual dynamic programming with stagewise dependent objective uncertainty Anthony Downward a,, Oscar Dowson a, Regan Baucke a a Department of Engineering Science, University of Auckland, New Zealand

More information

Stochastic Dual Dynamic Programming with CVaR Risk Constraints Applied to Hydrothermal Scheduling. ICSP 2013 Bergamo, July 8-12, 2012

Stochastic Dual Dynamic Programming with CVaR Risk Constraints Applied to Hydrothermal Scheduling. ICSP 2013 Bergamo, July 8-12, 2012 Stochastic Dual Dynamic Programming with CVaR Risk Constraints Applied to Hydrothermal Scheduling Luiz Carlos da Costa Junior Mario V. F. Pereira Sérgio Granville Nora Campodónico Marcia Helena Costa Fampa

More information

CONVERGENCE ANALYSIS OF SAMPLING-BASED DECOMPOSITION METHODS FOR RISK-AVERSE MULTISTAGE STOCHASTIC CONVEX PROGRAMS

CONVERGENCE ANALYSIS OF SAMPLING-BASED DECOMPOSITION METHODS FOR RISK-AVERSE MULTISTAGE STOCHASTIC CONVEX PROGRAMS CONVERGENCE ANALYSIS OF SAMPLING-BASED DECOMPOSITION METHODS FOR RISK-AVERSE MULTISTAGE STOCHASTIC CONVEX PROGRAMS VINCENT GUIGUES Abstract. We consider a class of sampling-based decomposition methods

More information

Regularized optimization techniques for multistage stochastic programming

Regularized optimization techniques for multistage stochastic programming Regularized optimization techniques for multistage stochastic programming Felipe Beltrán 1, Welington de Oliveira 2, Guilherme Fredo 1, Erlon Finardi 1 1 UFSC/LabPlan Universidade Federal de Santa Catarina

More information

Upper bound for optimal value of risk averse multistage problems

Upper bound for optimal value of risk averse multistage problems Upper bound for optimal value of risk averse multistage problems Lingquan Ding School of Industrial and Systems Engineering Georgia Institute of Technology Atlanta, GA 30332-0205 Alexander Shapiro School

More information

MEDIUM-TERM HYDROPOWER SCHEDULING BY STOCHASTIC DUAL DYNAMIC INTEGER PROGRAMMING: A CASE STUDY

MEDIUM-TERM HYDROPOWER SCHEDULING BY STOCHASTIC DUAL DYNAMIC INTEGER PROGRAMMING: A CASE STUDY MEDIUM-TERM HYDROPOWER SCHEDULING BY STOCHASTIC DUAL DYNAMIC INTEGER PROGRAMMING: A CASE STUDY Martin Hjelmeland NTNU Norwegian University of Science and Technology, Trondheim, Norway. martin.hjelmeland@ntnu.no

More information

A deterministic algorithm for solving multistage stochastic programming problems

A deterministic algorithm for solving multistage stochastic programming problems A deterministic algorithm for solving multistage stochastic programming problems Regan Bauce a,b,, Anthony Downward a, Golbon Zaeri a,b a Electrical Power Optimization Centre at The University of Aucland,

More information

Risk neutral and risk averse approaches to multistage stochastic programming.

Risk neutral and risk averse approaches to multistage stochastic programming. Risk neutral and risk averse approaches to multistage stochastic programming. A. Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205, USA

More information

Stochastic Integer Programming

Stochastic Integer Programming IE 495 Lecture 20 Stochastic Integer Programming Prof. Jeff Linderoth April 14, 2003 April 14, 2002 Stochastic Programming Lecture 20 Slide 1 Outline Stochastic Integer Programming Integer LShaped Method

More information

RESEARCH ARTICLE. On cutting plane algorithms and dynamic programming for hydroelectricity generation

RESEARCH ARTICLE. On cutting plane algorithms and dynamic programming for hydroelectricity generation Optimization Methods and Software Vol. 00, No. 00, Month 2011, 1 20 RESEARCH ARTICLE On cutting plane algorithms and dynamic programming for hydroelectricity generation Andy Philpott a, Anes Dallagi b

More information

A deterministic algorithm for solving stochastic minimax dynamic programmes

A deterministic algorithm for solving stochastic minimax dynamic programmes A deterministic algorithm for solving stochastic minimax dynamic programmes Regan Baucke a,b,, Anthony Downward a,b, Golbon Zakeri a,b a The Department of Engineering Science, The University of Auckland,

More information

Dual Dynamic Programming for Long-Term Receding-Horizon Hydro Scheduling with Head Effects

Dual Dynamic Programming for Long-Term Receding-Horizon Hydro Scheduling with Head Effects Dual Dynamic Programming for Long-Term Receding-Horizon Hydro Scheduling with Head Effects Benjamin Flamm, Annika Eichler, Joseph Warrington, John Lygeros Automatic Control Laboratory, ETH Zurich SINTEF

More information

The policy graph decomposition of multistage stochastic optimization problems

The policy graph decomposition of multistage stochastic optimization problems O R I G I N A L A R T I C L E The policy graph decomposition of multistage stochastic optimization problems Oscar Dowson Correspondence Department of Industrial Engineering and Management Sciences, Northwestern

More information

An Adaptive Partition-based Approach for Solving Two-stage Stochastic Programs with Fixed Recourse

An Adaptive Partition-based Approach for Solving Two-stage Stochastic Programs with Fixed Recourse An Adaptive Partition-based Approach for Solving Two-stage Stochastic Programs with Fixed Recourse Yongjia Song, James Luedtke Virginia Commonwealth University, Richmond, VA, ysong3@vcu.edu University

More information

Birgit Rudloff Operations Research and Financial Engineering, Princeton University

Birgit Rudloff Operations Research and Financial Engineering, Princeton University TIME CONSISTENT RISK AVERSE DYNAMIC DECISION MODELS: AN ECONOMIC INTERPRETATION Birgit Rudloff Operations Research and Financial Engineering, Princeton University brudloff@princeton.edu Alexandre Street

More information

On the relation between concavity cuts and the surrogate dual for convex maximization problems

On the relation between concavity cuts and the surrogate dual for convex maximization problems On the relation between concavity cuts and the surrogate dual for convex imization problems Marco Locatelli Dipartimento di Ingegneria Informatica, Università di Parma Via G.P. Usberti, 181/A, 43124 Parma,

More information

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions Angelia Nedić and Asuman Ozdaglar April 15, 2006 Abstract We provide a unifying geometric framework for the

More information

Solution Methods for Stochastic Programs

Solution Methods for Stochastic Programs Solution Methods for Stochastic Programs Huseyin Topaloglu School of Operations Research and Information Engineering Cornell University ht88@cornell.edu August 14, 2010 1 Outline Cutting plane methods

More information

Multistage Stochastic Unit Commitment Using Stochastic Dual Dynamic Integer Programming

Multistage Stochastic Unit Commitment Using Stochastic Dual Dynamic Integer Programming Multistage Stochastic Unit Commitment Using Stochastic Dual Dynamic Integer Programming Jikai Zou Shabbir Ahmed Xu Andy Sun May 14, 2017 Abstract Unit commitment (UC) is a key operational problem in power

More information

Branch-and-cut Approaches for Chance-constrained Formulations of Reliable Network Design Problems

Branch-and-cut Approaches for Chance-constrained Formulations of Reliable Network Design Problems Branch-and-cut Approaches for Chance-constrained Formulations of Reliable Network Design Problems Yongjia Song James R. Luedtke August 9, 2012 Abstract We study solution approaches for the design of reliably

More information

Operations Research Letters. On a time consistency concept in risk averse multistage stochastic programming

Operations Research Letters. On a time consistency concept in risk averse multistage stochastic programming Operations Research Letters 37 2009 143 147 Contents lists available at ScienceDirect Operations Research Letters journal homepage: www.elsevier.com/locate/orl On a time consistency concept in risk averse

More information

3.10 Lagrangian relaxation

3.10 Lagrangian relaxation 3.10 Lagrangian relaxation Consider a generic ILP problem min {c t x : Ax b, Dx d, x Z n } with integer coefficients. Suppose Dx d are the complicating constraints. Often the linear relaxation and the

More information

The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 44

The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 44 1 / 44 The L-Shaped Method Operations Research Anthony Papavasiliou Contents 2 / 44 1 The L-Shaped Method [ 5.1 of BL] 2 Optimality Cuts [ 5.1a of BL] 3 Feasibility Cuts [ 5.1b of BL] 4 Proof of Convergence

More information

Distributed Optimization. Song Chong EE, KAIST

Distributed Optimization. Song Chong EE, KAIST Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links

More information

Stochastic Programming: From statistical data to optimal decisions

Stochastic Programming: From statistical data to optimal decisions Stochastic Programming: From statistical data to optimal decisions W. Römisch Humboldt-University Berlin Department of Mathematics (K. Emich, H. Heitsch, A. Möller) Page 1 of 24 6th International Conference

More information

4y Springer NONLINEAR INTEGER PROGRAMMING

4y Springer NONLINEAR INTEGER PROGRAMMING NONLINEAR INTEGER PROGRAMMING DUAN LI Department of Systems Engineering and Engineering Management The Chinese University of Hong Kong Shatin, N. T. Hong Kong XIAOLING SUN Department of Mathematics Shanghai

More information

Bounds on Risk-averse Mixed-integer Multi-stage Stochastic Programming Problems with Mean-CVaR

Bounds on Risk-averse Mixed-integer Multi-stage Stochastic Programming Problems with Mean-CVaR Bounds on Risk-averse Mixed-integer Multi-stage Stochastic Programming Problems with Mean-CVaR Ali rfan Mahmuto ullar, Özlem Çavu³, and M. Selim Aktürk Department of Industrial Engineering, Bilkent University,

More information

DETERMINISTIC AND STOCHASTIC SELECTION DYNAMICS

DETERMINISTIC AND STOCHASTIC SELECTION DYNAMICS DETERMINISTIC AND STOCHASTIC SELECTION DYNAMICS Jörgen Weibull March 23, 2010 1 The multi-population replicator dynamic Domain of analysis: finite games in normal form, G =(N, S, π), with mixed-strategy

More information

Exact converging bounds for Stochastic Dual Dynamic Programming via Fenchel duality

Exact converging bounds for Stochastic Dual Dynamic Programming via Fenchel duality Exact converging bounds for Stochastic Dual Dynamic Programming via Fenchel duality Vincent Leclère, Pierre Carpentier, Jean-Philippe Chancelier, Arnaud Lenoir and François Pacaud April 12, 2018 Abstract

More information

Network Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini

Network Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini In the name of God Network Flows 6. Lagrangian Relaxation 6.3 Lagrangian Relaxation and Integer Programming Fall 2010 Instructor: Dr. Masoud Yaghini Integer Programming Outline Branch-and-Bound Technique

More information

MDP Preliminaries. Nan Jiang. February 10, 2019

MDP Preliminaries. Nan Jiang. February 10, 2019 MDP Preliminaries Nan Jiang February 10, 2019 1 Markov Decision Processes In reinforcement learning, the interactions between the agent and the environment are often described by a Markov Decision Process

More information

Risk Measures. A. Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia , USA ICSP 2016

Risk Measures. A. Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia , USA ICSP 2016 Risk Measures A. Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205, USA ICSP 2016 Min-max (distributionally robust) approach to stochastic

More information

Stochastic Shortest Path Problems

Stochastic Shortest Path Problems Chapter 8 Stochastic Shortest Path Problems 1 In this chapter, we study a stochastic version of the shortest path problem of chapter 2, where only probabilities of transitions along different arcs can

More information

INEXACT CUTS IN BENDERS' DECOMPOSITION GOLBON ZAKERI, ANDREW B. PHILPOTT AND DAVID M. RYAN y Abstract. Benders' decomposition is a well-known techniqu

INEXACT CUTS IN BENDERS' DECOMPOSITION GOLBON ZAKERI, ANDREW B. PHILPOTT AND DAVID M. RYAN y Abstract. Benders' decomposition is a well-known techniqu INEXACT CUTS IN BENDERS' DECOMPOSITION GOLBON ZAKERI, ANDREW B. PHILPOTT AND DAVID M. RYAN y Abstract. Benders' decomposition is a well-known technique for solving large linear programs with a special

More information

Lecture 1. Stochastic Optimization: Introduction. January 8, 2018

Lecture 1. Stochastic Optimization: Introduction. January 8, 2018 Lecture 1 Stochastic Optimization: Introduction January 8, 2018 Optimization Concerned with mininmization/maximization of mathematical functions Often subject to constraints Euler (1707-1783): Nothing

More information

Stability of Stochastic Programming Problems

Stability of Stochastic Programming Problems Stability of Stochastic Programming Problems W. Römisch Humboldt-University Berlin Institute of Mathematics 10099 Berlin, Germany http://www.math.hu-berlin.de/~romisch Page 1 of 35 Spring School Stochastic

More information

CS261: Problem Set #3

CS261: Problem Set #3 CS261: Problem Set #3 Due by 11:59 PM on Tuesday, February 23, 2016 Instructions: (1) Form a group of 1-3 students. You should turn in only one write-up for your entire group. (2) Submission instructions:

More information

Generalized Dual Dynamic Programming for Infinite Horizon Problems in Continuous State and Action Spaces

Generalized Dual Dynamic Programming for Infinite Horizon Problems in Continuous State and Action Spaces Generalized Dual Dynamic Programming for Infinite Horizon Problems in Continuous State and Action Spaces Joseph Warrington, Paul N. Beuchat, and John Lygeros Abstract We describe a nonlinear generalization

More information

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Michael Patriksson 0-0 The Relaxation Theorem 1 Problem: find f := infimum f(x), x subject to x S, (1a) (1b) where f : R n R

More information

Can Li a, Ignacio E. Grossmann a,

Can Li a, Ignacio E. Grossmann a, A generalized Benders decomposition-based branch and cut algorithm for two-stage stochastic programs with nonconvex constraints and mixed-binary first and second stage variables Can Li a, Ignacio E. Grossmann

More information

Propp-Wilson Algorithm (and sampling the Ising model)

Propp-Wilson Algorithm (and sampling the Ising model) Propp-Wilson Algorithm (and sampling the Ising model) Danny Leshem, Nov 2009 References: Haggstrom, O. (2002) Finite Markov Chains and Algorithmic Applications, ch. 10-11 Propp, J. & Wilson, D. (1996)

More information

ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko

ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko Indirect Utility Recall: static consumer theory; J goods, p j is the price of good j (j = 1; : : : ; J), c j is consumption

More information

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018 EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal

More information

Stochastic Programming Models in Design OUTLINE

Stochastic Programming Models in Design OUTLINE Stochastic Programming Models in Design John R. Birge University of Michigan OUTLINE Models General - Farming Structural design Design portfolio General Approximations Solutions Revisions Decision: European

More information

WE propose an approximate dynamic programming

WE propose an approximate dynamic programming IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL 58, NO 12, DECEMBER 2013 2995 An Optimal Approximate Dynamic Programming Algorithm for Concave, Scalar Storage Problems With Vector-Valued Controls Juliana Nascimento

More information

Slides II - Dynamic Programming

Slides II - Dynamic Programming Slides II - Dynamic Programming Julio Garín University of Georgia Macroeconomic Theory II (Ph.D.) Spring 2017 Macroeconomic Theory II Slides II - Dynamic Programming Spring 2017 1 / 32 Outline 1. Lagrangian

More information

Optimization Tools in an Uncertain Environment

Optimization Tools in an Uncertain Environment Optimization Tools in an Uncertain Environment Michael C. Ferris University of Wisconsin, Madison Uncertainty Workshop, Chicago: July 21, 2008 Michael Ferris (University of Wisconsin) Stochastic optimization

More information

arxiv: v3 [math.oc] 25 Apr 2018

arxiv: v3 [math.oc] 25 Apr 2018 Problem-driven scenario generation: an analytical approach for stochastic programs with tail risk measure Jamie Fairbrother *, Amanda Turner *, and Stein W. Wallace ** * STOR-i Centre for Doctoral Training,

More information

Stability-based generation of scenario trees for multistage stochastic programs

Stability-based generation of scenario trees for multistage stochastic programs Stability-based generation of scenario trees for multistage stochastic programs H. Heitsch and W. Römisch Humboldt-University Berlin Institute of Mathematics 10099 Berlin, Germany http://www.math.hu-berlin.de/~romisch

More information

CS 7180: Behavioral Modeling and Decisionmaking

CS 7180: Behavioral Modeling and Decisionmaking CS 7180: Behavioral Modeling and Decisionmaking in AI Markov Decision Processes for Complex Decisionmaking Prof. Amy Sliva October 17, 2012 Decisions are nondeterministic In many situations, behavior and

More information

Value and Policy Iteration

Value and Policy Iteration Chapter 7 Value and Policy Iteration 1 For infinite horizon problems, we need to replace our basic computational tool, the DP algorithm, which we used to compute the optimal cost and policy for finite

More information

Complexity of two and multi-stage stochastic programming problems

Complexity of two and multi-stage stochastic programming problems Complexity of two and multi-stage stochastic programming problems A. Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205, USA The concept

More information

Worst case analysis for a general class of on-line lot-sizing heuristics

Worst case analysis for a general class of on-line lot-sizing heuristics Worst case analysis for a general class of on-line lot-sizing heuristics Wilco van den Heuvel a, Albert P.M. Wagelmans a a Econometric Institute and Erasmus Research Institute of Management, Erasmus University

More information

Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise

Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise Robert J. Elliott 1 Samuel N. Cohen 2 1 Department of Commerce, University of South Australia 2 Mathematical Insitute, University

More information

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018 Section Notes 9 Midterm 2 Review Applied Math / Engineering Sciences 121 Week of December 3, 2018 The following list of topics is an overview of the material that was covered in the lectures and sections

More information

A Benders Algorithm for Two-Stage Stochastic Optimization Problems With Mixed Integer Recourse

A Benders Algorithm for Two-Stage Stochastic Optimization Problems With Mixed Integer Recourse A Benders Algorithm for Two-Stage Stochastic Optimization Problems With Mixed Integer Recourse Ted Ralphs 1 Joint work with Menal Güzelsoy 2 and Anahita Hassanzadeh 1 1 COR@L Lab, Department of Industrial

More information

Can Li a, Ignacio E. Grossmann a,

Can Li a, Ignacio E. Grossmann a, A generalized Benders decomposition-based branch and cut algorithm for two-stage stochastic programs with nonconvex constraints and mixed-binary first and second stage variables Can Li a, Ignacio E. Grossmann

More information

On the Approximate Linear Programming Approach for Network Revenue Management Problems

On the Approximate Linear Programming Approach for Network Revenue Management Problems On the Approximate Linear Programming Approach for Network Revenue Management Problems Chaoxu Tong School of Operations Research and Information Engineering, Cornell University, Ithaca, New York 14853,

More information

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm You have 80 minutes. The exam is closed book, closed notes except a one-page crib sheet, basic calculators only.

More information

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions Angelia Nedić and Asuman Ozdaglar April 16, 2006 Abstract In this paper, we study a unifying framework

More information

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.

More information

DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION

DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION UNIVERSITY OF MARYLAND: ECON 600. Alternative Methods of Discrete Time Intertemporal Optimization We will start by solving a discrete time intertemporal

More information

Introduction to Nonlinear Stochastic Programming

Introduction to Nonlinear Stochastic Programming School of Mathematics T H E U N I V E R S I T Y O H F R G E D I N B U Introduction to Nonlinear Stochastic Programming Jacek Gondzio Email: J.Gondzio@ed.ac.uk URL: http://www.maths.ed.ac.uk/~gondzio SPS

More information

Lecture notes for Analysis of Algorithms : Markov decision processes

Lecture notes for Analysis of Algorithms : Markov decision processes Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with

More information

Stochastic Decision Diagrams

Stochastic Decision Diagrams Stochastic Decision Diagrams John Hooker CORS/INFORMS Montréal June 2015 Objective Relaxed decision diagrams provide an generalpurpose method for discrete optimization. When the problem has a dynamic programming

More information

Scenario Grouping and Decomposition Algorithms for Chance-constrained Programs

Scenario Grouping and Decomposition Algorithms for Chance-constrained Programs Scenario Grouping and Decomposition Algorithms for Chance-constrained Programs Siqian Shen Dept. of Industrial and Operations Engineering University of Michigan Joint work with Yan Deng (UMich, Google)

More information

Lagrangian Relaxation in MIP

Lagrangian Relaxation in MIP Lagrangian Relaxation in MIP Bernard Gendron May 28, 2016 Master Class on Decomposition, CPAIOR2016, Banff, Canada CIRRELT and Département d informatique et de recherche opérationnelle, Université de Montréal,

More information

Project Discussions: SNL/ADMM, MDP/Randomization, Quadratic Regularization, and Online Linear Programming

Project Discussions: SNL/ADMM, MDP/Randomization, Quadratic Regularization, and Online Linear Programming Project Discussions: SNL/ADMM, MDP/Randomization, Quadratic Regularization, and Online Linear Programming Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305,

More information

Lagrange Relaxation: Introduction and Applications

Lagrange Relaxation: Introduction and Applications 1 / 23 Lagrange Relaxation: Introduction and Applications Operations Research Anthony Papavasiliou 2 / 23 Contents 1 Context 2 Applications Application in Stochastic Programming Unit Commitment 3 / 23

More information

Adaptive Dynamic Cost Updating Procedure for Solving Fixed Charge Network Flow Problems.

Adaptive Dynamic Cost Updating Procedure for Solving Fixed Charge Network Flow Problems. Adaptive Dynamic Cost Updating Procedure for Solving Fixed Charge Network Flow Problems. Artyom Nahapetyan, Panos Pardalos Center for Applied Optimization Industrial and Systems Engineering Department

More information

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. 35, No., May 010, pp. 84 305 issn 0364-765X eissn 156-5471 10 350 084 informs doi 10.187/moor.1090.0440 010 INFORMS On the Power of Robust Solutions in Two-Stage

More information

6.254 : Game Theory with Engineering Applications Lecture 7: Supermodular Games

6.254 : Game Theory with Engineering Applications Lecture 7: Supermodular Games 6.254 : Game Theory with Engineering Applications Lecture 7: Asu Ozdaglar MIT February 25, 2010 1 Introduction Outline Uniqueness of a Pure Nash Equilibrium for Continuous Games Reading: Rosen J.B., Existence

More information

Basic Deterministic Dynamic Programming

Basic Deterministic Dynamic Programming Basic Deterministic Dynamic Programming Timothy Kam School of Economics & CAMA Australian National University ECON8022, This version March 17, 2008 Motivation What do we do? Outline Deterministic IHDP

More information

Column Generation in Integer Programming with Applications in Multicriteria Optimization

Column Generation in Integer Programming with Applications in Multicriteria Optimization Column Generation in Integer Programming with Applications in Multicriteria Optimization Matthias Ehrgott Department of Engineering Science The University of Auckland, New Zealand email: m.ehrgott@auckland.ac.nz

More information

Optimization methods

Optimization methods Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,

More information

Multistage Robust Mixed Integer Optimization with Adaptive Partitions

Multistage Robust Mixed Integer Optimization with Adaptive Partitions Vol. 00, No. 0, Xxxxx 0000, pp. 000 000 issn 0000-0000 eissn 0000-0000 00 0000 0001 INFORMS doi 10.187/xxxx.0000.0000 c 0000 INFORMS Multistage Robust Mixed Integer Optimization with Adaptive Partitions

More information

A Simplified Lagrangian Method for the Solution of Non-linear Programming Problem

A Simplified Lagrangian Method for the Solution of Non-linear Programming Problem Chapter 7 A Simplified Lagrangian Method for the Solution of Non-linear Programming Problem 7.1 Introduction The mathematical modelling for various real world problems are formulated as nonlinear programming

More information

Mixed Integer Linear Programming Formulation for Chance Constrained Mathematical Programs with Equilibrium Constraints

Mixed Integer Linear Programming Formulation for Chance Constrained Mathematical Programs with Equilibrium Constraints Mixed Integer Linear Programming Formulation for Chance Constrained Mathematical Programs with Equilibrium Constraints ayed A. adat and Lingling Fan University of outh Florida, email: linglingfan@usf.edu

More information

QUADRATIC MAJORIZATION 1. INTRODUCTION

QUADRATIC MAJORIZATION 1. INTRODUCTION QUADRATIC MAJORIZATION JAN DE LEEUW 1. INTRODUCTION Majorization methods are used extensively to solve complicated multivariate optimizaton problems. We refer to de Leeuw [1994]; Heiser [1995]; Lange et

More information

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness. CS/ECE/ISyE 524 Introduction to Optimization Spring 2016 17 14. Duality ˆ Upper and lower bounds ˆ General duality ˆ Constraint qualifications ˆ Counterexample ˆ Complementary slackness ˆ Examples ˆ Sensitivity

More information

Math 273a: Optimization Subgradients of convex functions

Math 273a: Optimization Subgradients of convex functions Math 273a: Optimization Subgradients of convex functions Made by: Damek Davis Edited by Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com 1 / 42 Subgradients Assumptions

More information

Optimality Results in Inventory-Pricing Control: An Alternate Approach

Optimality Results in Inventory-Pricing Control: An Alternate Approach Optimality Results in Inventory-Pricing Control: An Alternate Approach Woonghee Tim Huh, Columbia University Ganesh Janakiraman, New York University May 9, 2006 Abstract We study a stationary, single-stage

More information

Mathematical Preliminaries for Microeconomics: Exercises

Mathematical Preliminaries for Microeconomics: Exercises Mathematical Preliminaries for Microeconomics: Exercises Igor Letina 1 Universität Zürich Fall 2013 1 Based on exercises by Dennis Gärtner, Andreas Hefti and Nick Netzer. How to prove A B Direct proof

More information

Near-Potential Games: Geometry and Dynamics

Near-Potential Games: Geometry and Dynamics Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo January 29, 2012 Abstract Potential games are a special class of games for which many adaptive user dynamics

More information

A Parametric Simplex Algorithm for Linear Vector Optimization Problems

A Parametric Simplex Algorithm for Linear Vector Optimization Problems A Parametric Simplex Algorithm for Linear Vector Optimization Problems Birgit Rudloff Firdevs Ulus Robert Vanderbei July 9, 2015 Abstract In this paper, a parametric simplex algorithm for solving linear

More information

16.4 Multiattribute Utility Functions

16.4 Multiattribute Utility Functions 285 Normalized utilities The scale of utilities reaches from the best possible prize u to the worst possible catastrophe u Normalized utilities use a scale with u = 0 and u = 1 Utilities of intermediate

More information

Column Generation for Extended Formulations

Column Generation for Extended Formulations 1 / 28 Column Generation for Extended Formulations Ruslan Sadykov 1 François Vanderbeck 2,1 1 INRIA Bordeaux Sud-Ouest, France 2 University Bordeaux I, France ISMP 2012 Berlin, August 23 2 / 28 Contents

More information

Course 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016

Course 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016 Course 16:198:520: Introduction To Artificial Intelligence Lecture 13 Decision Making Abdeslam Boularias Wednesday, December 7, 2016 1 / 45 Overview We consider probabilistic temporal models where the

More information

A Shadow Simplex Method for Infinite Linear Programs

A Shadow Simplex Method for Infinite Linear Programs A Shadow Simplex Method for Infinite Linear Programs Archis Ghate The University of Washington Seattle, WA 98195 Dushyant Sharma The University of Michigan Ann Arbor, MI 48109 May 25, 2009 Robert L. Smith

More information

Benders Decomposition

Benders Decomposition Benders Decomposition Yuping Huang, Dr. Qipeng Phil Zheng Department of Industrial and Management Systems Engineering West Virginia University IENG 593G Nonlinear Programg, Spring 2012 Yuping Huang (IMSE@WVU)

More information

Basing Decisions on Sentences in Decision Diagrams

Basing Decisions on Sentences in Decision Diagrams Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence Basing Decisions on Sentences in Decision Diagrams Yexiang Xue Department of Computer Science Cornell University yexiang@cs.cornell.edu

More information

Risk-Averse Dynamic Optimization. Andrzej Ruszczyński. Research supported by the NSF award CMMI

Risk-Averse Dynamic Optimization. Andrzej Ruszczyński. Research supported by the NSF award CMMI Research supported by the NSF award CMMI-0965689 Outline 1 Risk-Averse Preferences 2 Coherent Risk Measures 3 Dynamic Risk Measurement 4 Markov Risk Measures 5 Risk-Averse Control Problems 6 Solution Methods

More information

Final Exam - Math Camp August 27, 2014

Final Exam - Math Camp August 27, 2014 Final Exam - Math Camp August 27, 2014 You will have three hours to complete this exam. Please write your solution to question one in blue book 1 and your solutions to the subsequent questions in blue

More information

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden 1 Selecting Efficient Correlated Equilibria Through Distributed Learning Jason R. Marden Abstract A learning rule is completely uncoupled if each player s behavior is conditioned only on his own realized

More information

Distributed power allocation for D2D communications underlaying/overlaying OFDMA cellular networks

Distributed power allocation for D2D communications underlaying/overlaying OFDMA cellular networks Distributed power allocation for D2D communications underlaying/overlaying OFDMA cellular networks Marco Moretti, Andrea Abrardo Dipartimento di Ingegneria dell Informazione, University of Pisa, Italy

More information