Decomposition with Branch-and-Cut Approaches for Two Stage Stochastic Mixed-Integer Programming

Similar documents
Decomposition with Branch-and-Cut Approaches for Two-Stage Stochastic Mixed-Integer Programming

Decomposition Algorithms with Parametric Gomory Cuts for Two-Stage Stochastic Integer Programs

Fenchel Decomposition for Stochastic Mixed-Integer Programming

Generation and Representation of Piecewise Polyhedral Value Functions

Disjunctive Decomposition for Two-Stage Stochastic Mixed-Binary Programs with GUB Constraints

Computations with Disjunctive Cuts for Two-Stage Stochastic Mixed 0-1 Integer Programs

Disjunctive Decomposition for Two-Stage Stochastic Mixed-Binary Programs with Random Recourse

Stochastic Integer Programming An Algorithmic Perspective

Change in the State of the Art of (Mixed) Integer Programming

Network Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini

Lift-and-Project Inequalities

Lagrangian Relaxation in MIP

Solution Methods for Stochastic Programs

An Adaptive Partition-based Approach for Solving Two-stage Stochastic Programs with Fixed Recourse

Strengthened Benders Cuts for Stochastic Integer Programs with Continuous Recourse

Stochastic Integer Programming

Valid Inequalities and Restrictions for Stochastic Programming Problems with First Order Stochastic Dominance Constraints

Part 4. Decomposition Algorithms

arxiv: v1 [cs.cc] 5 Dec 2018

A Benders Algorithm for Two-Stage Stochastic Optimization Problems With Mixed Integer Recourse

LP Relaxations of Mixed Integer Programs

Integer Programming. Wolfram Wiesemann. December 6, 2007

Can Li a, Ignacio E. Grossmann a,

Decomposition Algorithms for Two-Stage Distributionally Robust Mixed Binary Programs

Reduced rst-level representations via the reformulation-linearization technique: results, counterexamples, and computations

Can Li a, Ignacio E. Grossmann a,

Disconnecting Networks via Node Deletions

Computational Integer Programming. Lecture 2: Modeling and Formulation. Dr. Ted Ralphs

Totally Unimodular Stochastic Programs

Integer Programming ISE 418. Lecture 13. Dr. Ted Ralphs

On Solving Two-Stage Distributionally Robust Disjunctive Programs with a General Ambiguity Set

Outline. Relaxation. Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING. 1. Lagrangian Relaxation. Lecture 12 Single Machine Models, Column Generation

Section Notes 8. Integer Programming II. Applied Math 121. Week of April 5, expand your knowledge of big M s and logical constraints.

Branch-and-cut Approaches for Chance-constrained Formulations of Reliable Network Design Problems

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Bilevel Integer Linear Programming

Discrete Two-Stage Stochastic Mixed-Integer Programs with Applications to Airline Fleet Assignment and Workforce Planning Problems

Decomposition Algorithms for Two-Stage Chance-Constrained Programs

Stochastic Mixed-Integer Programming

Operations Research Lecture 6: Integer Programming

Bilevel Integer Optimization: Theory and Algorithms

Integer Programming ISE 418. Lecture 8. Dr. Ted Ralphs

Breaking the Rectangle Bound Barrier against Formula Size Lower Bounds

Column Generation for Extended Formulations

Computational Integer Programming Universidad de los Andes. Lecture 1. Dr. Ted Ralphs

DECOMPOSITION ALGORITHMS FOR TWO-STAGE STOCHASTIC INTEGER PROGRAMMING

BCOL RESEARCH REPORT 07.04

Two-stage stochastic (and distributionally robust) p-order conic mixed integer programs: Tight second stage formulations

Cutting planes from extended LP formulations

arxiv:math/ v1 [math.co] 23 May 2000

Solving Bilevel Mixed Integer Program by Reformulations and Decomposition

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

Multiobjective Mixed-Integer Stackelberg Games

The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 44

The use of predicates to state constraints in Constraint Satisfaction is explained. If, as an

ON MIXING SETS ARISING IN CHANCE-CONSTRAINED PROGRAMMING

On the Approximate Linear Programming Approach for Network Revenue Management Problems

Benders' Method Paul A Jensen

The L-Shaped Method. Operations Research. Anthony Papavasiliou 1 / 38

Applications. Stephen J. Stoyan, Maged M. Dessouky*, and Xiaoqing Wang

Reformulation and Decomposition of Integer Programs

On a Level-Set Characterization of the Value Function of an Integer Program and Its Application to Stochastic Programming

Multicommodity Flows and Column Generation

Integer Programming Duality

3.7 Cutting plane methods

Mixed-integer linear representability, disjunctions, and variable elimination

Bilevel Integer Programming

Improvements to Benders' decomposition: systematic classification and performance comparison in a Transmission Expansion Planning problem

Decision Procedures An Algorithmic Point of View

Decision Diagram Relaxations for Integer Programming

Benders Decomposition

Reduced RLT Representations for Nonconvex Polynomial Programming Problems

Graph Coloring Inequalities from All-different Systems

Lagrange Relaxation: Introduction and Applications

Extended Formulations, Lagrangian Relaxation, & Column Generation: tackling large scale applications

Polyhedral Approach to Integer Linear Programming. Tepper School of Business Carnegie Mellon University, Pittsburgh

CO 250 Final Exam Guide

Operations Research Letters

is called an integer programming (IP) problem. model is called a mixed integer programming (MIP)

K-Adaptability in Two-Stage Mixed-Integer Robust Optimization

A Note on Representations of Linear Inequalities in Non-Convex Mixed-Integer Quadratic Programs

Split Rank of Triangle and Quadrilateral Inequalities

Scenario Grouping and Decomposition Algorithms for Chance-constrained Programs

to work with) can be solved by solving their LP relaxations with the Simplex method I Cutting plane algorithms, e.g., Gomory s fractional cutting

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones

An improved L-shaped method for two-stage convex 0-1 mixed integer nonlinear stochastic programs

Inference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples

Stochastic Dual Dynamic Integer Programming

Linear Programming in Matrix Form

Integer Programming ISE 418. Lecture 13b. Dr. Ted Ralphs

An Integer Cutting-Plane Procedure for the Dantzig-Wolfe Decomposition: Theory

3.10 Lagrangian relaxation

Integer Programming Part II

c 2000 Society for Industrial and Applied Mathematics

A Tighter Variant of Jensen s Lower Bound for Stochastic Programs and Separable Approximations to Recourse Functions

Column Generation in Integer Programming with Applications in Multicriteria Optimization

Critical Reading of Optimization Methods for Logical Inference [1]

A Principled Approach to Mixed Integer/Linear Problem Formulation

Lecture 1. Stochastic Optimization: Introduction. January 8, 2018

The Simplex Method: An Example

Transcription:

Decomposition with Branch-and-Cut Approaches for Two Stage Stochastic Mixed-Integer Programming by Suvrajeet Sen SIE Department, University of Arizona, Tucson, AZ 85721 and Hanif D. Sherali ISE Department, VPI&SU Blacksburg, VA 24061 August 2002 Keywords: Decomposition, Branch-and-Cut, Divide-and-Conquer, Integer Programming Abstract Decomposition has proved to be one of the more effective tools for the solution of large scale problems, especially those arising in stochastic programming. A decomposition method with wide applicability is Benders decomposition, and has been applied to both stochastic programming, as well as integer programming problems. However, this method of decomposition relies on convexity of the value function of linear programming subproblems. This paper is devoted to a class of problems in which the second stage subproblem(s) may impose integer restrictions on some variables. The value function of such integer subproblem(s) is not convex, and new approaches must be designed. In this paper, we discuss alternative decomposition methods in which the second stage integer subproblems are solved using branch-and-cut methods. One of the main advantages of our decomposition scheme is that Stochastic Mixed-Integer Programming (SMIP) problems can be solved by dividing a large problem into smaller MIP subproblems which can be solved in parallel. This paper lays the foundation for such decomposition methods for two-stage stochastic integer programs.

2 Decomposition with Branch-and-Cut Approaches for Two Stage Stochastic Integer Programming 1. Introduction This paper continues the line of work initiated in two earlier papers dealing with convexification of stochastic integer programming problems. One of the papers (Sherali and Fraticelli [2000]) drew upon results from the theory of reformulation-linearization techniques (Sherali and Adams [1990]) to develop a sequence of relaxations for 0-1 mixed-integer second stage problems, and these were incorporated within a Benders decomposition algorithm for a two stage stochastic program having binary first-stage variables. The other motivating paper is that of Sen and Higle [2000] in which methods of disjunctive programming were used to develop set convexifications in a sequential manner. As one might expect, there are definite connections between these approaches, although their development, as well as the algorithmic schemes, are different. The current paper is intended to enhance these approaches by adding a major piece of the algorithmic puzzle. In essence, this paper revolves around the design of branch-and-cut (BAC) algorithms for Stochastic Mixed-Integer Programming (SMIP) problems. BAC algorithms provide one of the more successful approaches for deterministic mized-integer programming (MIP). However, their extension to SMIP problems is far from obvious. The main challenges arise from the need to combine decomposition-based methods that have worked well in the Stochastic Programming arena (see Ruszczyński [1997]) with BAC methods that have worked well for MIP problems. We begin this paper by exploring a method that combines decomposition with a BAC method in a deterministic setting. This (deterministic) setting allows some simplifications that permit a few algorithmic liberties. The development not only provides the first taste of the algorithmic issues for decomposition-based BAC (D-BAC) algorithms, but also sets the stage for a BAC algorithm in the stochastic programming setting. The D-BAC algorithm will guide us towards Disjunctive Decomposition (D 2 ) algorithms to accommodate SMIP problems in which both first and second stages have integer variables. The new method will be designated as the D 2 -BAC algorithm. These extensions provide powerful additions to the algorithmic toolkit for SMIP.

3 2. Decomposing Deterministic Integer Programs This section is devoted mainly to a deterministic binary mixed-integer programming problem. This problem also provides us the vehicle that will be used to set the stage for designing algorithms for the SMIP problem. Consider the problem Min c x + g y (1.1) T x + W y r (1.2) x X B, y Y B (1.3) where B (B ) denotes the set of binary (mixed 0-1) vectors, and where X and Y are nonempty polyhedra. The set X is represented as {x Ax b, x 0}, where the constraints x 1 are included in the linear inequalities Ax b. Similarly, we assume that Y is bounded and represent it as follows Y = {y Dy f, y 0, y j binary for j J 2 }. (1.4) According to our notation, inequalities of the form y j 1 are also included in the constraints Dy f, and the index set J 2 provides the subset of the second stage variables that are restricted to be binary. 2.1 A Decomposition-based Cutting Plane Approach This subsection is based on ideas presented in Sherali and Fraticelli [2000]. At iteration k, let x k be given, and consider the following MIP in which x are treated as continuous variables, and y are treated as (mixed) binary variables. Min M x j + (1 x j ) + g y (2.1) {j x k j =0} {j x k j =1} T x + W y r (2.2) 0 x 1, y Y B. (2.3) Note that M is assumed to be sufficiently large, so that under (2.3), which includes the requirement that 0 x 1, the above problem yields x = x k as an optimal solution of (2). Hence, x is not quite a variable in (2), but is essentially fixed, thereby providing

a convenient conceptual basis for developing valid Benders cuts. The derivation of these cuts is predicated on the fact that (1) is equivalent to the problem obtained by replacing its constraints by the convex hull of (2.2), (2.3), along with the restrictions that x X (Sherali and Fraticelli [2000]). Now, suppose that (2) is solved using a cutting plane method, and that one obtains a partial polyhedral representation of the mixed-integer feasible set through additional constraints of the form G k x + H k y h k. These constraints, together with the linear inequalities in (2.2) and the structural inequalities describing Y in (1.4), yield the following LP representation of (2). Min M x j + (1 x j ) + g y (3.1) {j x k j =0} {j x k j =1} T x + W y r (3.2) G k x + H k y h k (3.3) Dy f (3.4) 0 x 1, y 0. (3.5) From a computational point of view, we should note that the use of M is only an artifice implying that the value of x remains fixed at x k throughout any iterative scheme to solve (3). However, when the cuts (3.3) are derived during the course of the algorithm for solving (2), we must ensure that the resulting inequalities are valid with respect to all (x, y) satisfying (2.2) and (2.3). It is for this reason that we choose to think of x as variables in (2). In order to represent (3) in a compact form, we append the sub-matrices associated with the x variables in rows (3.2)-(3.4) to form the matrix T k. Similarly, the sub-matrices associated with the y variables in (3.2)-(3.4) are taken to form a matrix W k, and the righthand side vector of rows (3.2)-(3.4) is recorded as the vector r k. Hence the above LP can be rewritten in a compact form as Min M x j + (1 x j ) + g y (4.1) {j x k j =0} {j x k j =1} T k x + W k y r k (4.2) 0 x 1, y 0 (4.3) 4 Letting θ k denote optimal dual multipliers associated with (4.2), Sherali and Fraticelli [2000] derive the following Benders cut that can be added to the master program,

where η represents the second stage value function. Then the master program has the following familiar form: η θ k (r k T k x). (5) Min x X B {c x + η η θt (r t T t x), t = 1,..., k}. It is important to note that since the inequalities of (3.3) derived at iteration k are valid in the space of both variables x as well as y, these inequalities can be used in all subsequent iterations. Hence, the subproblem solved during iteration k + 1 can begin by using the inequalities obtained through the first k iterations, and more globally valid inequalities may be added during iteration k + 1 and beyond. Finally, if the cuts (3.3) used for solving (2) have the property that they ultimately construct the necessary facets of the convex hull of the set described by (2.2) and (2.3), then finite convergence of the resulting algorithm is immediate. There are various families of valid inequalities that satisfy this property (see Sherali and Adams [1990, 1994], Lovasz and Schrijver [1991], and Balas, Ceria and Cornuéjols [1993]). 2.2 A Decomposition-based BAC Approach (D-BAC) The results of this subsection extend the approach of Sherali and Fraticelli [2000] to solving (1) for the case in which the second stage mixed-binary optimization problem is solved using BAC. Here, the term decomposition is to be interpreted in the sense that the integer restrictions on the x and y variables are treated separately. That is, the variables may appear in both the master as well as subproblems simultaneously; however, the integer restrictions on both these variables (x and y) are not imposed simultaneously in either the master or subproblems. Suppose that in a BAC approach for solving (2), we obtain an optimal solution at some node, denoted, of the branch-and-bound tree. Let us denote the index sets of fixed variables at node by J2 = {j J 2 y j 0}, and J 2 + = {j J 2 y j 1}. Then the LP problem for node in a notation analogous to (4) is Min M x j + (1 x j ) + g y (6.1) {j x k j =0} {j x k j =1} T k x + W k y r k (6.2) 0 x 1, y 0 (6.3) y j 0, j J 2, y j 1, j J + 2. (6.4) 5

Again, we emphasize that the cuts included in (6.2) must be valid for all (x, y) satisfying (2.2) and (2.3). Furthermore, as before, the solution of an LP at any node of the tree should not treat x as a variable; instead x is fixed equal to x k. 6 Proposition 1. Let θ k denote a vector of optimal dual multipliers associated with (6.2). Moreover, let ψkj and ψ+ kj denote optimal dual multipliers for the bounding constraints associated with J2 and J 2 +, respectively, in (6.4). Then the following inequality η θk (r k T k x) + ψ + kj y j ψ kj y j (7) j J + 2 j J 2 provides a lower bounding function on g y over (x, y) feasible to (2). Proof. Since θ k 0, (6.2) implies that 0 θ k [r k T k x W k y]. Hence, η g y + θ k [r k T k x W k y] = θ k [r k T k x] + [g θ k W k]y. (8) Furthermore θ k is dual feasible for (6), and y 0. Hence, [g θk W k]y ψ kj + y j ψ kj y j. j J + 2 Substituting this inequality in (8) yields the result. j J 2 Although we have presented Proposition 1 in such a manner as to add only one cut, it is possible to add more cuts to obtain a stronger approximation. Note that by incorporating an artificial variable column (with arbitarily high costs) in the subproblem, we can assume that every node q of the BAC tree is associated with a feasible LP, and that nodes are fathomed when their LP lower bounds exceed the best upper bound obtained. If θ kq, ψ + kjq, ψ kjq denote the dual multipliers with each fathomed node q, then one can add as many cuts as there are nodes in the BAC tree. That is, the following inequalities may be added to the master program. η θkq (r k T k x) + ψ kjq + y j ψkjq y j, q. (9) j J + 2q j J 2q The arguments supporting the validity of (7) also support the validity of (9). Proposition 2. Consider the following partial relaxed master program in which the binary restrictions on the y variables are enforced. Min{c x + η (η, x, y) satisfy (9), y j binary j J 2 } (10)

(Note that (9) requires the inequalities based on all nodes of the BAC tree to be included.) Then, for x = x k (fixed), the optimal value of η in (10) is the same as the value of the subproblem (2). 7 Proof. Note that all inequalities indexed by q are valid lower bounding inequalities, and moreover, the inequality for the node that yields the optimal value is included in the description of (10). Furthermore, for any inequality q in (9), when we fix x = x k along with y j = 1 j J 2q + and y j = 0 j J2q, the right-hand side of (9) yields (by duality) the value of the corresponding nodal subproblem associated with (6) in the enumeration tree. Since any binary solution (ŷ j, j J 2 ) relates to precisely one terminal node q in this enumeration tree for which ŷ j = 1, j J 2q +, and ŷ j = 0, j J2q, the result follows. Notwithstanding this result, one may wish to relax the requirement that the second stage binary variables (in y) be restricted to be integer within the master program. This would result in a more tractable, albeit weaker, master program. Indeed, this can be achieved by incorporating the following additional valid inequality within the master program whenever one solves the subproblem to optimality (as we have in this section). Let I k = {i x k i = 1}, Z k = {1,..., n} I k. Next define the linear function δ k (x) = I k [ i I k x i i Z k x i ]. (11) It is easily seen that when x = x k (assumed binary), δ k (x) = 0; whereas, for all other binary x x k, at least one of the variables must switch states. Hence for x x k, we have [ i I k x i i Z k x i ] I k 1, i.e. δ k (x) 1. Next suppose that a lower bound on the second stage, denoted l, is available. Furthermore, let η(x k ) denote the optimal value of the subproblem, given x k. Then the following inequality is valid and may be included whenever η(x k ) is available. η η(x k ) δ k (x)[η(x k ) l]. (12) This is essentially the optimality cut of Laporte and Louveaux [1993]. To verify its validity, first observe that when x = x k, the second term in (12) vanishes, and hence the

master program recovers the value of the corresponding subproblem. On the other hand, if x x k, then, δ k (x)[η(x k ) l] [η(x k ) l]. Hence, for all x x k, the right-hand side of (12) obeys η(x k ) δ k (x)[η(x k ) l] η(x k ) η(x k ) + l = l. Hence imposing (12) does not delete any viable first stage solution. Moreover, since (12) itself asserts that for x = x k, η η(x k ) in the master program, we need not explicitly impose either the integrality restrictions on any of the binary components of y (as in Proposition 2), or the terminal node inequalities (9). Thus, while there is a price to be paid in solving the second stage problem to optimality, including (12) in addition to (9) may provide stronger relaxations in the master program. Furthermore this construct obtains finite convergence of Benders decomposition as in Sherali and Fraticelli [2000]. Remark 1. While implementing (9), if we wish to dispense with the inclusion of y in the master program, then we can derive a projection that provides an inequality for which the coefficients of y vanish. To do so, one would have to select nonnegative multipliers on y j 0, y j 1, j J 2, together with a multiplier on (9) in such a way that the aggregated coefficient for y is zero. Remark 2. In the foregoing analysis, we have assumed that by adding appropriate artificial variables, each nodal subproblem in the tree is feasible. It is instructive to note that (11) provides the facility of generating feasibility cuts in the absence of the above assumption. Specifically, if some node yields an infeasible restriction, then in lieu of (9), one may add the constraint (1 y j ) + y j 1 δ k (x). (13) 8 j J + 2q j J 2q To see that (13) is valid, note that whenever x = x k, δ(x k ) = 0, and then (13) asserts that at least one of the binary variables for y j, j J 2, must assume a value different from the restrictions imposed at node q of the branch-and-bound tree. On the other hand, when x x k, we have δ(x k ) 1, and then (13) is simply redundant. Moreover, by incorporating (13) within (9) for all infeasible nodes q, the assertion of Proposition 2 continues to hold. 2.3 An Illustration of the D-BAC Algorithm Consider the following problem

9 Min x 1 2y 1 + 4y 2 4x 1 3y 1 + y 2 6 (x 1, y 1 ) binary, y 2 0. We will refer to x 1 as the first stage decision variable, and y will be designated a second stage decision vector. It is easily seen that for binary values of y 1, and nonnegative y 2, a lower bound on the expression 2y 1 + 4y 2 is 2; that is η 2. Hence, we may initialize the process with the following master program. Min x 1 + η x 1 binary. η 2 The optimal value of the above problem is 3, and this value also provides a lower bound on the optimal value of the original problem. The optimal solution to this approximation is x 1 = 1, η = 2. Using x 1 = 1, we formulate the second stage problem as follows. Min 2y 1 + 4y 2 3y 1 + y 2 2 y 1 binary, y 2 0. The LP relaxation provides a fractional solution (2/3, 0). Suppose that we solve this problem using a branch-and-bound scheme. The nodes of the tree are analyzed below. Node 1(y 1 = 1): We solve the following LP: Min 2y 1 + 4y 2 3y 1 + y 2 2 y 1 1 y 1 1 y 1, y 2 0. Relating this problem to (6), note that the first two rows form the matrix W k of (6.2), and the third constraint above is a lower bound of the form in (6.4). Solving this LP yields (y 1, y 2 ) = (1, 1) and dual multipliers θ = (4, 0), and ψ 1 + = 10. Then (9) yields the following approximation of the second stage value function. η 4( 6 + 4x 1 ) + 10y 1.

10 Node 2 (y 1 = 0): The LP for this node is given by Min 2y 1 + 4y 2 3y 1 + y 2 2 y 1 1 y 1 0 y 1, y 2 0. Upon solving this LP, we have (y 1, y 2 ) = (0, 0), and the dual multipliers are θ = (0, 0), and ψ 1 = 2. Applying (9), we obtain η 2y 1. The upper bound for the original problem at this iteration is 1 (obtained at node 2). Moreover, an inequality of type (12) may be obtained by letting x k correspond to x 1 = 1, η(x k ) = 0, l = 2, and using (11) to define δ k (x) = 1 x 1. The resulting inequality is η 2(1 x 1 ). (14) Using inequalities of type (9) and (12) (and omitting the dominated inequality η 2), the updated master program is as follows. Min x 1 + η 2x 1 + η 2 (14) 16x 1 + η 10y 1 24 η + 2y 1 0 x 1 binary, 0 y 1 1. Having updated the master program, we have now completed one iteration. At this point in the algorithm, the upper bound is 1, and the lower bound is 3. Starting the next iteration, we solve the updated master problem. The solution to this problem yields (x 1, η, y 1 ) = (0, 2, 1), and the updated lower bound is 2. Fixing x 1 = 0, the LP relaxation of the second stage problem yields (y 1, y 2 ) = (1, 0). The resulting upper bound is therefore 2, which equals the lower bound. Hence the method stops with an optimal first stage solution, which is x 1 = 0. Note that because of our use of a decomposition scheme, no individual problem (master or subproblem) has more than one integer variable, although the original problem has two

integer variables. Thus, similar to Benders decomposition for continuous second stage decisions, our D-BAC algorithm solves MIP problems of the above type by solving a sequence of easier mixed-integer programs. 11 3. Decomposition for Stochastic Mixed-Integer Programming We now turn to a study of SMIPs in which a subset of both first and second stage variables are required to satisfy integer restrictions. We state the problem as follows. Min x X B c x + E[f(x, ω)], where X, and B are sets defined in (1), ω is a random variable defined on a probability space (Ω, A, P), and for any realization ω of ω, f(x, ω) = Min g y W y r(ω) T (ω)x y 0, y j binary, j J 2. Within the stochastic programming literature, a realization of ω is known as a scenario. As such, the second stage problem is often referred to as a scenario subproblem. The method presented here is applicable to a more general class of problems than the one stated above. In particular, our approach is applicable to cases where the second stage problem includes general integers (not simply 0-1), and the first stage decisions are required to be extreme points of X. Of course, the class of problems stated above satisfy these requirements. The choice to restrict our discussion to 0-1 problems is motivated by the need for clarity and consistency with the previous section. We begin this section by examining what it takes to make the transition from deterministic MIPs and stochastic MIPs. Simply put, it is a matter of scalability. The manner in which an algorithm accommodates the presence of a large number of alternative scenarios determines its effectiveness for solving SMIP. While this presentation will assume that there are only finitely many scenarios in the SMIP problem, realistic models often result in a relatively large number of scenarios. Recognizing this, let us first discuss the merits of obtaining an optimal solution for each scenario subproblem for a given first stage solution x. Since these subproblems are generally NP-hard, the decomposition method may get bogged down in attempts to solve subproblems, even while the particular first stage decision x may not be near a small neighborhood of an optimal solution. In essence, our view

is that the algorithm should allow partial (i.e. suboptimal) solves of the MIP subproblems, but as iterations proceed, the method should learn enough about the structure of the MIP subproblems, so that ultimately, the partial solves begin to yield optimal solutions. This can be accomplished via a sequential convexification process in which only a small subset of facets are generated during any iteration. Scalability of an algorithm is not only determined by the computational effort per iteration, but also by its memory requirements reflected through the size of each optimization problem solved during the algorithmic process. This is one of the main motivations for the C 3 theorem in Sen and Higle [2000]. Because of the common cut coefficients, approximations of the second stage can be generated without storing cuts separately for each scenario. In contrast, the cuts presented in 2.1 are intended to be recorded explicitly for each scenario, and although they may be reused for different first stage decisions, there is the potential of dealing with a very memory-intensive algorithm. In making the transition from deterministic to stochastic integer programming problems, care must be taken to design algorithms that acknowledge the potential for a large number of scenarios. We note that the BAC approach for subproblems (see 2.2) can help reduce the number of valid inequalities recorded for each scenario. However, the augmented cuts of 2.2 introduce other complications for SIPs. For instance, a naive extension of the method of 2.2 to SIP problems would lead us to include as many second stage variables as there are scenarios in the SIP. Notwithstanding the fact that these are treated as continuous variables in the master problem, the size of the resulting MIP can quickly get out of hand for all but the most modest SMIP problems. We therefore adopt the D 2 approach whereby the second stage variables are projected out of the first stage master program as in Sen and Higle [2000]. The remainder of this section is devoted to the incorporation of BAC methods for the second stage problem in a D 2 algorithm. Consider a partial branch-and-bound tree generated during a partial solve of the second stage problem. Let Q(ω) denote the set of nodes of the tree that have been explored for the subproblem associated with scenario ω. As in 2.2, we will assume that all nodes of the branch-and-bound tree are associated with a feasible LP relaxation, and that nodes are fathomed when the LP lower bound exceeds the best available upper bound. This may be accomplished by introducing artificial variables, if necessary. Our strategy revolves around using the dual problem associated with the LP relaxation (one for each node), and then stating a disjunction that will provide a valid inequality for the first stage problem. When 12

compared with the inequalities obtained previously in (9), these cuts involve only the first stage variables (x). (The reader may also refer to Remark 1 for projecting (9) onto the space of first stage variables x.) In the following, we use k as the index of iterations, where at each iteartion, the second stage subproblems are solved to some degree of accuracy. For any node q Q(ω), let z ql (ω) and z qh (ω) denote vectors whose elements are used to define lower and upper bounds, respectively, on the second stage (integer) variables. In some cases, an element (z qh ) j may be +, and in this case, the associated constraint may be ignored, implying that the associated dual multiplier is fixed at 0. In any event, the LP relaxation for node q may be written as Min g y W k y r k (ω) T k (ω)x y 0 y z ql (ω) y z qh (ω), and, the corresponding dual LP is Max θ q (ω) [r k (ω) T k (ω)x] + ψ ql (ω) z ql (ω) ψ qh (ω) z qh (ω) θ q (ω) W k + ψ ql (ω) ψ qh (ω) θ q (ω) 0, ψ ql (ω) 0, ψ qh (ω) 0, g where the vectors ψ ql (ω), and ψ qh (ω) are appropriately dimensioned vectors. We should remind the reader that the above form of the relaxation assumes that the cut coefficients in the matrix W k do not depend on the scenario ω. This follows from the common cut coefficients (C 3 ) theorem which allows the disjunctive decomposition method to share cut coefficients among various second stage scenarios (Sen and Higle [2000]). This theorem suggests that the convex hull of mixed-integer (binary) solutions in the second stage (for any outcome ω) can be obtained by adding a sequence of cuts of the form π y π 0 (x; ω). Note that the cut coefficients π are common, but the right-hand side is a function of both x and ω. It turns out that for any given ω, the function π 0 is a piecewise linear concave function of x that can be convexified using disjunctive programming (Sen and Higle [2000]). The resulting convexified right hand side functions can be represented as affine functions, denoted π c (x; ω). We assume that the rows of r k (ω) and T k (ω) represent coefficients of the affine function π c (x; ω). We now turn our attention to approximating the value function of the second stage MIP. As noted in Blair and Jeroslow [1982], and subsequently by Blair [1995], IP and MIP 13

value functions are complicated objects; however, the branch-and-bound tree, together with the LP relaxations at these nodes, provide important information that can be used to approximate MIP value functions. The main observation that we use in this development is that the branch-and-bound tree embodies a disjunction, and when coupled with piecewise linear value functions of the LP relaxations of each node, we obtain a disjunctive description of an approximation to the MIP value function. By using the disjunctive cut principle, we will then obtain linear inequalities (cuts) that can be used to build value function approximations. In order to do so, we assume that we have a lower bound l such that f(x, ω) l (almost surely) for all x. Without loss of generality, this bound may be assumed to be 0. Consider a node q Q(ω) and let (θq k (ω), ψql k (ω), ψk qh (ω)) denote optimal dual multipliers for node q. Then a lower bounding function may be obtained by requiring that x X and that the following disjunction holds. η θ k q (ω) [r k (ω) T k (ω)x] + ψ k ql (ω) z ql (ω) ψ k qh (ω) z qh (ω) for at least one q Q(ω). (15) Note that each inequality in (15) corresponds to a second stage value function approximation that is valid only when the restrictions (on the y-variables) associated with node q Q(ω) hold true. Since any optimal solution of the second stage must be associated with at least one of the nodes q Q(ω), the disjunction (15) is valid. By assumption, we have η 0. Hence, x X and (15) leads to the following disjunction: Π k (ω) = {(η, x) q Q(ω) E k q (ω)}, 14 where with, and E k q (ω) = {(η, x) η ν k q (ω) γ k q (ω) x, Ax b, x 0, η 0}, ν k q (ω) = θ k q (ω) r k (ω) + ψ k ql (ω) z ql (ω) ψ k qh (ω) z qh (ω), γ k q (ω) = T k (ω) θ k q (ω). Following the development in Sen and Higle [2000], the epi-reverse polar of this set, which we denote by Π k (ω), is as follows. Π k (ω) = {σ 0(ω) R, σ(ω), R n 1, ζ(ω) R q Q(ω), τ q (ω) 0, τ 0q (ω) R + s.t.

15 σ 0 (ω) τ 0q (ω) τ 0q (ω) = 1 q Q(ω) q Q(ω) σ j (ω) τ q (ω) A j + τ 0q (ω) γ k qj(ω) q Q(ω), j = 1,..., n 1 ζ(ω) τ q (ω) b + τ 0q (ω) ν k q (ω) q Q(ω) τ q (ω) 0, τ 0q (ω) 0 q Q(ω)}. (16) Hence the cut is obtained by choosing non-negative multipliers τ k 0q (ω), τ k q (ω) for all q, and then using the Min and Max operations as follows: σ0(ω) k = Max τ k q 0q(ω) σj k (ω) = Max{τ k q q (ω) A j + τ0q(ω) γ k qj(ω)} k j ζ(ω) = Min{[τ k q q (ω)] b + τ0q(ω) ν k q k (ω)}. The disjunctive cut for any outcome ω is then obtained as follows. σ k 0(ω)η + j σ j (ω)x j ζ(ω), where the conditions in (16) imply that σ 0 (ω) Max q τ 0q (ω) > 0. We assume that the cuts used by the algorithm are based on using extreme points of (16). Hence, the epireverse polar only allows those facets (of the convex hull of Π k (ω)) that have a positive coefficient for the variable η. Because of its connections with the disjunctive decomposition (D 2 ) algorithm of Sen and Higle [2000], we refer to the current method as the D 2 -BAC algorithm. The optimality cut to be included in the first stage master in iteration k is given by η E[ ζk ( ω) σ k 0 ( ω)] E[σk ( ω) σ k 0 ( ω)] x. (17.k) It is obvious that one can also devise a multi-cut method in which the above optimality cut is disaggregated into several inequalities (e.g. Birge and Louveaux [1997]). Remark 3. It is important to draw the distinction between the disjunctive cuts that have appeared in the MIP literature, and the cut proposed in (17.k). Disjunctive cuts in the MIP literature are intended to provide tight relaxations of the set of feasible integer points. The D 2 algorithm with set convexification (D 2 -SC, Sen and Higle [2000], Sen, Higle and

Ntaimo [2002]) provides a sequential approach for tightening second stage linear relaxations. Such sequential convexification allows us to carry information from one iteration to the next, thus avoiding the need to re-start the convexification process from scratch. Nevertheless, the goal of D 2 -SC remains one in which linear relaxations of the subproblems are tightened using valid inequalities. In contrast, the D 2 -BAC approach provides a completely novel application of disjunctive programming. We have used disjunctive programming to approximate the value function of MIP problems. This viewpoint facilitates the design of decomposition algorithms for special structured MIP problems, especially SMIP problems. Remark 4. It is instructive to study the relationship between the cuts proposed in section 2 with those presented here. Disjunction (15) is stated only in terms of the first stage variables, and hence leads to a disjunctive cut in this space. Thus the projection process occurs prior to the derivation of (17.k). In contrast, the process suggested in Remark 1 requires a projection after the derivation of (9). 16 Proposition 3. Let the first stage approximation solved in iteration k be Min{c x + η η 0, x X B, (η, x) satisfies (17.1),..., (17.k)}. Moreover, assume that the second stage subproblem is a mixed-integer (binary) linear program whose partial solutions are obtained using a branch-and-bound method in which all LP relaxations are feasible, and nodes are fathomed only when the lower bound (on the second stage) exceeds the best available upper bound (for the second stage). Suppose that there exists an iteration K such that for k K, the branch-and-bound method (for each second stage subproblem) provides an optimal second stage solution for all ω Ω, thus yielding an upper bound on the two-stage problem. Then the resulting D 2 -BAC algorithm provides an optimal first stage solution. Proof. We consider iteration indices k K, so that for such k, the second stage objective values provide MIP values for each outcome, and an upper bound on the original problem becomes available. Moreover, the master program provides a non-decreasing sequence of lower bounds (on the original problem) as usual. Since all nodes possess feasible LP relaxations, disjunctive programming theory (Balas [1979]) implies that the collection (17.k) provides facets of the convex hull of Π k (ω). Moreover, since the binary points in the master program are extreme points of X, it follows that the (optimal) value for x = x k in the master program provides the same value as that given by the closure of the convex hull

of Π k (ω). Hence if x k = x k+τ for any τ 1, then the upper and lower bounds obtained from the procedure must be equal and such a point must be optimal. Since there are only finitely many binary vectors (x), there must exist an integers k and τ 1 such that x k = x k+τ, implying optimality. Note that the main property of the binary first stage solutions that we have used above is that such points are extreme points of X which helps us conclude that the closure of the convex hull of the epigraph of the value function agrees with the collection of cuts (17.k) at extreme points of X. Hence, the D 2 -BAC method is applicable to a larger class of problems called extreme point mathematical programs which require feasible solutions to be extreme points of a polyhedral set. 3.2 An Illustration of the D 2 -BAC Algorithm 17 For illustrative purposes, we use the same example as in 2.3. While the instance happens to be deterministic, it will be interesting to compare the master program approximations resulting from the D 2 -BAC method with those of the D-BAC method. The application to SIP problems should be evident. For the sake of convenience, the instance from 2.3 is restated below. Min x 1 2y 1 + 4y 2 4x 1 3y 1 + y 2 6 (x 1, y 1 ) binary, y 2 0. The D 2 -BAC method requires a lower bound on the objective value of the second stage. By noting that the value of 2y 1 +4y 2 on the feasible set must be at least 2, we conclude that η 2. In order to be consistent with our development that requires l = 0, we make the translation that ν = η + 2, so that ν 0. Hence the first master program (see also 2.3) is given by 2 + Min x 1 + ν ν 0 x 1 binary. The above approximation yields the first iterate x 1 = 1, ν = 0. Hence the lower bound is 3. Using the given value of x 1, we proceed to the second stage problem Min 2y 1 + 4y 2 3y 1 + y 2 2 y 1 binary, y 2 0.

As before, the solution to the LP relaxation is fractional, and we solve this problem using a branch-and-bound method which results in the same two nodes as in 2.3. During the branch-and-bound process, the value of the second stage problem is η = 0, and so, an upper bound on the original problem is 1. From the dual solutions obtained for LP relaxations at each node, we arrive at the following disjunction. 18 either or {η 4( 6 + 4x 1 ) + 10, x 1 1, x 1 0, η 2}, {η 0, x 1 1, x 1 0, η 2}. Once again, translating ν = η + 2, this disjunction can be stated as either { 16x 1 + ν 12, x 1 1, x 1 0, ν 0}, or {ν 2, x 1 1, x 1 0, ν 0}. Using the notation of (16), we choose multipliers τ 01 = 1/3, τ 02 = 2/3, τ 1 = (0, 0, 0), τ 2 = (16/3, 0, 0). The resulting disjunctive cut is Max{ 16/3, 16/3}x 1 + Max{1/3, 2/3}ν Min{ 4, 4}. This cut yields the facet 8x 1 + ν 6. Adding this inequality results in an updated master program 2 + Min x 1 + ν 8x 1 + ν 6 ν 0 x 1 binary. This completes one iteration. Observe that the inequality 8x 1 + ν 6 is implied by the constraints in the final master program listed in 2.3. To see this, note that by using a multiplier 1 for the inequality 2x 1 +η 2 (see (14)) and a multiplier 6 for the inequality x 1, we obtain an aggregated inequality 8x 1 + η 8. Substituting η = ν 2 we obtain the same inequality as the facet obtained in the above master program. We begin the next iteration by solving the master program obtained at the end of the first iteration. The optimal solution to this problem is (x 1, ν) = (0, 0). Thus, the lower bound is 2. Solving the second stage problem with x 1 = 0 provides a solution yielding an upper bound of 2 with η = 2. Since the upper and lower bounds are both 2, we declare x 1 = 0 as an optimal solution.

The reader may also find it interesting to use the D 2 -BAC method to solve the following modified instance involving general integer variables in the second stage. (One may also use the D-BAC method for this problem.) 19 Min x 1 2y 1 + 4y 2 4x 1 3y 1 + y 2 7 x 1 binary, y 1 integer, y 1, y 2 0. In the interest of brevity, we only provide the sequence of master programs that are generated by the method. The initial master program is the same as the one given for the previous instance, with the understanding that ν = η + 6, where l = 6 is a lower bound on the second stage value. The second master programs is and the third one is 6 + Min x 1 + ν (8/3)x 1 + ν 4/3 ν 0 x 1 binary, 6 + Min x 1 + ν (8/3)x 1 + ν 4/3 ν 2 ν 0 x 1 binary. It so happens that the solution x 1 provided by the second master program is x 1 = 0, and the branch-and-bound process for the subproblem yields an upper bound of 4 on the original problem. The optimal solution for the third master program is (x 1, ν) = (0, 2), and since ν = η + 6, we conclude that the lower bound on the original problem is 4. Since the upper and lower bounds are equal at this point, we conclude that the first stage solution x 1 = 0 is optimal.

20 4. Conclusions In this paper, we have developed extensions of decomposition-based cutting plane algorithms to allow the use of branch-and-cut methods for stochastic mixed-integer programming (SMIP) problems. These approaches allow us to solve the original MIP using a sequence of smaller MIPs. Given the complexity of these problems, such reductions are valuable. We have investigated two alternative forms: the D-BAC algorithm, and the D 2 -BAC method. The former uses second stage variables as continuous decisions in the first stage, while the latter creates master problems using only first stage decisions. As pointed out in Remark 1, the D-BAC approach may be modified to generate projected inequalities that use only the first stage decisions. The D-BAC method is convenient in the deterministic setting, or for SMIP problems with only a few scenarios in the second stage. However, if the number of outcomes in the second stage is very large (as in large scale SIPs), the D 2 -BAC method allows a more direct facility for creating a master program that uses only the first stage variables. Finally, due to the use of branch-and-bound methods for the second stage problem, these algorithms allow general integer variables in the second stage. Acknowledgements: This research has been supported by grants DMI-9812047, DMI- 9978780, and CISE-9975050 from the National Science Foundation.

References Balas, E. [1979], Disjunctive programming, Annals of Discrete Mathematics, 5, pp. 3-51. Balas, E., S. Ceria and G. Cornuéjols [1993], A lift-and-project cutting plane algorithm for mixed 0-1 programs, Mathematical Programming, 58, pp. 295-324. Blair, C. [1995], A closed-form representation of mixed-integer program value functions, Mathematical Programming, 71, pp. 127-136. Blair, C and R. Jeroslow [1982], The value function of an integer program, Mathematical Programming, 23, pp. 237-273. Birge, J.R. and F. Louveaux [1997], Introduction to Stochastic Programming, Springer. Jeroslow, R. [1980], A cutting plane game for facial disjunctive programs, SIAM Journal on Control and Optimization, 18, pp. 264-281. Laporte, G. and F.V. Louveaux [1993], The integer L-shaped methods for stochastic integer programs with complete recourse, Operations Research Letters, 13, pp. 133-142. Lovász, L. and A. Schriver [1991], Cones of matrices and set functions and 0-1 optimization, SIAM J. on Optimization, 1, pp. 166-190. Ruszczyński, A. [1997], Decomposition methods in stochastic programming, Mathematical Programming B, (Liebling and de Werra (eds.)), 79, pp. 333-353. Sen, S. and J.L. Higle [2000], The C 3 theorem and a D 2 algorithm for large scale stochastic optimization: set convexification, working paper SIE Department, University of Arizona, Tucson, AZ 85721 (also available on Stochastic Programming E-Print Series web site). Sen, S., J.L. Higle and L.Ntaimo [2002], A Summary and Illustration of Disjunctive Decomposition with Set Convexification, to appear in Stochastic Integer Programming and Network Interdiction Models (D.L. Woodruff ed.), Kluwer Academic Press, Dordrecht, The Netherlands. Sherali, H.D. and W.P. Adams [1990], A hierarchy of relaxations between the continuous and convex hull representations for zero-one programming problems, SIAM Journal on Discrete Mathematics, 3, pp. 411-430. Sherali, H.D. and W.P. Adams [1994], A hierarchy of relaxations and convex hull characterizations for mixed integer zero-one programming problems, Discrete Applied Mathematics, 52, pp. 83-106. Sherali, H.D. and B.M.P Fraticelli [2002], A modification of Benders decomposition algorithm for discrete subproblems: an approach for stochastic programs with integer recourse, Journal of Global Optimization, 22, pp. 319-342. 21