Variants of Simplex Method

Similar documents
Review Solutions, Exam 2, Operations Research

February 17, Simplex Method Continued

Linear Programming: Simplex Algorithm. A function of several variables, f(x) issaidtobelinear if it satisþes the two

The Simplex Method. Standard form (max) z c T x = 0 such that Ax = b.

Chapter 4 The Simplex Algorithm Part II

The Simplex Method. Lecture 5 Standard and Canonical Forms and Setting up the Tableau. Lecture 5 Slide 1. FOMGT 353 Introduction to Management Science

Systems Analysis in Construction

Week_4: simplex method II

Simplex Method for LP (II)

The Big M Method. Modify the LP

Linear Programming, Lecture 4

OPTIMISATION 3: NOTES ON THE SIMPLEX ALGORITHM

OPERATIONS RESEARCH. Linear Programming Problem

Standard Form An LP is in standard form when: All variables are non-negativenegative All constraints are equalities Putting an LP formulation into sta

Math 210 Finite Mathematics Chapter 4.2 Linear Programming Problems Minimization - The Dual Problem

c) Place the Coefficients from all Equations into a Simplex Tableau, labeled above with variables indicating their respective columns

Dr. Maddah ENMG 500 Engineering Management I 10/21/07

MATH 445/545 Homework 2: Due March 3rd, 2016

Simplex Algorithm Using Canonical Tableaus

IE 400: Principles of Engineering Management. Simplex Method Continued

Introduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

The Simplex Algorithm

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

21. Solve the LP given in Exercise 19 using the big-m method discussed in Exercise 20.

Worked Examples for Chapter 5

Understanding the Simplex algorithm. Standard Optimization Problems.

Example. 1 Rows 1,..., m of the simplex tableau remain lexicographically positive

Math Models of OR: Some Definitions

Chapter 4 The Simplex Algorithm Part I

1 Review Session. 1.1 Lecture 2

1. Introduce slack variables for each inequaility to make them equations and rewrite the objective function in the form ax by cz... + P = 0.

Contents. 4.5 The(Primal)SimplexMethod NumericalExamplesoftheSimplexMethod

Yinyu Ye, MS&E, Stanford MS&E310 Lecture Note #06. The Simplex Method

IE 5531: Engineering Optimization I

CO350 Linear Programming Chapter 6: The Simplex Method

Lecture 10: Linear programming duality and sensitivity 0-0

Prelude to the Simplex Algorithm. The Algebraic Approach The search for extreme point solutions.

Gauss-Jordan Elimination for Solving Linear Equations Example: 1. Solve the following equations: (3)

Lesson 27 Linear Programming; The Simplex Method

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Linear Programming Duality P&S Chapter 3 Last Revised Nov 1, 2004

ECE 307 Techniques for Engineering Decisions

Review Questions, Final Exam

Linear Programming and the Simplex method

TIM 206 Lecture 3: The Simplex Method

Fundamental Theorems of Optimization

Dr. S. Bourazza Math-473 Jazan University Department of Mathematics

Slide 1 Math 1520, Lecture 10

CLASSICAL FORMS OF LINEAR PROGRAMS, CONVERSION TECHNIQUES, AND SOME NOTATION

IE 400 Principles of Engineering Management. The Simplex Algorithm-I: Set 3

Lecture 10: Linear programming. duality. and. The dual of the LP in standard form. maximize w = b T y (D) subject to A T y c, minimize z = c T x (P)

Optimization Methods in Management Science

Chap6 Duality Theory and Sensitivity Analysis

AM 121: Intro to Optimization Models and Methods

The Simplex Method. Formulate Constrained Maximization or Minimization Problem. Convert to Standard Form. Convert to Canonical Form

The Simplex Algorithm

Solutions to Review Questions, Exam 1

Lecture 4: Algebra, Geometry, and Complexity of the Simplex Method. Reading: Sections 2.6.4, 3.5,

Notes on Simplex Algorithm

Sensitivity Analysis

ORF 522. Linear Programming and Convex Analysis

Chapter 2: Linear Programming Basics. (Bertsimas & Tsitsiklis, Chapter 1)

Summary of the simplex method

The Simplex Method for Solving a Linear Program Prof. Stephen Graves

The simplex algorithm

4.4 The Simplex Method and the Standard Minimization Problem

Generating Gomory's Cuts for linear integer programming problems: the HOW and WHY

The augmented form of this LP is the following linear system of equations:

Slack Variable. Max Z= 3x 1 + 4x 2 + 5X 3. Subject to: X 1 + X 2 + X x 1 + 4x 2 + X X 1 + X 2 + 4X 3 10 X 1 0, X 2 0, X 3 0

Math Models of OR: The Revised Simplex Method

Week 3: Simplex Method I

AM 121: Intro to Optimization

AM 121: Intro to Optimization Models and Methods Fall 2018

Summary of the simplex method

9.5 THE SIMPLEX METHOD: MIXED CONSTRAINTS

The Simplex Algorithm and Goal Programming

Simplex Method in different guises

IP Cut Homework from J and B Chapter 9: 14, 15, 16, 23, 24, You wish to solve the IP below with a cutting plane technique.

OPRE 6201 : 3. Special Cases

MATH 4211/6211 Optimization Linear Programming

Foundations of Operations Research

The Graphical Method & Algebraic Technique for Solving LP s. Métodos Cuantitativos M. En C. Eduardo Bustos Farías 1

F 1 F 2 Daily Requirement Cost N N N

March 13, Duality 3

The Simplex Method of Linear Programming

Linear programming: algebra

In Chapters 3 and 4 we introduced linear programming

Introduction to the Simplex Algorithm Active Learning Module 3

OPTIMISATION /09 EXAM PREPARATION GUIDELINES

Integer Programming. The focus of this chapter is on solution techniques for integer programming models.

9.1 Linear Programs in canonical form

1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations

Operations Research. Duality in linear programming.

"SYMMETRIC" PRIMAL-DUAL PAIR

Decision Procedures An Algorithmic Point of View

Linear Programming Inverse Projection Theory Chapter 3

END3033 Operations Research I Sensitivity Analysis & Duality. to accompany Operations Research: Applications and Algorithms Fatih Cavdur

maxz = 3x 1 +4x 2 2x 1 +x 2 6 2x 1 +3x 2 9 x 1,x 2

Lecture 2: The Simplex method

Transcription:

Variants of Simplex Method All the examples we have used in the previous chapter to illustrate simple algorithm have the following common form of constraints; i.e. a i x + a i x + + a in x n b i, i =,,,m with all these b i being non-negative. ThesigniÞcance of b i being non-negative is that the initialization step of the Simplex algorithm becomes very simple we just put in the slack variables {s i }, andobtainabfsbysettingx j = 0 for all original decision variables and s i = b i, for every slack variable. Indeed, initialization is the serious problem caused by constraints of other forms. In this chapter, we will introduce the concept of artiþcial variable to Þnd a starting BFS, and the Big-M method, as well as the Two-Phase Method, that solves the expanded LP problem. The Big-M method We will illustrate the main idea by solving the following simple example. Example: Solve the LP problem: Minimize Z =x +x and x,x 0. x + x 4 x + x Solution: This problem can be transformed into canonical form by adding slack variables and change the minimization to maximization: Maximize Z = x x x + x s = 4 x + x + s = and x j,s i 0. If we did it the old way, setting x i =0,wehaves = 4 ands =,whichisnot feasible since s is now negative. To overcome this difficulty, we introduce an artiþcial variable ā and write the constraints as x + x s + ā = 4 x + x + s =

with x j 0,s i 0andā 0. However, the consequences of the introduction of the artiþcial variable are two-folds: () it is easy now to determine an initial BFS corresponding to the new constraints, namely, NBV =(x,x,s )=0, BV =(s, ā )=(, 4). () there is, however, no guarantee that the optimal solution to the new constraints will be the same as that to the original constraints; in other words, the artiþcial variable in the new optimal solution may take strictly positive value. Indeed, if we solve the LP with new constraints, we will have an optimal solution (x,x )=(0, 0), (s,s )=(0, 0), ā =4, and the optimal value Z = 0, which is wrong because the graphical method will easily show that the optimal solution is (x,x )=(, 0), Z = 4, Z = Z =4. Remark: Why the optimal solution could be different? Indeed, the introduction of artiþcial variable could change the feasible region the expanded LP is now equivalent to Maximize Z = x x x + x s 4 x + x, which is further equivalent to (why?) Maximize Z = x x x + x. The change of feasible region could very possible change the optimal solution. Big-M method: One way to guarantee that the new optimal solution is optimal for the original LP, is to modify the objective function, so that the artiþcial variable will take value zero in the new optimal solution. In other words, a very large penalization is added to the objective function if the artiþcial variable takes positive value. Consider the following LP: Maximize Z = x x Mā x + x s +ā = 4 x + x + s =. Here M is a symbolic big positive number. It is so big that even if ā is slightly big than 0, the penalization Mā will be very severe. In this case, it is reasonable that the optimal solution to this new LP will take value 0 for the artiþcial variable ā, and hence an optimal solution for the original LP. Let us perform the simplex algorithm to Þnd the optimal solution to this new LP. We write down the following table:

Basic Variable Row Z x x s s ā RHS Ratios Z (0) 0 0 M 0 ā () 0-0 4 s () 0-0 0 You might want to jump to the conclusion that the current BFS is optimal, since all the coefficients in Row (0) is non-negative. This is wrong! The validity of this optimality test is based on the assumption that the objective function Z is written in terms of non-basic variable only. However, in the above table, Row (0) contains a basic variable ā with non-zero coefficient. Therefore, we need to perform Gaussian elimination Þrst to make the coefficients of all basic variables be 0. We have Basic Variable Row Z x x s s ā RHS Ratios Z (0) -M -M M 0 0-4M ā () 0-0 4 4/ = min s () 0-0 0 Basic Variable Row Z x x s s ā RHS Ratios Z (0) 0 0 M- -4 x () 0 0.5-0.5 0 0.5 s () 0 0.5-0.5 0.5 We have the optimal solution (x,x )=(, 0), (s,s )=(0, ), ā =0, Z = 4, Z = Z =4. Remark: The symbol M represents a huge number, and it will only appear in Row (0). The symbolic quantity am +b (in case a 6= 0)almostequalsaM since M is much larger compared with the constant b. Therefore, am + b is positive (resp. negative) as long as a is positive (resp. negative). We have, for example 0.M 000 > 0, 0.M +00< 0, M 00 >M+000, M +00 < M 000. Remark: It is possible that in the new optimal solution the artiþcial variable takes a strictly positive value. If this is the case, the original problem has no feasible solution. Indeed, for example, say we are to Maximize Z = c x + c x Ma where a is an artiþcial variable. Suppose in the optimal solution to this LP, the artiþcial variable a > 0, then we have Z = a M + b, for some constant b. IftheoriginalLPhasafeasiblesolution,say( x, x ), then ( x, x,a= 0) is a feasible solution to the expanded LP. Z c x + c x M 0=c x + c x. Combining these two inequalities, we have a M + b c x + c x, which is impossible since M is a really big constant, and a is strictly positive. A contradiction. Hence the original LP has no feasible solution.

Equality constraints: Suppose one of the constraints is, say, a i x + a i x + + a in x n = b i. in a maximization LP. Without loss of generality, assume b i 0 (otherwise, just multiply on both sides). One can write this equality constraint as two inequality constraints (i.e. and ) and then use the Big-M method. However, more conveniently, we can directly introduce the artiþcial variable ā and rewrite the LP as Maximize Z = Mā while change the constraint to a i x + a i x + + a in x n +ā = b i Example: Solve the following LP: Maximize Z = x + x x + x x + x = 6. Solution: The new LP will be Maximize Z = x + x Mā Mā x + x s + ā = x + x + ā = 6. We have the following table: Z (0) - 0 M M 0 ā () 0-0 ā () 0 0 0 6 To start the simplex algorithm, we need to write the value function Z as a function of NBV variables. Z (0) -4M --M M 0 0-7M ā () 0-0 / = min ā () 0 0 0 6 6/ = Z (0) 0 M- -M+ 4M- 0 -M- x () 0-0 ā () 0 0 - - / = min 4

Z (0) 0-5 0 M M- - x () 0 0 0 / = min s () 0 0 - - Z (0).5 0 0 M M +0.5 x () 0.5 0 0 0.5 s () 0 0.5 0-0.5 Therefore, the optimal solution is Two-Phase method (x,x )=(0, ), s =, (ā, ā )=(0, 0), Z =. The two-phase method and big-m method are equivalent. In practice, however, most computer codes utilizes the two-phased method. The reasons are that the inclusion of the big number M may cause round-off error and other computational difficulties. The two-phase method, on the other hand, does not involve the big number M and hence all the problems are avoided. Thetwo-phasemethod,asitiscalled,dividestheprocessintotwophases. Phase : The goal is to Þnd a BFS for the original LP. Indeed, we will ignore the original objective for a while, and instead try to minimize the sum of all artiþcial variable. At the end of phase, a BFS is obtained if the minimal value of this LP is zero (why?). Phase : Drop all the artiþcial variables, change the objective function back to the original one. Use just the regular simplex algorithm, with the starting BFS obtained in Phase. We will illustrate the idea re-doing the last example in the preceding section. Example (revisited): Solve the following LP: Maximize Z = x + x x + x x + x = 6. Solution: Phase : The LP for this phase is Minimize Z =ā +ā or Maximize Z = ā ā x + x s + ā = x + x + ā = 6. We have the following table. Z (0) 0 0 0 0 ā () 0-0 ā () 0 0 0 6 5

Note in this Þrst table, the objective function Z = ā ā iswrittenintermofthebasicvariables, so Þrst we need to express it in terms of the non-basic variable. This leads to the following table, from which the simplex algorithm starts. Z (0) -4-0 0-7 ā () 0-0 / = min ā () 0 0 0 6 6/ = Z (0) 0-4 0 - x () 0-0 ā () 0 0 - - / = min Z (0) 0 0 0 0 x () 0 0 0 s () 0 0 - - This is the end of phase, and we indeed obtain an BFS with NBV =(x, ā, ā )=(0, 0, 0), BV =(x,s )=(, ). Phase : We will just drop all the columns of artiþcial variables in the Þnal tableau of Phase, and rewrite the objective function back to the original one. Basic Variable Row Z x x s RHS Ratios Z (0) - 0 0 x () 0 0 s () 0 0 - However, again, the objective function Z is not expressed in terms of non-basic variables x,we will do Gaussian elimination to make it so. Basic Variable Row Z x x s RHS Ratios Z (0) 0 5 0 - x () 0 0 / = min s () 0 0 - Basic Variable Row Z x x s RHS Ratios Z (0).5 0 0 x () 0.5 0 s () 0 0.5 0 The optimal solution is (x,x )=(0, ), s =, Z =. 6

Remark: If in Phase, the optimal value for ā +ā is strictly positive (or ā ā is strictly negative), then the original Lp has no feasible solution. Because otherwise, there must exist a feasible solution with (ā, ā ) = 0, which is a contradiction. Remark: In the example given, in the Þnal tableau of the LP in Phase, the artiþcial variables are both non-basic variables. It is possible that in the Þnal tableau of Phase, some artiþcial variables are basic variable, while the value of Z already reaches 0. In this case, the problem is degenerate, and we will not go into details about this. But keep in mind, even in this case, it is also very easy to move to Phase, and Þnd the optimal solution for the original LP. Remark: When a variable, say x, has no sign constraint, we just write x = u v, u 0, v 0, and then perform the simplex algorithm. It is easy to see that, in all the tables, the column for u is always the negative of that of v. Therefore, it will never be the case that u and v are basic variables simultaneously (why?). In other words, it will never be the case that u,v are both taking strictly positive values. 7