OPERATIONS RESEARCH. Linear Programming Problem

Similar documents
Standard Form An LP is in standard form when: All variables are non-negativenegative All constraints are equalities Putting an LP formulation into sta

Simplex Algorithm Using Canonical Tableaus

UNIT-4 Chapter6 Linear Programming

Slack Variable. Max Z= 3x 1 + 4x 2 + 5X 3. Subject to: X 1 + X 2 + X x 1 + 4x 2 + X X 1 + X 2 + 4X 3 10 X 1 0, X 2 0, X 3 0

Systems Analysis in Construction

Linear Programming and the Simplex method

Linear Programming, Lecture 4

Part 1. The Review of Linear Programming

MATH2070 Optimisation

AM 121: Intro to Optimization Models and Methods

9.1 Linear Programs in canonical form

Review Solutions, Exam 2, Operations Research

Contents. 4.5 The(Primal)SimplexMethod NumericalExamplesoftheSimplexMethod

In Chapters 3 and 4 we introduced linear programming

F 1 F 2 Daily Requirement Cost N N N

Yinyu Ye, MS&E, Stanford MS&E310 Lecture Note #06. The Simplex Method

Lecture 2: The Simplex method

The Simplex Method. Lecture 5 Standard and Canonical Forms and Setting up the Tableau. Lecture 5 Slide 1. FOMGT 353 Introduction to Management Science

THE UNIVERSITY OF HONG KONG DEPARTMENT OF MATHEMATICS. Operations Research I

min 4x 1 5x 2 + 3x 3 s.t. x 1 + 2x 2 + x 3 = 10 x 1 x 2 6 x 1 + 3x 2 + x 3 14

IE 5531: Engineering Optimization I


CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

Introduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

1 Review Session. 1.1 Lecture 2

Developing an Algorithm for LP Preamble to Section 3 (Simplex Method)

Lecture 2: The Simplex method. 1. Repetition of the geometrical simplex method. 2. Linear programming problems on standard form.

Simplex Method for LP (II)

Farkas Lemma, Dual Simplex and Sensitivity Analysis

Lecture 4: Algebra, Geometry, and Complexity of the Simplex Method. Reading: Sections 2.6.4, 3.5,

Part 1. The Review of Linear Programming

Lesson 27 Linear Programming; The Simplex Method

AM 121: Intro to Optimization

TRANSPORTATION PROBLEMS

The Simplex Method. Formulate Constrained Maximization or Minimization Problem. Convert to Standard Form. Convert to Canonical Form

21. Solve the LP given in Exercise 19 using the big-m method discussed in Exercise 20.

Chapter 1: Linear Programming

MATH 4211/6211 Optimization Linear Programming

Linear Programming. Linear Programming I. Lecture 1. Linear Programming. Linear Programming

The Simplex Method for Solving a Linear Program Prof. Stephen Graves

Summary of the simplex method

Math 273a: Optimization The Simplex method

CO 602/CM 740: Fundamentals of Optimization Problem Set 4

OPTIMISATION 3: NOTES ON THE SIMPLEX ALGORITHM

Fundamental Theorems of Optimization

"SYMMETRIC" PRIMAL-DUAL PAIR

3 The Simplex Method. 3.1 Basic Solutions

3. THE SIMPLEX ALGORITHM

CO350 Linear Programming Chapter 8: Degeneracy and Finite Termination

The Simplex Algorithm and Goal Programming

Ω R n is called the constraint set or feasible set. x 1

AM 121: Intro to Optimization Models and Methods Fall 2018

Operations Research Lecture 2: Linear Programming Simplex Method

LINEAR PROGRAMMING I. a refreshing example standard form fundamental questions geometry linear algebra simplex algorithm

MATH 445/545 Homework 2: Due March 3rd, 2016

Lecture 6 Simplex method for linear programming

Lecture slides by Kevin Wayne

Gauss-Jordan Elimination for Solving Linear Equations Example: 1. Solve the following equations: (3)

Special cases of linear programming

Introduce the idea of a nondegenerate tableau and its analogy with nondenegerate vertices.

OPERATIONS RESEARCH. Michał Kulej. Business Information Systems

Metode Kuantitatif Bisnis. Week 4 Linear Programming Simplex Method - Minimize

Example. 1 Rows 1,..., m of the simplex tableau remain lexicographically positive

New Artificial-Free Phase 1 Simplex Method

Week_4: simplex method II

OPERATIONS RESEARCH. Transportation and Assignment Problems

Note 3: LP Duality. If the primal problem (P) in the canonical form is min Z = n (1) then the dual problem (D) in the canonical form is max W = m (2)

Dr. Maddah ENMG 500 Engineering Management I 10/21/07

1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations

Chapter 4 The Simplex Algorithm Part I

Chapter 5 Linear Programming (LP)

System of Linear Equations

4.6 Linear Programming duality

Duality Theory, Optimality Conditions

Linear Programming Duality P&S Chapter 3 Last Revised Nov 1, 2004

2. Linear Programming Problem

Distributed Real-Time Control Systems. Lecture Distributed Control Linear Programming

Chapter 1 Linear Programming. Paragraph 5 Duality

Dr. S. Bourazza Math-473 Jazan University Department of Mathematics

Understanding the Simplex algorithm. Standard Optimization Problems.

TIM 206 Lecture 3: The Simplex Method

END3033 Operations Research I Sensitivity Analysis & Duality. to accompany Operations Research: Applications and Algorithms Fatih Cavdur

4.5 Simplex method. LP in standard form: min z = c T x s.t. Ax = b

Lecture 10: Linear programming duality and sensitivity 0-0

LINEAR PROGRAMMING 2. In many business and policy making situations the following type of problem is encountered:

Chap6 Duality Theory and Sensitivity Analysis

Lecture 11 Linear programming : The Revised Simplex Method

Linear programming. Starch Proteins Vitamins Cost ($/kg) G G Nutrient content and cost per kg of food.

ECE 307 Techniques for Engineering Decisions

4. Duality and Sensitivity

Chapter 4 The Simplex Algorithm Part II

OPRE 6201 : 3. Special Cases

4. Duality Duality 4.1 Duality of LPs and the duality theorem. min c T x x R n, c R n. s.t. ai Tx = b i i M a i R n

Summary of the simplex method

February 17, Simplex Method Continued

3 Development of the Simplex Method Constructing Basic Solution Optimality Conditions The Simplex Method...

MATH 445/545 Test 1 Spring 2016

Math Models of OR: Some Definitions

CPS 616 ITERATIVE IMPROVEMENTS 10-1

Prelude to the Simplex Algorithm. The Algebraic Approach The search for extreme point solutions.

Transcription:

OPERATIONS RESEARCH Chapter 1 Linear Programming Problem Prof. Bibhas C. Giri Department of Mathematics Jadavpur University Kolkata, India Email: bcgiri.jumath@gmail.com

MODULE - 2: Simplex Method for Solving LPP and Big-M Method 2.1 Simplex Method In 1947, George Dantzig developed an efficient method called simplex method for solving LP problems having many variables. The concept of simplex method is similar to the graphical method in which extreme points of the feasible region are examined in order to find the optimal solution. Here, the optimal solution lies at an extreme point of a multi-dimensional polyhedron. The simplex method is based on the property that the optimal solution, if exists, can always be found in one of the basic feasible solutions. 2.1.1 Canonical and Standard Forms of An LPP An LPP is said to be in canonical form when it is expressed as Maximize Z = c 1 x 1 + c 2 x 2 +... + c n x n a i1 x 1 + a i2 x 2 +... + a ij x j +... + a in x n b i, i = 1,2,,m x 1,x 2,...,x n 0 The characteristics of this form are as follows: (i) The objective function is of maximization type (Maximize Z). In case of Minimize Z, it can be written as Maximize ( Z). (ii) All constraints are of type, except the non-negative restrictions. (iii) All variables are non-negative. 2

An LPP in the following form is known as standard form: Maximize (or Minimize) Z = c 1 x 1 + c 2 x 2 +... + c n x n a i1 x 1 + a i2 x 2 +... + a ij x j +... + a in x n = b i, i = 1,2,,m and x 1,x 2,...,x n 0 or Maximize( or Minimize) Z = cx Ax = b x 0 (null vector) where c = (c 1,c 2,,c n ) an n-component row vector; x = [x 1,x 2,,x m ] an m-component column vector; b = [b 1,b 2,,b m ] an m-component column vector and the matrix A = (a ij ) m n. The characteristics of this form are as follows: (i) All constraints are expressed in the form of equations, except the non-negative restrictions. (ii) The RHS of each constraint equation is non-negative. 2.1.2 Slack and Surplus Variables Slack variable - A variable which is added to the LHS of a type constraint to convert the constraint into an equality is called slack variable. Surplus variable - A variable which is subtracted from the LHS of a type constraint to convert the constraint into an equality is called surplus variable. 2.1.3 Basic Solution Consider a set of m linear simultaneous equations of n (n > m) variables Ax = b, where A is an m n matrix of rank m. If any m m non-singular matrix B is chosen from A and if all the (n m) variables not associated with the chosen matrix are set equal to zero, then the solution to the resulting system of equations is a basic solution (BS). Basic solution has not more than m non-zero variables called basic variables. Thus the m vectors associated with m basic variables are linearly independent. The variables which are not basic, are termed as non-basic variables. If the number of non-zero

basic variables is less than m, then the solution is called degenerate basic solution. On the other hand, if none of the basic variables vanish, then the solution is called nondegenerate basic solution. equations in n unknowns is n C m = n! m!(n m)!. The possible number of basic solutions in a system of m Theorem 2.1: The necessary and sufficient condition for the existence and non-degeneracy of all the basic solutions of Ax = b is that every set of m columns of the augmented matrix [A, b] is linearly independent. Proof: Let us suppose that all the basic solutions exist and none of them is degenerate. Then, if a 1,a 2,,a m be one set of m column vectors of A, corresponding to the set of basic variables x 1,x 2,,x m, we have x 1 a 1 + x 2 a 2 + + x m a m = b x i 0,i = 1,2, m. Hence, by the replacement theorem of vector, a 1 can be replaced by b in the basis (a 1,a 2,,a m ) as x 1 0. Then the set of vectors {b,a 2,,a m } forms a basis. Similarly, a 2 can be replaced by b in the basis as x 2 0 and therefore, {a 1,b,a 3,,a m } forms a basis. Proceeding in this way, we can show that b along with any (m 1) vectors of A forms a basis of E m and are linearly independent. Thus, every set of m columns of [A, b] is linearly independent. Hence the condition is necessary. Next, we suppose that the set of vectors {a 1,a 2,,a m } of m columns of [A, b] is linearly independent. Hence b can be expressed as the linear combination of these vectors as b = x 1 a 1 + x 2 a 2 + + x m a m where (x 1,x 2,,x m ) is the corresponding basic solution. Now, if one of them, say x 1, is equal to zero, then 0 = 1b + x 2 a 2 + + x m a m so the vectors (b,a 2,,a m ) are linearly dependent m column vectors of the augmented matrix [A, b] which is a contradiction to the assumption. Thus, since b,a 2,,a m are linearly independent, the coefficient x 1 of a 1 cannot vanish as b can replace a 1 maintaining its basic character. By similar argument, vectors a 1,b,a 3,,a m are linearly independent and the coefficient x 2 of a 2 cannot vanish. Thus we see that none of x i s can vanish and the solution is non-degenerate. Hence all basic solutions exist and are non-degenerate.

2.1.4 Basic Feasible Solution A solution which satisfies all the constraints and non-negativity restrictions of an LPP is called a feasible solution. If again the feasible solution is basic, then it is called a basic feasible solution (BFS). Theorem 2.2: The necessary and sufficient condition for the existence and non-degeneracy of all possible basic feasible solutions of Ax = b, x 0 is the linear independent of every set of m columns of the augmented matrix [A, b], where A is the m n coefficient matrix. The proof is omitted as it is similar to that of Theorem 2.1 Example 2.1: Find the basic feasible solutions of the following system of equations 2x 1 + 3x 2 x 3 + 4x 4 = 8 x 1 2x 2 + 6x 3 7x 4 = 3 x 1,x 2,x 3,x 4 0. Solution: The given system of equations can be written as a 1 x 1 +a 2 x 2 + +a 4 x 4 = b where a 1 = [2, 1], a 2 = [3, 2], a 3 = [ 1, 6], a 4 = [4, 7] and b = [8, 3]. The maximum number of basic solutions that can be obtained is 4 C 2 = 6. The six sets of 2 vectors out of 4 are B 1 = [a 1, a 2 ], B 2 = [a 1, a 3 ], B 3 = [a 1, a 4 ] B 4 = [a 2, a 3 ], B 5 = [a 2, a 4 ], B 6 = [a 3, a 4 ]. Here B 1 = 7, B 2 = 18, B 3 = 18, B 4 = 16, B 5 = 13, and B 6 = 17. Since none of these determinants vanishes, hence every set B i of two vectors is linearly independent. Therefore, the vectors of the basic variables associated to each set B i, i = 1,2,3,4,5,6 are given by

x B1 = B 1 1 b = 1 7 x B2 = B 1 2 b = 1 13 x B3 = B 1 3 b = 1 18 x B4 = B 1 4 b = 1 16 x B5 = B 1 5 b = 1 13 and x B6 = B 1 6 b = 1 17 2 3 8 1 2 3 6 1 8 1 2 3 7 4 8 1 2 3 6 1 8 2 3 3 7 4 8 2 3 3 7 6 = 1 2 = 45/13 14/13 = 22/9 7/9 = 45/16 7/16 = 44/13 7/13 4 8 1 3 = 44/17 45/17 From above, we see that the possible basic feasible solutions are x 1 = [1, 2, 0, 0], x 2 = [22/9, 0, 0, 7/9], x 3 = [0, 45/16, 7/16, 0] and x 4 = [0, 0, 44/17, 45/17] which are also non-degenerate. The other basic solutions are not feasible. Theorem 2.3 (Fundamental Theorem of Linear Programming): If a linear programming problem admits of an optimal solution, then the optimal solution will coincide with at least one basic feasible solution of the problem. Proof: Let us assume that x is an optimal solution of the following LPP : Maximize z = cx Ax = b, x 0 (2.1) Without any loss of generality, we assume that the first p components of x are nonzero and the remaining (n p) components of x are non-zero. Thus x = [x 1,x 2,,x p,0,0,,0]. Then, from (2.1), Ax = b gives p a ij x j = b i, i = 1,2,,m. Also, A = [a 1,a 2,,a p,a p+1,,a n ] gives p Also z = z max = c j x j. a 1 x 1 + a 2 x 2 + + a p x p = b. (2.2) (2.3)

Now, if the vectors a 1, a 2,, a p corresponding to the non-zero components of x are linearly independent, then, by definition, x is a basic solution and hence the theorem holds in this case. If p = m, then the basic feasible solution is non-degenerate. On the other hand, if p < m then it will form a degenerate basic feasible solution with (m p) basic variables equal to zero. However, if the vectors a 1, a 2,, a p are not linearly independent, then they must be linearly dependent and there exists scalars λ j, j = 1,2,,p of which at least one of the λ j s is non-zero such that λ 1 a 1 + λ 2 a 2 + + λ p a p = 0. (2.4) Suppose that at least one λ j > 0. If the non-zero λ j is not positive, then we can multiply (2.4) by ( 1) to get a positive λ j. } Let µ = Max 1 j p { λj x j (2.5) Then µ is positive as x j > 0 for all j = 1,2,,p and at least one λ j is positive. Dividing (2.4) by µ and subtracting it from (2.2), we get ( x 1 λ ) ( 1 a µ 1 + x 2 λ ) ( 2 a µ 2 + + x p λ ) p a µ p = b [( and hence x 1 = x 1 λ ) ( 1, x µ 2 λ ) ( 2,, x µ p λ ) ] p,0,0,,0 µ is a solution of the system of equations Ax = b. Again from (2.5), we have or µ λ j x j for j = 1,2,,p x j λ j µ 0 for j = 1,2,,p (2.6) This implies that all the components of x 1 are non-negative and hence x 1 is a feasible solution of Ax = b, x 0. Again, for at least one value of j, we have, from (2.5), x j λ j µ = 0, for at least one value of j. Thus we see that the feasible solution x 1 will contain one more zero than it was shown to have in (2.6). Thus, the feasible solution x 1 cannot contain more than (p 1) non-zero variables. Therefore, we have shown that the number of positive variables giving an optimal solution can be reduced. Now, we have to show that even after this reduction x 1 remains optimal. Let z be the value of the objective function for this value of x. Then p ( z = cx 1 = c j x j λ ) p p j λ j = c µ j x j c j µ = z 1 µ p c j λ j, by (2.3) (2.7)

Now, if we can show that p c j λ j = 0 (2.8) then z = z and this will prove that x 1 is an optimal solution. We assume that (2.8) does not hold and we find a suitable real number γ, such that γ(c 1 λ 1 + c 2 λ 2 + + c p λ p ) > 0 i.e., c 1 (γλ 1 ) + c 2 (γλ 2 ) + + c p (γλ p ) > 0. Adding (c 1 x 1 + c 2 x 2 + + c p x p ) to both sides, we get c 1 (x 1 + γλ 1 ) + c 2 (x 2 + γλ 2 ) + + c p (x p + γλ p ) > c 1 x 1 + c 2 + x 2 + + c p x p = z (2.9) Again, multiplying (2.4) by γ and adding to (2.2), we get (x 1 + γλ 1 )a 1 + (x 2 + γλ 2 )a 2 + + (x p + γλ p )a p = b so that [(x 1 + γλ 1 ),(x 2 + γλ 2 ), + (x p + γλ p ),0,0,,0] (2.10) is also a solution of the system Ax = b. Now, we choose γ such that or and x j + γλ j 0 for all j = 1,2,,p γ x j λ j if λ j > 0 γ x j λ j if λ j > 0 and γ is unrestricted, if λ j = 0. Now (2.10) becomes a feasible solution of Ax = b, x 0. Thus choosing γ in a manner Max j λ j > 0 { x } j γ λ j Min j λ j < 0 { x } j λ j we see, from (2.9), that the feasible solution (2.10) gives a greater value of the objective function than z. This contradicts our assumption that z is the optimal value and thus we see that p c j λ j = 0

Hence x 1 is also an optimal solution. Thus we show that from the given optimal solution, the number of non-zero variables in it is less than that of the given solution. If the vectors associated with the new non-zero variables is linearly independent, then the new solution will be a basic feasible solution and hence the theorem follows. If again the new solution is not a basic feasible solution, then we can further diminish the number of non-zero variables as above to get a new set of optimal solutions. We may continue the process until the optimal solution obtained is a basic feasible solution. 2.1.5 Simplex Algorithm For the solution of any LP problem by simplex algorithm, the existence of an initial BFS is always assumed. Here we will discuss the simplex algorithm of maximization type LP problem. The steps for computation of an optimal solution are as follows: Step 1: Convert the objective function into maximization type if the given LPP is of minimization type. Also, convert all the m constraints such that b i s (i = 1, 2,..., m) are all non-negative. Then convert each inequality constraint into equation by introducing slack or surplus variable and assign a zero cost coefficient to such a variable in the objective function. Step 2: If needed, introduce artificial variable(s) and take (- M) as the coefficient of each artificial variable in the objective function. Step 3: Obtain the initial basic feasible solution x B = B 1 b where B is the basis matrix which is an identity matrix here. Step 4: Calculate the net evaluations z j c j = c B x Bj - c j. (i) If z j c j 0 for all j then x B is an optimum BFS. (ii) If at least one z j c j < 0, then proceed to improve the solution in the next step. Step 5: If there are more than one negative z j c j, then choose the most negative of them. Let it be z k c k for some j = k. (i) If all a ik < 0 (i = 1,2,...,m), then there exists an unbounded solution to the given problem. (ii) If at least one a ik > 0 (i = 1,2,...,m) then the corresponding vector a k enters the basis B. This column is called the key or pivot column.

Step 6: Divide each value of x B (i.e., b i ) by the corresponding (but positive) number in the key column and select a row which has the ratio non-negative and minimum, i.e., { } x Br xbi = Min ;a a rk a ik > 0 ik This rule is called minimum ratio rule. The row selected in this manner is called the key or pivot row and it represents the variable which will leave the basic solution. The element that lies in the intersection of key row and key column of the simplex table is called the key or pivot element (say a rk ). Step 7: Convert the leading element to unity by dividing its row by the key element itself and all other elements in its column to zeros by making use of the relation: â rj = a rj a rk â ij = a ij a rj a rk a ik and ˆx Br = x Br a rk, i = r;j = 1,2,...n and ˆx Bi = x Bi x Br a rk a ik, i = 1,2,...,m;i r Step 8: Go to step 4 and repeat the procedure until all entries in (z j c j ) are either positive or zero, or there is an indication of an unbounded solution. 2.1.6 Simplex Table The simplex table for a standard LPP is given below: where Maximize z = cx Ax = b x 0 B = (a B1, a B2,, a Bm ), basis matrix x B = (x B1, x B2,, x Bm ), basic variables c B = [c B1, c B2,, c Bm ] A = ( a ij ) m n b = [b 1,b 2,,b m ] c = (c 1, c 2,, c n ) x = (x 1, x 2,, x n )

c j c 1 c 2... c n c B B x B b a 1 a 2... a n c B1 a B1 x B1 b 1 a 11 a 12... a 1n c B2 a B2 x B2 b 2 a 21 a 22... a 2n........................ c Bm a Bm x Bm b m a m1 a m2... a mn z j c j z 1 c 1 z 2 c 2... z n c n Table 2.1: Simplex table Example 2.2: Solve the following LP problem by simplex method: Minimize Z =x 1 3x 2 + 2x 3 3x 1 x 2 + 2x 3 7 2x 1 + 4x 2 12 4x 1 + 3x 2 + 8x 3 10 x 1,x 2,x 3 0 Solution: This is a minimization problem. Therefore, converting the objective function for maximization, we have Max Z 1 = Min ( Z) = x 1 +3x 2 2x 3. After introducing the slack variables x 4,x 5 and x 6, the problem can be put in the standard form as Max Z 1 = x 1 + 3x 2 2x 3 + 0x 4 + 0x 5 + 0x 6 3x 1 x 2 + 2x 3 + x 4 = 7 2x 1 + 4x 2 + x 5 = 12 4x 1 + 3x 2 + 8x 3 + x 6 = 10 x 1,x 2,x 3,x 4,x 5,x 6 0 Now, we apply simplex algorithm. The results of successive iteration are shown in Table 2.2. Since z j c j 0 for all j in the last iteration of Table 2.2, optimality condition is satisfied. The optimal solution is x 1 = 4, x 2 = 5 and x 3 = 0 and the corresponding value of the objective function is (Z 1 ) max = 11. Hence, the solution of the original problem is x 1 = 4, x 2 = 5, x 3 = 0 and Z min = 11.

c j -1 3-2 0 0 0 Mini c B B x B b a 1 a 2 a 3 a 4 a 5 a 6 Ratio 0 a 4 x 4 7 3-1 2 1 0 0-0 a 5 x 5 12-2 4 0 0 1 0 12/4=3 0 a 6 x 6 10-4 3 8 0 0 1 10/3=3.33 z j c j 1-3 2 0 0 0 0 a 4 x 4 10 5/2 0 2 1 1/4 0 3 a 2 x 2 3-1/2 1 0 0 1/4 0 0 a 6 x 6 1-5/2 0 8 0-3/4 1 z j c j -1/2 0 2 0 3/4 0-1 a 1 x 1 4 1 0 4/5 2/5 1/10 0 3 a 2 x 2 5 0 1 2/5 1/5 3/10 0 0 a 6 x 6 11 0 0 10 1-1/2 1 z j c j 0 0 12/5 1/5 16/20 0 Table 2.2: Simplex Table for Example 2.2 2.2 Artificial Variable and Big-M Method In the standard form of an LPP, when the decision variables together with slack and surplus variables cannot afford initial basic variables, a non-negative variable is introduced to the LHS of each equation which lacks the starting basic variable. This variable is known as artificial variable. Big M method is a method of solving LPP having artificial variables. In this method, a very large negative price ( M) (M is positive) is assigned to each artificial variable in the objective function of maximization type. After introducing the artificial variable(s), the problem can be solved in the usual simplex method. However, while solving in simplex, the following conclusions are drawn from the final table: (a) The current solution is an optimal BFS, if no artificial variable remains in the basis and the optimality condition is satisfied. (b) The current solution is an optimal degenerate BFS, if at least one artificial variable appears in the basis at zero level and the optimality condition is satisfied. (c) The problem has no feasible solution, if at least one artificial variable appears in the basis at positive level and the optimality condition is satisfied.

Example 2.3: Solve the following LP problem by simplex method: Minimize Z = 2x 1 + 3x 2 x 1 + x 2 5 x 1 + 2x 2 6 x 1,x 2 0 Solution: Introducing surplus and artificial variables, the given problem can be written in standard form as Max Z = Mini ( Z) = 2x 1 3x 2 + 0x 3 + 0x 4 Mx 5 Mx 6 x 1 + x 2 x 3 + x 5 = 5 x 1 + 2x 2 x 4 + x 6 = 6 x 1,x 2,x 3,,x 6 0 c j -2-3 0 0 -M -M Mini c B B x B b a 1 a 2 a 3 a 4 a 5 a 6 Ratio -M a 5 x 5 5 1 1-1 0 1 0 5/1=5 -M a 6 x 6 6 1 2 0-1 0 1 6/2=2 z j c j -2M+2-3M +3 M M 0 0 -M a 5 x 5 2 1/2 0-1 1/2 1 2/(1/2)=4-3 a 2 x 2 3 1/2 1 0-1/2 0 3/(1/2)=6 z j c j (-M+1)/2 0 M (-M+3)/2 0-2 a 1 x 1 4 1 0-2 1-3 a 2 x 2 1 0 1 1-1 z j c j 0 0 1 1 Table 2.3: Simplex Table for Example 2.3 From the last iteration in Table 2.3, we see that z j c j 0 for all j. Hence the optimality condition is satisfied. The optimal basic feasible solution is x 1 = 4 and x 2 = 1 and the corresponding Z max = 11, i.e. Z min = 11.