Homework Assignment 4 Solutions
|
|
- Brandon Copeland
- 6 years ago
- Views:
Transcription
1 MTAT.03.86: Advanced Methods in Algorithms Homework Assignment 4 Solutions University of Tartu 1 Probabilistic algorithm Let S = {x 1, x,, x n } be a set of binary variables of size n 1, x i {0, 1}. Consider a binary equation φ with m 1 unknowns of the form x t1 + x t + + x tm = 0, where all computations are done modulo, and t i {1,,, n} for all i = 1,,, m. (a) Consider a simple probabilistic algorithm that assigns values 0 and 1 to all x i, i = 1,,, n, as follows Prob(x i = 0) = Prob(x i = 1) = 1 (the value to each variable is assigned independently of other variables). Show that the probability that the proposed algorithm does not find a solution of φ is exactly 1. Hint: for each assignment of values that does not solve φ there is an assignment that solves it. (b) Consider now a system of equations φ 1, φ,, φ r as above, r 1, where any two equations φ i and φ j use a disjoint set of unknowns. What is the probability that the algorithm in part (a) solves the system in (b) (i.e., all r equations are solved at the same time)? (c) How many times should the algorithm in (a) be executed in order to solve the system given in (b) with probability at least 1? at least 0.99? 1.1 Failure probability First, consider the case where m = 1, then the equation is x t1 = 0 and by definition Prob(x t1 = 0) = 1 as required. Now, assume that there are more variables m. Then we have the following using the law of total probability and some simplifications. Prob(x t1 + x t + + x tm = 0) = Prob(x tm = 0) Prob(x t1 + x t + + x tm 1 = 0) + Prob(x tm = 1) Prob(x t1 + x t + + x tm 1 = 1) = 1 (Prob(x t 1 + x t + + x tm 1 = 1) + Prob(x t1 + x t + + x tm 1 = 0)) = 1 1 = 1 1
2 1. System of equations As each equation has distinct variables, then the probability of solving any one equation is 1. All r equations are independent, therefore the probability of the joint event that all equations are solved is Prob(solved) = 1 r. 1.3 Improving the probability with repetitions Repetitions can help to increase the success probability. Let the number of repetitions be k. The probability that the algorithm fails to solve the system is 1 1 and the probability that all k r repetitions independently fail is therefore (1 1 ) k. We have the following r Prob(k repetitions solve the system) = Prob(at least one repetition solves the system) = 1 Prob(no single repetition solves the problem) = 1 (1 1 r )k To ensure success probability 1 we need to find k from the following inequality using the given simplifications and taking the logarithm from both sides of the inequality. 1 (1 1 r )k 1 1 (1 1 r )k 1 log (1 1 r ) k 1 k log (1 1 r ) Where note that log (1 1 r ) is negative and when dividing the inequality with that value we need to change the direction of the inequality. The other case in analogous giving log (0.01) log (1 1 r ) k. Maximum flow LP Consider the following flow network N.
3 a 3 5 s 7 t 4 b 1. Write the problem of finding maximum flow from s to t in N as a linear program.. Write down the dual of this linear program. There should be a dual variable for each edge of the network and for each vertex other than s and t. Now, consider a general flow network. Recall the linear program formulation for a general maximum flow problem, which was shown in the class. 3. Write down the dual of this general flow linear-programming problem, using a variable y e for each edge and x u for each vertex u s, t. 4. Show that any solution to the general dual problem must satisfy the following property: for any directed path from s to t in the network, the sum of y e values along the path must be at least 1..1 Maximum flow Let s use a variable f e to denote the flow in each edge e. The objective function is to maximize the incoming flow in vertex t and in this example it has no outgoing edges, therefore it is max f s,t + f a,t. 3
4 For each edge, we have to satisfy the edge rule, giving us the following constraints. f s,a 3 f s,t 7 f s,b 4 f b,a f a,t 5 f s,a 0 f s,t 0 f s,b 0 f b,a 0 f a,t 0 In addition, the flow must satisfy the vertex rule for a and b. f s,a + f b,a f a,t = 0 f s,b f b,a = 0 In total the problem becomes max f s,t + f a,t s. t. f s,a + f b,a f a,t = 0 f s,b f b,a = 0 f s,a 3 f s,t 7 f s,b 4 f b,a f a,t 5 f s,a 0 f s,t 0 f s,b 0 f b,a 0 f a,t 0 4
5 . Dual problem To find the dual problem, let s at first reformat the primal problem to make the matrix A more explicit. In addition, let s use variable x v for each vertex rule of v and y e for each edge rule on edge e. max f s,t +f a,t s. t. f s,a +f b,a f a,t = 0 x a f s,b f b,a = 0 x b f s,a 3 y s,a f s,t 7 y s,t f s,b 4 y s,b f b,a y b,a f a,t 5 y a,t f s,t, f s,a, f s,b, f b,a, f a,t 0 It is straightforward to write out the dual problem using the columns for the matrix A that describes the constraints. In addition, all our initial variables have the non-negativity constraint meaning that all resulting constraints in the dual problem are inequalities. However, in the dual problem we only have non-negativity constraints for y e variables as these correspond to inequalities of the primal problem. min 3y s,a + 7y s,t + 4y s,b + y b,a + 5y a,t s. t. y s,t 1 x a + y s,a 0 x b + y s,b 0 x a x b + y b,a 1 x a + y a,t 0 y s,a 0 y s,t 0 y s,b 0 y b,a 0 y a,t 0 5
6 .3 Dual of the general problem The general linear program for a flow network N(G(V, E), s, t, c) can be written out using a variable f e for each e E that denotes the flow in the corresponding edge. max f e e in(t) e out(t) s. t. e E : f e c(e) v V \ {s, t} : f e f e = 0 e in(v) e out(v) e E : f e 0 f e The initial problem has V \ {s, t} equalities and E inequalities if we do not count the basic restrictions for f e 0. Hence, the dual problem will have V \{s, t} + E variables. As we have the constraint f e 0 for all initial variables then we know that all the constraints in the dual problem will be inequalities. We use a variable y e for each edge e and x u for each vertex u that is not a source or the sink. Here x u corresponds to the equality for the vertex rule of u and y e corresponds to the inequality for the edge rule of e. We assume that there are no self loops because these are not interesting for maximum flow problems. The objective function becomes minimizing c(e) y e. e E The constraints will be divided based on to which edge the derived inequality corresponds to. We will get one constraint for each edge in the graph. The main difference between them is if the edge is connected to the source or sink or not. For each edge that has one endpoint in the sink t we know that it was also used in the initial objective function, therefore the corresponding inequality will have the form... 1 or The rest of the equations will have the form The most special edges are between s and t because they are not part of any vertex rule, but participate in the objective function. Therefore, we only have some rule for the edges as y (s,t) 1 and y (t,s) 1. For each e = (u, t) E, u V we have x u + y e 1. Because initially f e is used in two constraints, one for the vertex rule of u and the other for the edge rule of e. In addition it is used in the objective function with a positive sign. Analogously, for e = (t, u) E we have x u + y e 1 because this edge is used for the vertex rule of u with a positive sign and in the corresponding edge rule. However, it was present in the objective function with the coefficient 1. Similarly, we get the special cases with e = (s, u) for u V \ {t} that gives us an inequality x u + y e 0 because we use it in the edge rule and vertex rule for u and for e = (u, s) E for u V \ {t} that gives x u + y e 0. 6
7 Finally, we have the general edges e = (u, v) where u, v V \{t, s}. All these edges are reflected by three new variables and give x u + x v + y e 0 because they are used in the vertex rules for u and v and for edge rule in e. In addition, we have y e 0 because the variables for the edges correspond to the inequalities in the original problem. Therefore, in total we obtain the following dual problem. min c(e) y e e E s. t. e = (t, s) E : y e 1 e = (s, t) E : y e 1 e = (t, u) E, u V \ {s} : x u + y e 1 e = (u, t) E, u V \ {s} : x u + y e 1 e = (s, u) E, u V \ {t} : x u + y e 0 e = (u, s) E, u V \ {t} : x u + y e 0 e = (u, v) E, u, v V \ {t, s} : x u + x v + y e 0 e E : y e 0.4 Directed path We know that none of the values y e are negative, therefore if the value of one edge is larger than one, then this is sufficient for the proof. The condition clearly holds for the only path with just one edge y (s,t) because we have the constraint y (s,t) 1. Assume that we have some path using the numbered vertices s 1... n t such that the source and sink are passed through only once. Then we consider the sum y (s,1) + y (1,) y (n,t). From the constraint x n + y (n,t) 1 we get y (n,t) 1 + x n. Therefore y (s,1) + y (1,) y (n,t) y (s,1) + y (1,) y (n 1,n) x n. Next, we can consider the constraints x n 1 + y (n 1,n) + x n 0 where y n 1,n + x n x n 1 gives y (s,1) + y (1,) y n 1,n + x n + 1 y (s,1) + y (1,) y (n,n 1) + x n We can keep using this rule on the intermediate edges (u, v) until we reach the sink and have only y s,1 + x where we have y s,1 + x 1 0, therefore y s,1 + x Following it back we get y (s,1) + y (1,) y (n,t) 1. We also have to consider the more general paths that pass through the source or the sink many times. Let s 1... n t be such path now where some 0 i n can be t or s. 7
8 We can consider this path in fragments that start and end in s or t. Each fragment is a path. Namely, we start from s and the first fragment ends when we first encounter t or s. We start the next fragment using this s or t and continue fragmenting until we reach the end of the path. The possible fragments are s s, s t, t s and t t. However, from the previous we know that the sum is greater than one for each fragment s t. In addition, trivially each path from s to t has to contain at least one fragment s t because otherwise we start from s and can only have fragments s s which means that we never reach the t in the end of the path however we have to reach it by definition. The sum of y e in each fragment is non-negative because each y e 0. Therefore, we can only look at the edges that belong to this fragment y e y e 1 e path e s t holds because the previous result for path from s to t clearly holds for this simple fragment s t. 3 Set cover and set packing You are given a set of elements U = {x 1, x,, x n } and m its subsets S i U, i = 1,,, m, such that U = S 1 S S m. Definition. A set packing of U is a collection P of subsets S i, such that any two sets in P are disjoint. The maximum set packing problem is the problem of finding a set packing with maximum number of sets. Definition. A set cover of U is a collection F of subsets S i, such that the sets in F cover U, i.e. Si FS i = U. The minimum set cover problem is the problem of finding a set cover with minimum number of sets. Example. Let n = 5 and m = 4. Take U = {1,, 3, 4, 5}, S 1 = {1,, 3}, S = {, 3, 4}, S 3 = {3, 4, 5}, S 4 = {4}. Then P = {S 1, S 4 } is a maximum set packing, and F = {S 1, S 3 } is a minimum set cover. 1. Formulate maximum set packing problem as integer linear-programming problem.. Formulate minimum set cover problem as integer linear-programming problem. 3. In the two formulations, relax constraints such that the resulting problems become linearprogramming problems, and the resulting two problems are dual. Explain why the problems are dual indeed. Assume that there are no empty sets S i. 8
9 3.1 Maximum set packing integer LP Let s i be the indicator variable that is 1 if S i is included in the packing solution. For each variable x U we need to ensure that at most one set is included in the packing, hence we get x U : m s i I(x, S i ) where I(x, S i ) = 1 if x S i and 0 otherwise. Hence, we obtain max s. t. x U : m s i m s i I(x, S i ) 1 i {1,..., m} : s i {0, 1} 3. Minimum set cover integer LP For each set S i let s i be an indicator variable such that s i = 1 if S i is in the set cover and s i = 0 otherwise. Then the problem is to minimize m s i. However, we need constraints to make sure that for each variable x j at least on set S i such that x j S i is in the cover. x U : m s i I(x, S i ) 1 where I(x, S i ) = 1 if x S i and 0 otherwise. Hence, we obtain min s. t. x U : m s i m s i I(x, S i ) 1 i {1,..., m} : s i {0, 1} 3.3 Duality Relaxing Set packing problem can be relaxed by setting 0 s i 1. However, the bound s i 1 is assured by the constraints as the constraints state that a sum of some s i has to be less than or equal to 1. Therefore we can omit this constraint and we get the following relaxed linear program. 9
10 max s. t. x U : m s i m s i I(x, S i ) 1 i {1,..., m} : s i 0 For the set cover problem, the minimization ensures that there is no reason for the variables s i to increase larger than 1 to satisfy any of the constraints. Hence we can omit s i 1 constraint as it is redundant. min s. t. x U : m s i m s i I(x, S i ) 1 i {1,..., m} : s i Explanation of duality Set packing dual Consider the case of finding the dual of the set packing solution. The dual problem has one constraint for each set S i and the objective function sums the variables for each element in U. min s. t. i {1,..., m} : n X i n X j I(x j, S i ) 1 j=1 i {1,..., n} : X i 0 We can see this as a relaxation of the following problem. 10
11 min s. t. i {1,..., m} : n X i n X j I(x j, S i ) 1 j=1 i {1,..., n} : X i {0, 1} where the constraints X i 1 are removed as this is ensured by the fact that the object is to minimize and none of the constraints require larger variables than 1 as long as the variables are non-negative. In fact, this is a set covering problem, but for different sets and elements than in the packing problem. Namely, the set of elements is U = {S 1,..., S m } and its subsets X i = {S j : x i S j }. Set covering dual Consider the dual problem of the set covering problem max s. t. i {1,..., m} : n X i n X j I(x j, S i ) 1 j=1 i {1,..., n} : X j 0 That can be seen as a relaxation of the following integer linear programming problem max s. t. i {1,..., m} : n X i n X j I(x j, S i ) 1 j=1 i {1,..., n} : X j {0, 1} This is a set packing problem for U = {S 1,..., S m } and its subsets X i = {S j : x i S j }. 11
12 4 Simplex algorithm Solve the following linear-programming problem using simplex algorithm: max x 1 + 7x + x 3 s.t. x 1 + x 3 x + x 3 4 x 1 + x 3 x 1 0, x 0, x Tableau format We write down the initial problem in the tableau format and then use this to perform the simplex algorithm on this format. The first part is the matrix A representing the left hand side of the constraints. The middle part stands for the slack variables and the last column corresponds to the right hand side of the constraints b. The last row holds the objective function information, especially it begins with c Steps of the simplex algorithm The simplex algorithm in requires us to follow two rules and do Gaussian elimination. 1. Choose a column (variable) with a positive coefficient in the last row. If no such column exists then the current solution is optimal.. For each row i where this column c has a positive entry, compute the ratio b[i] c[i] between this entry and the right hand side of the corresponding constraint. Choose the row with the smallest ratio. The chosen element is used as a pivot and then Gaussian elimination is performed. First we can choose column 1, or 3. Let s choose and the respective ratios are 3 and 4 meaning that we have to select row. We obtain the following matrix after applying the Gaussian elimination. (First row is (1)- 1 (), Second row is 1 (), Third remains the same and Fourth row is (4)- 7 ())
13 Now the only column that we can choose is the first column and based on the respective ratios, we have to take the first row. The second row remains the same, but we need to subtract the first row from the thirds and fourth row In here we can not do any more steps as all variables have negative entries in the last row. 4.3 Result The result of the Simplex algorithm can be written out from the final tableau based on the basic variables. Right now the basic variables are x 1, x and z 3 as these are the variables with a single 1 in the respective columns. We can find the values for these variables in the last column of in the row where the respective columns have 1. Hence, the optimum is obtained at x 1 = 1 and x =. We also know that z 3 = 1, but the value of the slack variable does not affect the objective function. In addition, the values of all non-basic variables is 0, hence x 3 = 0 and the value of the objective function is 15 (as represented in the bottom right entry in the tableau). 13
Review Solutions, Exam 2, Operations Research
Review Solutions, Exam 2, Operations Research 1. Prove the weak duality theorem: For any x feasible for the primal and y feasible for the dual, then... HINT: Consider the quantity y T Ax. SOLUTION: To
More informationUnderstanding the Simplex algorithm. Standard Optimization Problems.
Understanding the Simplex algorithm. Ma 162 Spring 2011 Ma 162 Spring 2011 February 28, 2011 Standard Optimization Problems. A standard maximization problem can be conveniently described in matrix form
More informationWorked Examples for Chapter 5
Worked Examples for Chapter 5 Example for Section 5.2 Construct the primal-dual table and the dual problem for the following linear programming model fitting our standard form. Maximize Z = 5 x 1 + 4 x
More informationSlide 1 Math 1520, Lecture 10
Slide 1 Math 1520, Lecture 10 In this lecture, we study the simplex method which is a powerful method for solving maximization/minimization problems developed in the late 1940 s. It has been implemented
More informationFarkas Lemma, Dual Simplex and Sensitivity Analysis
Summer 2011 Optimization I Lecture 10 Farkas Lemma, Dual Simplex and Sensitivity Analysis 1 Farkas Lemma Theorem 1. Let A R m n, b R m. Then exactly one of the following two alternatives is true: (i) x
More informationSlack Variable. Max Z= 3x 1 + 4x 2 + 5X 3. Subject to: X 1 + X 2 + X x 1 + 4x 2 + X X 1 + X 2 + 4X 3 10 X 1 0, X 2 0, X 3 0
Simplex Method Slack Variable Max Z= 3x 1 + 4x 2 + 5X 3 Subject to: X 1 + X 2 + X 3 20 3x 1 + 4x 2 + X 3 15 2X 1 + X 2 + 4X 3 10 X 1 0, X 2 0, X 3 0 Standard Form Max Z= 3x 1 +4x 2 +5X 3 + 0S 1 + 0S 2
More information(P ) Minimize 4x 1 + 6x 2 + 5x 3 s.t. 2x 1 3x 3 3 3x 2 2x 3 6
The exam is three hours long and consists of 4 exercises. The exam is graded on a scale 0-25 points, and the points assigned to each question are indicated in parenthesis within the text. Problem 1 Consider
More informationMath Models of OR: Some Definitions
Math Models of OR: Some Definitions John E. Mitchell Department of Mathematical Sciences RPI, Troy, NY 12180 USA September 2018 Mitchell Some Definitions 1 / 20 Active constraints Outline 1 Active constraints
More informationPart III: A Simplex pivot
MA 3280 Lecture 31 - More on The Simplex Method Friday, April 25, 2014. Objectives: Analyze Simplex examples. We were working on the Simplex tableau The matrix form of this system of equations is called
More information- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs
LP-Duality ( Approximation Algorithms by V. Vazirani, Chapter 12) - Well-characterized problems, min-max relations, approximate certificates - LP problems in the standard form, primal and dual linear programs
More informationDiscrete Optimization
Prof. Friedrich Eisenbrand Martin Niemeier Due Date: April 15, 2010 Discussions: March 25, April 01 Discrete Optimization Spring 2010 s 3 You can hand in written solutions for up to two of the exercises
More informationThe augmented form of this LP is the following linear system of equations:
1 Consider the following LP given in standard form: max z = 5 x_1 + 2 x_2 Subject to 3 x_1 + 2 x_2 2400 x_2 800 2 x_1 1200 x_1, x_2 >= 0 The augmented form of this LP is the following linear system of
More informationDEPARTMENT OF STATISTICS AND OPERATIONS RESEARCH OPERATIONS RESEARCH DETERMINISTIC QUALIFYING EXAMINATION. Part I: Short Questions
DEPARTMENT OF STATISTICS AND OPERATIONS RESEARCH OPERATIONS RESEARCH DETERMINISTIC QUALIFYING EXAMINATION Part I: Short Questions August 12, 2008 9:00 am - 12 pm General Instructions This examination is
More informationLesson 27 Linear Programming; The Simplex Method
Lesson Linear Programming; The Simplex Method Math 0 April 9, 006 Setup A standard linear programming problem is to maximize the quantity c x + c x +... c n x n = c T x subject to constraints a x + a x
More informationThe Simplex Algorithm
8.433 Combinatorial Optimization The Simplex Algorithm October 6, 8 Lecturer: Santosh Vempala We proved the following: Lemma (Farkas). Let A R m n, b R m. Exactly one of the following conditions is true:.
More informationChapter 1 Linear Programming. Paragraph 5 Duality
Chapter 1 Linear Programming Paragraph 5 Duality What we did so far We developed the 2-Phase Simplex Algorithm: Hop (reasonably) from basic solution (bs) to bs until you find a basic feasible solution
More information4.3 Minimizing & Mixed Constraints
Mathematics : Mattingly, Fall 6 8 4. Minimizing & Mixed Constraints So far, you have seen how to solve one type of problem: Standard Maximum. The objective function is to be maximized.. Constraints use..
More information21. Solve the LP given in Exercise 19 using the big-m method discussed in Exercise 20.
Extra Problems for Chapter 3. Linear Programming Methods 20. (Big-M Method) An alternative to the two-phase method of finding an initial basic feasible solution by minimizing the sum of the artificial
More informationDual Basic Solutions. Observation 5.7. Consider LP in standard form with A 2 R m n,rank(a) =m, and dual LP:
Dual Basic Solutions Consider LP in standard form with A 2 R m n,rank(a) =m, and dual LP: Observation 5.7. AbasisB yields min c T x max p T b s.t. A x = b s.t. p T A apple c T x 0 aprimalbasicsolutiongivenbyx
More informationYinyu Ye, MS&E, Stanford MS&E310 Lecture Note #06. The Simplex Method
The Simplex Method Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye (LY, Chapters 2.3-2.5, 3.1-3.4) 1 Geometry of Linear
More information1. (7pts) Find the points of intersection, if any, of the following planes. 3x + 9y + 6z = 3 2x 6y 4z = 2 x + 3y + 2z = 1
Math 125 Exam 1 Version 1 February 20, 2006 1. (a) (7pts) Find the points of intersection, if any, of the following planes. Solution: augmented R 1 R 3 3x + 9y + 6z = 3 2x 6y 4z = 2 x + 3y + 2z = 1 3 9
More informationLectures 6, 7 and part of 8
Lectures 6, 7 and part of 8 Uriel Feige April 26, May 3, May 10, 2015 1 Linear programming duality 1.1 The diet problem revisited Recall the diet problem from Lecture 1. There are n foods, m nutrients,
More informationStandard Form An LP is in standard form when: All variables are non-negativenegative All constraints are equalities Putting an LP formulation into sta
Chapter 4 Linear Programming: The Simplex Method An Overview of the Simplex Method Standard Form Tableau Form Setting Up the Initial Simplex Tableau Improving the Solution Calculating the Next Tableau
More informationSection Notes 8. Integer Programming II. Applied Math 121. Week of April 5, expand your knowledge of big M s and logical constraints.
Section Notes 8 Integer Programming II Applied Math 121 Week of April 5, 2010 Goals for the week understand IP relaxations be able to determine the relative strength of formulations understand the branch
More informationLecture Simplex Issues: Number of Pivots. ORIE 6300 Mathematical Programming I October 9, 2014
ORIE 6300 Mathematical Programming I October 9, 2014 Lecturer: David P. Williamson Lecture 14 Scribe: Calvin Wylie 1 Simplex Issues: Number of Pivots Question: How many pivots does the simplex algorithm
More informationSystems Analysis in Construction
Systems Analysis in Construction CB312 Construction & Building Engineering Department- AASTMT by A h m e d E l h a k e e m & M o h a m e d S a i e d 3. Linear Programming Optimization Simplex Method 135
More informationExample. 1 Rows 1,..., m of the simplex tableau remain lexicographically positive
3.4 Anticycling Lexicographic order In this section we discuss two pivoting rules that are guaranteed to avoid cycling. These are the lexicographic rule and Bland s rule. Definition A vector u R n is lexicographically
More informationCS Algorithms and Complexity
CS 50 - Algorithms and Complexity Linear Programming, the Simplex Method, and Hard Problems Sean Anderson 2/15/18 Portland State University Table of contents 1. The Simplex Method 2. The Graph Problem
More informationDual fitting approximation for Set Cover, and Primal Dual approximation for Set Cover
duality 1 Dual fitting approximation for Set Cover, and Primal Dual approximation for Set Cover Guy Kortsarz duality 2 The set cover problem with uniform costs Input: A universe U and a collection of subsets
More informationmin 4x 1 5x 2 + 3x 3 s.t. x 1 + 2x 2 + x 3 = 10 x 1 x 2 6 x 1 + 3x 2 + x 3 14
The exam is three hours long and consists of 4 exercises. The exam is graded on a scale 0-25 points, and the points assigned to each question are indicated in parenthesis within the text. If necessary,
More informationLinear Programming and the Simplex method
Linear Programming and the Simplex method Harald Enzinger, Michael Rath Signal Processing and Speech Communication Laboratory Jan 9, 2012 Harald Enzinger, Michael Rath Jan 9, 2012 page 1/37 Outline Introduction
More information1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations
The Simplex Method Most textbooks in mathematical optimization, especially linear programming, deal with the simplex method. In this note we study the simplex method. It requires basically elementary linear
More informationc) Place the Coefficients from all Equations into a Simplex Tableau, labeled above with variables indicating their respective columns
BUILDING A SIMPLEX TABLEAU AND PROPER PIVOT SELECTION Maximize : 15x + 25y + 18 z s. t. 2x+ 3y+ 4z 60 4x+ 4y+ 2z 100 8x+ 5y 80 x 0, y 0, z 0 a) Build Equations out of each of the constraints above by introducing
More informationThe dual simplex method with bounds
The dual simplex method with bounds Linear programming basis. Let a linear programming problem be given by min s.t. c T x Ax = b x R n, (P) where we assume A R m n to be full row rank (we will see in the
More informationSupplementary lecture notes on linear programming. We will present an algorithm to solve linear programs of the form. maximize.
Cornell University, Fall 2016 Supplementary lecture notes on linear programming CS 6820: Algorithms 26 Sep 28 Sep 1 The Simplex Method We will present an algorithm to solve linear programs of the form
More informationLecture 11 Linear programming : The Revised Simplex Method
Lecture 11 Linear programming : The Revised Simplex Method 11.1 The Revised Simplex Method While solving linear programming problem on a digital computer by regular simplex method, it requires storing
More informationLecture 10: Linear programming duality and sensitivity 0-0
Lecture 10: Linear programming duality and sensitivity 0-0 The canonical primal dual pair 1 A R m n, b R m, and c R n maximize z = c T x (1) subject to Ax b, x 0 n and minimize w = b T y (2) subject to
More information4.4 The Simplex Method and the Standard Minimization Problem
. The Simplex Method and the Standard Minimization Problem Question : What is a standard minimization problem? Question : How is the standard minimization problem related to the dual standard maximization
More informationSummary of the simplex method
MVE165/MMG631,Linear and integer optimization with applications The simplex method: degeneracy; unbounded solutions; starting solutions; infeasibility; alternative optimal solutions Ann-Brith Strömberg
More informationIP Cut Homework from J and B Chapter 9: 14, 15, 16, 23, 24, You wish to solve the IP below with a cutting plane technique.
IP Cut Homework from J and B Chapter 9: 14, 15, 16, 23, 24, 31 14. You wish to solve the IP below with a cutting plane technique. Maximize 4x 1 + 2x 2 + x 3 subject to 14x 1 + 10x 2 + 11x 3 32 10x 1 +
More informationLinear Programming Redux
Linear Programming Redux Jim Bremer May 12, 2008 The purpose of these notes is to review the basics of linear programming and the simplex method in a clear, concise, and comprehensive way. The book contains
More informationLinear Programming Duality P&S Chapter 3 Last Revised Nov 1, 2004
Linear Programming Duality P&S Chapter 3 Last Revised Nov 1, 2004 1 In this section we lean about duality, which is another way to approach linear programming. In particular, we will see: How to define
More informationOptimization methods NOPT048
Optimization methods NOPT048 Jirka Fink https://ktiml.mff.cuni.cz/ fink/ Department of Theoretical Computer Science and Mathematical Logic Faculty of Mathematics and Physics Charles University in Prague
More informationTechnische Universität München, Zentrum Mathematik Lehrstuhl für Angewandte Geometrie und Diskrete Mathematik. Combinatorial Optimization (MA 4502)
Technische Universität München, Zentrum Mathematik Lehrstuhl für Angewandte Geometrie und Diskrete Mathematik Combinatorial Optimization (MA 4502) Dr. Michael Ritter Problem Sheet 1 Homework Problems Exercise
More informationExam 3 Review Math 118 Sections 1 and 2
Exam 3 Review Math 118 Sections 1 and 2 This exam will cover sections 5.3-5.6, 6.1-6.3 and 7.1-7.3 of the textbook. No books, notes, calculators or other aids are allowed on this exam. There is no time
More informationCOT 6936: Topics in Algorithms! Giri Narasimhan. ECS 254A / EC 2443; Phone: x3748
COT 6936: Topics in Algorithms! Giri Narasimhan ECS 254A / EC 2443; Phone: x3748 giri@cs.fiu.edu https://moodle.cis.fiu.edu/v2.1/course/view.php?id=612 Gaussian Elimination! Solving a system of simultaneous
More informationBBM402-Lecture 20: LP Duality
BBM402-Lecture 20: LP Duality Lecturer: Lale Özkahya Resources for the presentation: https://courses.engr.illinois.edu/cs473/fa2016/lectures.html An easy LP? which is compact form for max cx subject to
More information3. Linear Programming and Polyhedral Combinatorics
Massachusetts Institute of Technology 18.433: Combinatorial Optimization Michel X. Goemans February 28th, 2013 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the introductory
More informationMATH 118 FINAL EXAM STUDY GUIDE
MATH 118 FINAL EXAM STUDY GUIDE Recommendations: 1. Take the Final Practice Exam and take note of questions 2. Use this study guide as you take the tests and cross off what you know well 3. Take the Practice
More informationLINEAR PROGRAMMING 2. In many business and policy making situations the following type of problem is encountered:
LINEAR PROGRAMMING 2 In many business and policy making situations the following type of problem is encountered: Maximise an objective subject to (in)equality constraints. Mathematical programming provides
More information3. THE SIMPLEX ALGORITHM
Optimization. THE SIMPLEX ALGORITHM DPK Easter Term. Introduction We know that, if a linear programming problem has a finite optimal solution, it has an optimal solution at a basic feasible solution (b.f.s.).
More information6.854 Advanced Algorithms
6.854 Advanced Algorithms Homework 5 Solutions 1 10 pts Define the following sets: P = positions on the results page C = clicked old results U = unclicked old results N = new results and let π : (C U)
More informationLecture 2: The Simplex method
Lecture 2 1 Linear and Combinatorial Optimization Lecture 2: The Simplex method Basic solution. The Simplex method (standardform, b>0). 1. Repetition of basic solution. 2. One step in the Simplex algorithm.
More informationLecture #21. c T x Ax b. maximize subject to
COMPSCI 330: Design and Analysis of Algorithms 11/11/2014 Lecture #21 Lecturer: Debmalya Panigrahi Scribe: Samuel Haney 1 Overview In this lecture, we discuss linear programming. We first show that the
More informationOptimization methods NOPT048
Optimization methods NOPT048 Jirka Fink https://ktiml.mff.cuni.cz/ fink/ Department of Theoretical Computer Science and Mathematical Logic Faculty of Mathematics and Physics Charles University in Prague
More informationLecture 11: Post-Optimal Analysis. September 23, 2009
Lecture : Post-Optimal Analysis September 23, 2009 Today Lecture Dual-Simplex Algorithm Post-Optimal Analysis Chapters 4.4 and 4.5. IE 30/GE 330 Lecture Dual Simplex Method The dual simplex method will
More informationthat nds a basis which is optimal for both the primal and the dual problems, given
On Finding Primal- and Dual-Optimal Bases Nimrod Megiddo (revised June 1990) Abstract. We show that if there exists a strongly polynomial time algorithm that nds a basis which is optimal for both the primal
More informationGoing from graphic solutions to algebraic
Going from graphic solutions to algebraic 2 variables: Graph constraints Identify corner points of feasible area Find which corner point has best objective value More variables: Think about constraints
More informationCHAPTER 2. The Simplex Method
CHAPTER 2 The Simplex Method In this chapter we present the simplex method as it applies to linear programming problems in standard form. 1. An Example We first illustrate how the simplex method works
More informationFoundations of Operations Research
Solved exercises for the course of Foundations of Operations Research Roberto Cordone The dual simplex method Given the following LP problem: maxz = 5x 1 +8x 2 x 1 +x 2 6 5x 1 +9x 2 45 x 1,x 2 0 1. solve
More informationMotivating examples Introduction to algorithms Simplex algorithm. On a particular example General algorithm. Duality An application to game theory
Instructor: Shengyu Zhang 1 LP Motivating examples Introduction to algorithms Simplex algorithm On a particular example General algorithm Duality An application to game theory 2 Example 1: profit maximization
More informationPrelude to the Simplex Algorithm. The Algebraic Approach The search for extreme point solutions.
Prelude to the Simplex Algorithm The Algebraic Approach The search for extreme point solutions. 1 Linear Programming-1 x 2 12 8 (4,8) Max z = 6x 1 + 4x 2 Subj. to: x 1 + x 2
More informationTRANSPORTATION PROBLEMS
Chapter 6 TRANSPORTATION PROBLEMS 61 Transportation Model Transportation models deal with the determination of a minimum-cost plan for transporting a commodity from a number of sources to a number of destinations
More informationACO Comprehensive Exam March 17 and 18, Computability, Complexity and Algorithms
1. Computability, Complexity and Algorithms (a) Let G(V, E) be an undirected unweighted graph. Let C V be a vertex cover of G. Argue that V \ C is an independent set of G. (b) Minimum cardinality vertex
More informationIntroduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs
Introduction to Mathematical Programming IE406 Lecture 10 Dr. Ted Ralphs IE406 Lecture 10 1 Reading for This Lecture Bertsimas 4.1-4.3 IE406 Lecture 10 2 Duality Theory: Motivation Consider the following
More informationIn the original knapsack problem, the value of the contents of the knapsack is maximized subject to a single capacity constraint, for example weight.
In the original knapsack problem, the value of the contents of the knapsack is maximized subject to a single capacity constraint, for example weight. In the multi-dimensional knapsack problem, additional
More informationIntroduction to Mathematical Programming IE406. Lecture 13. Dr. Ted Ralphs
Introduction to Mathematical Programming IE406 Lecture 13 Dr. Ted Ralphs IE406 Lecture 13 1 Reading for This Lecture Bertsimas Chapter 5 IE406 Lecture 13 2 Sensitivity Analysis In many real-world problems,
More information1 Review Session. 1.1 Lecture 2
1 Review Session Note: The following lists give an overview of the material that was covered in the lectures and sections. Your TF will go through these lists. If anything is unclear or you have questions
More informationLecture 10: Linear programming. duality. and. The dual of the LP in standard form. maximize w = b T y (D) subject to A T y c, minimize z = c T x (P)
Lecture 10: Linear programming duality Michael Patriksson 19 February 2004 0-0 The dual of the LP in standard form minimize z = c T x (P) subject to Ax = b, x 0 n, and maximize w = b T y (D) subject to
More informationCOMP3121/9101/3821/9801 Lecture Notes. Linear Programming
COMP3121/9101/3821/9801 Lecture Notes Linear Programming LiC: Aleks Ignjatovic THE UNIVERSITY OF NEW SOUTH WALES School of Computer Science and Engineering The University of New South Wales Sydney 2052,
More informationSensitivity Analysis
Dr. Maddah ENMG 500 /9/07 Sensitivity Analysis Changes in the RHS (b) Consider an optimal LP solution. Suppose that the original RHS (b) is changed from b 0 to b new. In the following, we study the affect
More information3. Duality: What is duality? Why does it matter? Sensitivity through duality.
1 Overview of lecture (10/5/10) 1. Review Simplex Method 2. Sensitivity Analysis: How does solution change as parameters change? How much is the optimal solution effected by changing A, b, or c? How much
More information1 Primals and Duals: Zero Sum Games
CS 124 Section #11 Zero Sum Games; NP Completeness 4/15/17 1 Primals and Duals: Zero Sum Games We can represent various situations of conflict in life in terms of matrix games. For example, the game shown
More informationIntroduction to the Simplex Algorithm Active Learning Module 3
Introduction to the Simplex Algorithm Active Learning Module 3 J. René Villalobos and Gary L. Hogg Arizona State University Paul M. Griffin Georgia Institute of Technology Background Material Almost any
More informationF 1 F 2 Daily Requirement Cost N N N
Chapter 5 DUALITY 5. The Dual Problems Every linear programming problem has associated with it another linear programming problem and that the two problems have such a close relationship that whenever
More informationCSC Design and Analysis of Algorithms. LP Shader Electronics Example
CSC 80- Design and Analysis of Algorithms Lecture (LP) LP Shader Electronics Example The Shader Electronics Company produces two products:.eclipse, a portable touchscreen digital player; it takes hours
More informationOptimization - Examples Sheet 1
Easter 0 YMS Optimization - Examples Sheet. Show how to solve the problem min n i= (a i + x i ) subject to where a i > 0, i =,..., n and b > 0. n x i = b, i= x i 0 (i =,...,n). Minimize each of the following
More informationThe Dual Simplex Algorithm
p. 1 The Dual Simplex Algorithm Primal optimal (dual feasible) and primal feasible (dual optimal) bases The dual simplex tableau, dual optimality and the dual pivot rules Classical applications of linear
More informationTopic: Primal-Dual Algorithms Date: We finished our discussion of randomized rounding and began talking about LP Duality.
CS787: Advanced Algorithms Scribe: Amanda Burton, Leah Kluegel Lecturer: Shuchi Chawla Topic: Primal-Dual Algorithms Date: 10-17-07 14.1 Last Time We finished our discussion of randomized rounding and
More informationMath Homework 3: solutions. 1. Consider the region defined by the following constraints: x 1 + x 2 2 x 1 + 2x 2 6
Math 7502 Homework 3: solutions 1. Consider the region defined by the following constraints: x 1 + x 2 2 x 1 + 2x 2 6 x 1, x 2 0. (i) Maximize 4x 1 + x 2 subject to the constraints above. (ii) Minimize
More informationOptimization Methods in Management Science
Optimization Methods in Management Science MIT 15.05 Recitation 8 TAs: Giacomo Nannicini, Ebrahim Nasrabadi At the end of this recitation, students should be able to: 1. Derive Gomory cut from fractional
More informationIE 400: Principles of Engineering Management. Simplex Method Continued
IE 400: Principles of Engineering Management Simplex Method Continued 1 Agenda Simplex for min problems Alternative optimal solutions Unboundedness Degeneracy Big M method Two phase method 2 Simplex for
More information4. Duality Duality 4.1 Duality of LPs and the duality theorem. min c T x x R n, c R n. s.t. ai Tx = b i i M a i R n
2 4. Duality of LPs and the duality theorem... 22 4.2 Complementary slackness... 23 4.3 The shortest path problem and its dual... 24 4.4 Farkas' Lemma... 25 4.5 Dual information in the tableau... 26 4.6
More informationMVE165/MMG631 Linear and integer optimization with applications Lecture 5 Linear programming duality and sensitivity analysis
MVE165/MMG631 Linear and integer optimization with applications Lecture 5 Linear programming duality and sensitivity analysis Ann-Brith Strömberg 2017 03 29 Lecture 4 Linear and integer optimization with
More information10.3 Matroids and approximation
10.3 Matroids and approximation 137 10.3 Matroids and approximation Given a family F of subsets of some finite set X, called the ground-set, and a weight function assigning each element x X a non-negative
More informationThe Simplex Method. Lecture 5 Standard and Canonical Forms and Setting up the Tableau. Lecture 5 Slide 1. FOMGT 353 Introduction to Management Science
The Simplex Method Lecture 5 Standard and Canonical Forms and Setting up the Tableau Lecture 5 Slide 1 The Simplex Method Formulate Constrained Maximization or Minimization Problem Convert to Standard
More informationCS 6820 Fall 2014 Lectures, October 3-20, 2014
Analysis of Algorithms Linear Programming Notes CS 6820 Fall 2014 Lectures, October 3-20, 2014 1 Linear programming The linear programming (LP) problem is the following optimization problem. We are given
More informationA Parametric Simplex Algorithm for Linear Vector Optimization Problems
A Parametric Simplex Algorithm for Linear Vector Optimization Problems Birgit Rudloff Firdevs Ulus Robert Vanderbei July 9, 2015 Abstract In this paper, a parametric simplex algorithm for solving linear
More information18.310A Final exam practice questions
18.310A Final exam practice questions This is a collection of practice questions, gathered randomly from previous exams and quizzes. They may not be representative of what will be on the final. In particular,
More informationIntroduction to optimization
Introduction to optimization Geir Dahl CMA, Dept. of Mathematics and Dept. of Informatics University of Oslo 1 / 24 The plan 1. The basic concepts 2. Some useful tools (linear programming = linear optimization)
More informationLecture 15 (Oct 6): LP Duality
CMPUT 675: Approximation Algorithms Fall 2014 Lecturer: Zachary Friggstad Lecture 15 (Oct 6): LP Duality Scribe: Zachary Friggstad 15.1 Introduction by Example Given a linear program and a feasible solution
More informationAM 121: Intro to Optimization
AM 121: Intro to Optimization Models and Methods Lecture 6: Phase I, degeneracy, smallest subscript rule. Yiling Chen SEAS Lesson Plan Phase 1 (initialization) Degeneracy and cycling Smallest subscript
More informationDuality of LPs and Applications
Lecture 6 Duality of LPs and Applications Last lecture we introduced duality of linear programs. We saw how to form duals, and proved both the weak and strong duality theorems. In this lecture we will
More informationIntroduction to Applied Linear Algebra with MATLAB
Sigam Series in Applied Mathematics Volume 7 Rizwan Butt Introduction to Applied Linear Algebra with MATLAB Heldermann Verlag Contents Number Systems and Errors 1 1.1 Introduction 1 1.2 Number Representation
More information5 Flows and cuts in digraphs
5 Flows and cuts in digraphs Recall that a digraph or network is a pair G = (V, E) where V is a set and E is a multiset of ordered pairs of elements of V, which we refer to as arcs. Note that two vertices
More informationPreliminaries and Complexity Theory
Preliminaries and Complexity Theory Oleksandr Romanko CAS 746 - Advanced Topics in Combinatorial Optimization McMaster University, January 16, 2006 Introduction Book structure: 2 Part I Linear Algebra
More informationAM 121 Introduction to Optimization: Models and Methods Example Questions for Midterm 1
AM 121 Introduction to Optimization: Models and Methods Example Questions for Midterm 1 Prof. Yiling Chen Fall 2018 Here are some practice questions to help to prepare for the midterm. The midterm will
More informationLinear Programming: Chapter 5 Duality
Linear Programming: Chapter 5 Duality Robert J. Vanderbei September 30, 2010 Slides last edited on October 5, 2010 Operations Research and Financial Engineering Princeton University Princeton, NJ 08544
More informationA Review of Linear Programming
A Review of Linear Programming Instructor: Farid Alizadeh IEOR 4600y Spring 2001 February 14, 2001 1 Overview In this note we review the basic properties of linear programming including the primal simplex
More informationLecture 5. x 1,x 2,x 3 0 (1)
Computational Intractability Revised 2011/6/6 Lecture 5 Professor: David Avis Scribe:Ma Jiangbo, Atsuki Nagao 1 Duality The purpose of this lecture is to introduce duality, which is an important concept
More information