Two Lectures on the Ellipsoid Method for Solving Linear Programs
|
|
- Osborne Gray
- 5 years ago
- Views:
Transcription
1 Two Lectures on the Ellipsoid Method for Solving Linear Programs Lecture : A polynomial-time algorithm for LP Consider the general linear programming problem: δ = max cx S.T. Ax b, x R n, () where A = (a ij ) i,j Z m n is an m n integer matrix, and b = (b i ) i Z m and c = (c j ) j Z n are integer vectors. We denote by a T i the i th row of A. In the unit-cost model of computation, the cost of multiplying two integers x and y is cost(x y) =. However, in the bit-model, it is cost(x y) = l(x)+l(y), where l(x) = log( + x ) + = number of bits used to represent x in binary. For instance, consider the following code for computing x = a 2k :. x a; 2. for i =,..., k, set x x 2. Then in the unit-cost model, the cost of this code is k, while in the bit-model, it is log x 2 k l(a). Let l = max{l(a ij ), l(b i ), l(c j )}, and let L = i,j l(a ij ) + i l(b i ) + j l(c j ) be the total number of bits needed to represent the input. In the bit-model of computation, an algorithm is said to run in polynomial-time, if the number of operations ({+,,, /}) is at most poly(n, m, l) and the bit length of all numbers involved in the computation is at most poly(n, m, l). As we shall see, linear programming is polynomial-time solvable in the bit-model. These notes were taken by Khaled Elbassioni (Max-Planck-Institut für Informatik, Saarbrücken, Germany; (elbassio@mpi-sb.mpg.de)) while attending a graduate course on Linear Programming taught by Leonid Khachiyan in Fall 996 at Rutgers University.
2 The Ellipsoid method Let > 0 be a given constant. We call x an -approximate solution of () if c x δ and a T i x b i +, for i =,...,m. Assumption We know R R +, such that () has an optimal solution x in the Euclidean ball B R = {x / x R}. Let h = max{ a ij, b i, c j }, i.e., h = 2 l. Then under Assumption, the Ellipsoid method computes an -approximate solution of () in O((n + m)n 3 log ( ) ( Rhn ) arithmetic operations over O(log Rhn ) )-bit numbers. Fact Let V be the unit ball V = {x / x }, and let V = V {x / x n 0}. Then there s an ellipsoid E such that V E vol E vol V e 2(n+) 2n. Proof. Let the center of E be (0, 0,...,0, n+ ) = n+ e n, and α = n+ n (see Figure ). The equation of E is ( ) 2 x n n+ α 2 + x2 β 2 + x2 2 β x2 n β 2. At point y in Figure : x n = 0 and x 2 + x x2 n (/n+) 2 (n/n+) + 2 β =, which implies 2 β = + n 2 + 2n 2. =, and we get In the plane, an ellipse E with principal axes lengths α and β, has an area of παβ, and hence the ratio of area of E to that of the unit circle V is αβ. In 3-dimensions the volume of an ellipsoid with principal axes lengths α, β, and γ is 4 3 παβγ, while the volume of the unit ball is 4 3π, giving a ratio of αβγ. (This follows specifically from the fact that the ellipse E [ is obtained ] by α 0 transforming the circle using the linear map x = Ax, where A =, and 0 β thus vol(e) = det(a) vol(v ).) Generalizing, we get that vol(e )/ vol(v ) = αβ n. Using the inequality + x e x, valid for all x R (see Figure 2), we get α e n+ and β e 2(n 2 ), and thus vole volv = αβn e n+ e n 2(n 2 ) = e 2(n+). 2
3 x n V α n+ β y E e x y = + x x, x 2,..., x n x Fig. Fig. 2 Fact 2 Let E be an ellipsoid in R n centered at η. Given a non-zero vector a R n, consider the hyperplane π = {x / a T (x η) = 0}, and let E + and E be the two halves of E obtained by cutting E with π. Then there s an ellipsoid E such that E E vol E vol E e 2(n+) 2n. Proof. This follows by from Fact by applying a linear transformation (rotation) that maps E to the unit ball V. Such a transformation does not change the ratio of volumes or inclusion. Note that an ellipsoid E can be represented as E = {x / x = η + Qz, z }, where η R n is the center, and Q is an n n matrix. Given η, Q and a vector a R n, we can get the center η and matrix Q, corresponding to the ellipsoid E E {x / a T (x η) 0} in Fact as follows. With respect to the z-coordinates, we can write the inequality determined by π as a T (x η) = a T Qz = (Q T a) T z 0. Thus the vector e n if Fact corresponds (by an orthonormal transformation) to the vector Q T a/ Q T a in our case. Hence, the center of E in the z-coordinates is z = Q T a n+ Q T a. Let µ = QT a Q T a (which can be computed with n2 operations), then returning to the x-coordinates we get η = η Qµ (2) n + [ ( ] n Q 2 n = n 2 Q + )Qµµ n + T. (3) Note that the computation in (2), (3) can be done with n 2 operations. Recall that δ = max{cx / a T i x b i, for i =,..., m}. Let X = {x / cx δ, a T i x b i +, for i =,...,m, x R}. Assumption implies that 3
4 0 a T j ˆx b j > 0 X a η 0 t Fig. 3 0 a T j x b j < 0 x Fig. 4 ˆx X because the exact solution x B R belongs to X solution exists). It also implies that ( Fact 3 vol X vol B R h nr) n. (we assume such a Proof. Let x be an exact optimal solution, i.e., δ = cx. Let y R n be such that y x r (we will select r later). Then writing y = x + η, we have η r and hence cy = cx + cη δ c η (by the Cauchy-Schwartz s Inequality). Then, if we set r = h n, we get c η c r h nr. Similarly, a T i x b i implies a T i (x i + η) b i + ǫ. Thus the ball of radius r, centered at x, is contained in the set X and hence vol X vol Br vol B R vol B R ( r n. R) 2 Lecture 2 Recall Facts 2 and 3. Now we prove the following: Fact 4 Suppose we are given an ellipsoid E centered at η such that X E but η X. Then in O((n + m)n) operations, we can compute a new ellipsoid E such that () X E and (2) vole e 2(n+) vole. Proof. This can be done as follows (see Figure 3). Check if a T i η b i +, for i M = {,..., m}. If for i M, a T i η > b i +, then a T i x > a T i η implies x X. Thus, for any x X, we must have a T i x a T i η, i.e., a T i (x η) 0. Thus in this case, we set a = a i as the normal of the cutting hyperplane in Fact 2. Now suppose that η is an -feasible solution. Then we check whether η R. If η > R, then η T (x η) 0 for all x X. Thus we set a = η in this case. Finally, suppose that a T i η b i + for all i M and η R. If η X, then cη < δ. Hence, any x such that cx < cη < δ is not in X. Thus, cx cη, or c(x η) 0, for any -approximate solution. So we set a to c in this case, and apply Fact 2. 4
5 The Ellipsoid method starts with E 0 = B R and generates a sequence of ellipsoids E 0 = B R, E, E 2,..., E K, E K+. As long as η k (the center of E k ) is not in X, we obtain a new ellipsoid E k+ such that X E k+, and vol E k+ vol E k e 2(n+). In particular, volx vole K e 2(n+) volbr. From Fact 3, we get ( ) n volb R h volx vole K e K 2(n+) volbr, nr from which we get an upper bound on K: K N = 2n(n + )ln R nh K = O(n 2 log Rnh ). Termination Criterion: We run the algorithm for K iterations and always maintain the -feasible center with highest objective value: suppose, for some k, a T i η k b i +, for all i M and η k R, then we store η k. Suppose also that, for some l > k, a T i η l b i +, for all i M and η l R. If cη l > cη k, then we replace η k by η l. Thus the total number of operations is O((n+m)n 3 log Rnh ) = O(mn3 log Rnh ) (if m < n, we can find in O(m) time a basis and switch to a smaller problem). It can be shown also that the required accuracy is O(log Rnh ). How to round and -approximate solution to an exact one Recall that we assume the matrices A, b, and c to be integral. Given an integral matrix A, denote by (A) = max{ det B / B is a square submatrix of A}. [ ] 7 3 For e.g. if A =, then (A) = 7. Recall that for a vector v R 3 2 n, v = max i { v i }. Lemma If problem (), with integer matrices A and b, has an optimal solution, then it has an optimal solution x such that x n b (A). Proof. By the fundamental Theorem of Linear Programming, there exists M M = {,...,m} such that a T i x = b i, for all i M. Then by Cramer s rule x j = j, where is a non-zero subdeterminant of A, and j is a subdeterminant of [A b]. Since is an integer, x j j. But j = i b i i, where i is a subdeterminant of A. Using b i b and i (A), we get that x j n b (A). Thus we can set R = hn 3/2 (A) to guarantee that x R (using the fact that n 2 ). 5
6 Lemma 2 Suppose that δ is finite, then δ = t s, where t and s are integers such that s (A). Indeed, δ = c x + c 2 x c n x n = c + + c n n. Lemma 3 Consider the system Ax b, x R n, (4) and let 0 = (n+2) (A). If there is an 0-approximate solution to (4) (i.e. a T i x b i + 0, for i M, is feasible), then (4) has an exact solution. Proof. Consider = min S.T. a T i x b i +, i M. (5) Suppose that Ax b is infeasible. Then ǫ > 0, and hence by Lemma 2, ǫ = t s for some integers t and s, where s ([A e]) (n + ) (A). This gives s (n+) (A). Thus using = 0 = (n+2) (A) will make (4) feasible. Now suppose that a T i x b i + 0, i M. We can compute an exact solution using mn 2 operations as follows. Let I 0 = {i M / a T i x b i 0 }. Then a T j x < b j 0 for j M \I 0. The system a T i x = b i, i I 0, is solvable, by Lemma 3, because x is an 0 -approximate solution ([ for it (the ]) system can be written as a T i x b i, a T i x b A i, i I 0, and I0 (A A I0 ) (A), where I0 A I0 is the submatrix of A with rows indexed by I 0 ). Let ˆx be an exact solution for this system, i.e., a T i ˆx = b i, for i I 0. If x does not satisfy the original system, then we proceed as follows. Define x(t) = ( t) x + tˆx, for 0 t. Then for i I 0, we have a T i x(t) b i = ( t)a T i x + ta T i ˆx ( t)b i tb i = ( t)[a T i x b i] + t[a T i ˆx b i] ( t) a T i x b i + t a T i ˆx b i ( t) 0 0. Thus x(t) satisfies the constraints in I 0 for all t [0, ]. Now we obtain a new 0 -approximate solution with a wider set of tight constraints, by choosing t as follows (see Figure 4): { } t = min j M\I 0 at j x b j a T j ˆx at j x / at j ˆx b j > 0 Let j min be such a minimizer in M \I 0. Then a T j min x(t) b jmin = ( t)(a T j min x b jmin )+t(a T j minˆx b jmin ) = 0. We add j min to I 0, replace x by x(t), and repeat the procedure again. So this way, we do at most n iterations of mn 2 operations. 6
7 each (solving a system of linear equations), but it can be done in O(mn 2 ) in total. To consider the objective function: We use, instead of 0, = 4n 5/2 3 (A) c. We do the same procedure as before without caring about the objective function. We claim that the result is optimal. Indeed, we know that a T i ˆx = b i and a T i x = b i +η i for i I 0, where η. Thus a T i (ˆx x) = η i, for i I 0, and setting ˆx x = y, we get (again from Cramer s rule) that y n η (A), and thus ˆx x n (A). This gives cˆx c x c ˆx x c n 3/2 (A). At the end of the rounding algorithm we obtain x such that Ax b and cx c x c n 5/2 (A) (each iteration of the procedure contributes at most c n 3/2 (A) to this difference). Thus cx c x c n 5/2 (A) δ c n 5/2 (A) δ 2 c n 5/2 (A) = δ 2 2 (A). But cx = t s, where t, s Z and s (A) (as in Lemma 2), and δ = t s, s (A). Since cx δ, we have cx δ 2 2 (A). On the other hand, if cx δ, then cx δ = t s t s = t s ts ss ss 2 (A) > 2 2 (A). Thus we must have cx = δ. Summary Using = 4n 5/2 3 (A) c and R = n3/2 h (A), the Ellipsoid method requires O(mn 3 log Rhn ) = O(mn 3 log(hn (A))) operations. Denote by L = L(A, b, c) the number of bits required to write down the problem, and let L(A) = i,j log( + a ij ). We note that (A) 2 L(A) since (A) i,j (+ a ij ). Also n 2 L(A) and h 2 L(A,b,c). Thus the total number of operations required by the Ellipsoid method is O(mn 3 L). 7
A strongly polynomial algorithm for linear systems having a binary solution
A strongly polynomial algorithm for linear systems having a binary solution Sergei Chubanov Institute of Information Systems at the University of Siegen, Germany e-mail: sergei.chubanov@uni-siegen.de 7th
More informationHermite normal form: Computation and applications
Integer Points in Polyhedra Gennady Shmonin Hermite normal form: Computation and applications February 24, 2009 1 Uniqueness of Hermite normal form In the last lecture, we showed that if B is a rational
More informationComplexity of linear programming: outline
Complexity of linear programming: outline I Assessing computational e ciency of algorithms I Computational e ciency of the Simplex method I Ellipsoid algorithm for LP and its computational e ciency IOE
More informationTheory and Internet Protocols
Game Lecture 2: Linear Programming and Zero Sum Nash Equilibrium Xiaotie Deng AIMS Lab Department of Computer Science Shanghai Jiaotong University September 26, 2016 1 2 3 4 Standard Form (P) Outline
More informationLINEAR PROGRAMMING III
LINEAR PROGRAMMING III ellipsoid algorithm combinatorial optimization matrix games open problems Lecture slides by Kevin Wayne Last updated on 7/25/17 11:09 AM LINEAR PROGRAMMING III ellipsoid algorithm
More informationCS 6820 Fall 2014 Lectures, October 3-20, 2014
Analysis of Algorithms Linear Programming Notes CS 6820 Fall 2014 Lectures, October 3-20, 2014 1 Linear programming The linear programming (LP) problem is the following optimization problem. We are given
More informationANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3
ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any
More informationCS675: Convex and Combinatorial Optimization Spring 2018 The Ellipsoid Algorithm. Instructor: Shaddin Dughmi
CS675: Convex and Combinatorial Optimization Spring 2018 The Ellipsoid Algorithm Instructor: Shaddin Dughmi History and Basics Originally developed in the mid 70s by Iudin, Nemirovski, and Shor for use
More informationCombinatorial Optimization Spring Term 2015 Rico Zenklusen. 2 a = ( 3 2 ) 1 E(a, A) = E(( 3 2 ), ( 4 0
3 2 a = ( 3 2 ) 1 E(a, A) = E(( 3 2 ), ( 4 0 0 1 )) 0 0 1 2 3 4 5 Figure 9: An example of an axis parallel ellipsoid E(a, A) in two dimensions. Notice that the eigenvectors of A correspond to the axes
More information15.081J/6.251J Introduction to Mathematical Programming. Lecture 18: The Ellipsoid method
15.081J/6.251J Introduction to Mathematical Programming Lecture 18: The Ellipsoid method 1 Outline Efficient algorithms and computational complexity Slide 1 The key geometric result behind the ellipsoid
More informationA polynomial relaxation-type algorithm for linear programming
A polynomial relaxation-type algorithm for linear programming Sergei Chubanov Institute of Information Systems at the University of Siegen, Germany e-mail: sergei.chubanov@uni-siegen.de 6th February 20
More information6.854J / J Advanced Algorithms Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.854J / 18.415J Advanced Algorithms Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 18.415/6.854 Advanced
More informationThe Ellipsoid (Kachiyan) Method
Yinyu Ye, MS&E, Stanford MS&E310 Lecture Note: Ellipsoid Method 1 The Ellipsoid (Kachiyan) Method Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.
More informationGrothendieck s Inequality
Grothendieck s Inequality Leqi Zhu 1 Introduction Let A = (A ij ) R m n be an m n matrix. Then A defines a linear operator between normed spaces (R m, p ) and (R n, q ), for 1 p, q. The (p q)-norm of A
More informationORIE 6334 Spectral Graph Theory October 25, Lecture 18
ORIE 6334 Spectral Graph Theory October 25, 2016 Lecturer: David P Williamson Lecture 18 Scribe: Venus Lo 1 Max Flow in Undirected Graphs 11 Max flow We start by reviewing the maximum flow problem We are
More informationThe Ellipsoid Algorithm
The Ellipsoid Algorithm John E. Mitchell Department of Mathematical Sciences RPI, Troy, NY 12180 USA 9 February 2018 Mitchell The Ellipsoid Algorithm 1 / 28 Introduction Outline 1 Introduction 2 Assumptions
More informationRECAP How to find a maximum matching?
RECAP How to find a maximum matching? First characterize maximum matchings A maximal matching cannot be enlarged by adding another edge. A maximum matching of G is one of maximum size. Example. Maximum
More informationTopics in Theoretical Computer Science: An Algorithmist's Toolkit Fall 2007
MIT OpenCourseWare http://ocw.mit.edu 18.409 Topics in Theoretical Computer Science: An Algorithmist's Toolkit Fall 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationAlgorithms and Theory of Computation. Lecture 13: Linear Programming (2)
Algorithms and Theory of Computation Lecture 13: Linear Programming (2) Xiaohui Bei MAS 714 September 25, 2018 Nanyang Technological University MAS 714 September 25, 2018 1 / 15 LP Duality Primal problem
More informationA primal-simplex based Tardos algorithm
A primal-simplex based Tardos algorithm Shinji Mizuno a, Noriyoshi Sukegawa a, and Antoine Deza b a Graduate School of Decision Science and Technology, Tokyo Institute of Technology, 2-12-1-W9-58, Oo-Okayama,
More informationWeek 8. 1 LP is easy: the Ellipsoid Method
Week 8 1 LP is easy: the Ellipsoid Method In 1979 Khachyan proved that LP is solvable in polynomial time by a method of shrinking ellipsoids. The running time is polynomial in the number of variables n,
More information47-831: Advanced Integer Programming Lecturer: Amitabh Basu Lecture 2 Date: 03/18/2010
47-831: Advanced Integer Programming Lecturer: Amitabh Basu Lecture Date: 03/18/010 We saw in the previous lecture that a lattice Λ can have many bases. In fact, if Λ is a lattice of a subspace L with
More information15-850: Advanced Algorithms CMU, Spring 2017 Lecture #17: The Ellipsoid Algorithm March 3, 2017
15-850: Advanced Algorithms CMU, Spring 2017 Lecture #17: The Ellipsoid Algorithm March 3, 2017 Lecturer: Anupam Gupta Scribe: Benjamin Berg, David K. Isenberg In this lecture, we discuss some polynomial-time
More information15.083J/6.859J Integer Optimization. Lecture 10: Solving Relaxations
15.083J/6.859J Integer Optimization Lecture 10: Solving Relaxations 1 Outline The key geometric result behind the ellipsoid method Slide 1 The ellipsoid method for the feasibility problem The ellipsoid
More informationENGR-1100 Introduction to Engineering Analysis. Lecture 21. Lecture outline
ENGR-1100 Introduction to Engineering Analysis Lecture 21 Lecture outline Procedure (algorithm) for finding the inverse of invertible matrix. Investigate the system of linear equation and invertibility
More informationMatrix Operations: Determinant
Matrix Operations: Determinant Determinants Determinants are only applicable for square matrices. Determinant of the square matrix A is denoted as: det(a) or A Recall that the absolute value of the determinant
More informationENGR-1100 Introduction to Engineering Analysis. Lecture 21
ENGR-1100 Introduction to Engineering Analysis Lecture 21 Lecture outline Procedure (algorithm) for finding the inverse of invertible matrix. Investigate the system of linear equation and invertibility
More information{ move v ars to left, consts to right { replace = by t wo and constraints Ax b often nicer for theory Ax = b good for implementations. { A invertible
Finish remarks on min-cost ow. Strongly polynomial algorithms exist. { Tardos 1985 { minimum mean-cost cycle { reducing -optimality { \xing" arcs of very high reduced cost { best running running time roughly
More information7. Lecture notes on the ellipsoid algorithm
Massachusetts Institute of Technology Michel X. Goemans 18.433: Combinatorial Optimization 7. Lecture notes on the ellipsoid algorithm The simplex algorithm was the first algorithm proposed for linear
More information1 The linear algebra of linear programs (March 15 and 22, 2015)
1 The linear algebra of linear programs (March 15 and 22, 2015) Many optimization problems can be formulated as linear programs. The main features of a linear program are the following: Variables are real
More informationPolynomiality of Linear Programming
Chapter 10 Polynomiality of Linear Programming In the previous section, we presented the Simplex Method. This method turns out to be very efficient for solving linear programmes in practice. While it is
More informationLecture 2: Lattices and Bases
CSE 206A: Lattice Algorithms and Applications Spring 2007 Lecture 2: Lattices and Bases Lecturer: Daniele Micciancio Scribe: Daniele Micciancio Motivated by the many applications described in the first
More informationLecture 23: 6.1 Inner Products
Lecture 23: 6.1 Inner Products Wei-Ta Chu 2008/12/17 Definition An inner product on a real vector space V is a function that associates a real number u, vwith each pair of vectors u and v in V in such
More informationAlgorithmic Convex Geometry
Algorithmic Convex Geometry August 2011 2 Contents 1 Overview 5 1.1 Learning by random sampling.................. 5 2 The Brunn-Minkowski Inequality 7 2.1 The inequality.......................... 8 2.1.1
More informationLecture 13: Linear programming I
A Theorist s Toolkit (CMU 18-859T, Fall 013) Lecture 13: Linear programming I October 1, 013 Lecturer: Ryan O Donnell Scribe: Yuting Ge A Generic Reference: Ryan recommends a very nice book Understanding
More informationJim Lambers MAT 610 Summer Session Lecture 2 Notes
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the
More informationNew Characterizations in Turnstile Streams with Applications
New Characterizations in Turnstile Streams with Applications Yuqing Ai 1, Wei Hu 1, Yi Li 2, and David P. Woodruff 3 1 The Institute for Theoretical Computer Science (ITCS), Institute for Interdisciplinary
More informationLMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009
LMI MODELLING 4. CONVEX LMI MODELLING Didier HENRION LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ Universidad de Valladolid, SP March 2009 Minors A minor of a matrix F is the determinant of a submatrix
More informationMathematical Optimisation, Chpt 2: Linear Equations and inequalities
Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson
More informationIntroduction to Algorithms
Introduction to Algorithms 6.046J/18.401J/SMA5503 Lecture 18 Prof. Erik Demaine Negative-weight cycles Recall: If a graph G = (V, E) contains a negativeweight cycle, then some shortest paths may not exist.
More informationMATRICES. a m,1 a m,n A =
MATRICES Matrices are rectangular arrays of real or complex numbers With them, we define arithmetic operations that are generalizations of those for real and complex numbers The general form a matrix of
More informationA notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations
A notion of for Convex, Semidefinite and Extended Formulations Marcel de Carli Silva Levent Tunçel April 26, 2018 A vector in R n is integral if each of its components is an integer, A vector in R n is
More informationLecture notes for quantum semidefinite programming (SDP) solvers
CMSC 657, Intro to Quantum Information Processing Lecture on November 15 and 0, 018 Fall 018, University of Maryland Prepared by Tongyang Li, Xiaodi Wu Lecture notes for quantum semidefinite programming
More information12x + 18y = 30? ax + by = m
Math 2201, Further Linear Algebra: a practical summary. February, 2009 There are just a few themes that were covered in the course. I. Algebra of integers and polynomials. II. Structure theory of one endomorphism.
More informationDeterminants Chapter 3 of Lay
Determinants Chapter of Lay Dr. Doreen De Leon Math 152, Fall 201 1 Introduction to Determinants Section.1 of Lay Given a square matrix A = [a ij, the determinant of A is denoted by det A or a 11 a 1j
More informationA Strongly Polynomial Simplex Method for Totally Unimodular LP
A Strongly Polynomial Simplex Method for Totally Unimodular LP Shinji Mizuno July 19, 2014 Abstract Kitahara and Mizuno get new bounds for the number of distinct solutions generated by the simplex method
More information3. Linear Programming and Polyhedral Combinatorics
Massachusetts Institute of Technology 18.433: Combinatorial Optimization Michel X. Goemans February 28th, 2013 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the introductory
More informationLecture # 3 Orthogonal Matrices and Matrix Norms. We repeat the definition an orthogonal set and orthornormal set.
Lecture # 3 Orthogonal Matrices and Matrix Norms We repeat the definition an orthogonal set and orthornormal set. Definition A set of k vectors {u, u 2,..., u k }, where each u i R n, is said to be an
More informationInteger Programming ISE 418. Lecture 12. Dr. Ted Ralphs
Integer Programming ISE 418 Lecture 12 Dr. Ted Ralphs ISE 418 Lecture 12 1 Reading for This Lecture Nemhauser and Wolsey Sections II.2.1 Wolsey Chapter 9 ISE 418 Lecture 12 2 Generating Stronger Valid
More informationc 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp , March
SIAM REVIEW. c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp. 93 97, March 1995 008 A UNIFIED PROOF FOR THE CONVERGENCE OF JACOBI AND GAUSS-SEIDEL METHODS * ROBERTO BAGNARA Abstract.
More informationLecture 3. 1 Polynomial-time algorithms for the maximum flow problem
ORIE 633 Network Flows August 30, 2007 Lecturer: David P. Williamson Lecture 3 Scribe: Gema Plaza-Martínez 1 Polynomial-time algorithms for the maximum flow problem 1.1 Introduction Let s turn now to considering
More informationLecture 1 Introduction
L. Vandenberghe EE236A (Fall 2013-14) Lecture 1 Introduction course overview linear optimization examples history approximate syllabus basic definitions linear optimization in vector and matrix notation
More informationLecture 23 Branch-and-Bound Algorithm. November 3, 2009
Branch-and-Bound Algorithm November 3, 2009 Outline Lecture 23 Modeling aspect: Either-Or requirement Special ILPs: Totally unimodular matrices Branch-and-Bound Algorithm Underlying idea Terminology Formal
More informationBBM402-Lecture 20: LP Duality
BBM402-Lecture 20: LP Duality Lecturer: Lale Özkahya Resources for the presentation: https://courses.engr.illinois.edu/cs473/fa2016/lectures.html An easy LP? which is compact form for max cx subject to
More informationAppendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS
Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS Here we consider systems of linear constraints, consisting of equations or inequalities or both. A feasible solution
More informationOptimization (168) Lecture 7-8-9
Optimization (168) Lecture 7-8-9 Jesús De Loera UC Davis, Mathematics Wednesday, April 2, 2012 1 DEGENERACY IN THE SIMPLEX METHOD 2 DEGENERACY z =2x 1 x 2 + 8x 3 x 4 =1 2x 3 x 5 =3 2x 1 + 4x 2 6x 3 x 6
More informationA matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and
Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.
More information(K + L)(c x) = K(c x) + L(c x) (def of K + L) = K( x) + K( y) + L( x) + L( y) (K, L are linear) = (K L)( x) + (K L)( y).
Exercise 71 We have L( x) = x 1 L( v 1 ) + x 2 L( v 2 ) + + x n L( v n ) n = x i (a 1i w 1 + a 2i w 2 + + a mi w m ) i=1 ( n ) ( n ) ( n ) = x i a 1i w 1 + x i a 2i w 2 + + x i a mi w m i=1 Therefore y
More informationCSC Linear Programming and Combinatorial Optimization Lecture 8: Ellipsoid Algorithm
CSC2411 - Linear Programming and Combinatorial Optimization Lecture 8: Ellipsoid Algorithm Notes taken by Shizhong Li March 15, 2005 Summary: In the spring of 1979, the Soviet mathematician L.G.Khachian
More informationLecture #21. c T x Ax b. maximize subject to
COMPSCI 330: Design and Analysis of Algorithms 11/11/2014 Lecture #21 Lecturer: Debmalya Panigrahi Scribe: Samuel Haney 1 Overview In this lecture, we discuss linear programming. We first show that the
More information3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions
A. LINEAR ALGEBRA. CONVEX SETS 1. Matrices and vectors 1.1 Matrix operations 1.2 The rank of a matrix 2. Systems of linear equations 2.1 Basic solutions 3. Vector spaces 3.1 Linear dependence and independence
More information16.1 Min-Cut as an LP
600.469 / 600.669 Approximation Algorithms Lecturer: Michael Dinitz Topic: LPs as Metrics: Min Cut and Multiway Cut Date: 4//5 Scribe: Gabriel Kaptchuk 6. Min-Cut as an LP We recall the basic definition
More informationA matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and
Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.
More informationLecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?
KTH ROYAL INSTITUTE OF TECHNOLOGY Norms for vectors and matrices Why? Lecture 5 Ch. 5, Norms for vectors and matrices Emil Björnson/Magnus Jansson/Mats Bengtsson April 27, 2016 Problem: Measure size of
More informationLinear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.
Matrices Operations Linear Algebra Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0 The rectangular array 1 2 1 4 3 4 2 6 1 3 2 1 in which the
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationLinear Systems and Matrices
Department of Mathematics The Chinese University of Hong Kong 1 System of m linear equations in n unknowns (linear system) a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.......
More informationMATHEMATICAL ENGINEERING TECHNICAL REPORTS
MATHEMATICAL ENGINEERING TECHNICAL REPORTS Combinatorial Relaxation Algorithm for the Entire Sequence of the Maximum Degree of Minors in Mixed Polynomial Matrices Shun SATO (Communicated by Taayasu MATSUO)
More informationLinear Programming Redux
Linear Programming Redux Jim Bremer May 12, 2008 The purpose of these notes is to review the basics of linear programming and the simplex method in a clear, concise, and comprehensive way. The book contains
More informationOnline Learning Summer School Copenhagen 2015 Lecture 1
Online Learning Summer School Copenhagen 2015 Lecture 1 Shai Shalev-Shwartz School of CS and Engineering, The Hebrew University of Jerusalem Online Learning Shai Shalev-Shwartz (Hebrew U) OLSS Lecture
More informationMath 240 Calculus III
The Calculus III Summer 2015, Session II Wednesday, July 8, 2015 Agenda 1. of the determinant 2. determinants 3. of determinants What is the determinant? Yesterday: Ax = b has a unique solution when A
More informationLinear Algebraic Equations
Linear Algebraic Equations 1 Fundamentals Consider the set of linear algebraic equations n a ij x i b i represented by Ax b j with [A b ] [A b] and (1a) r(a) rank of A (1b) Then Axb has a solution iff
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J Olver 8 Numerical Computation of Eigenvalues In this part, we discuss some practical methods for computing eigenvalues and eigenvectors of matrices Needless to
More informationLet x be an approximate solution for Ax = b, e.g., obtained by Gaussian elimination. Let x denote the exact solution. Call. r := b A x.
ESTIMATION OF ERROR Let x be an approximate solution for Ax = b, e.g., obtained by Gaussian elimination. Let x denote the exact solution. Call the residual for x. Then r := b A x r = b A x = Ax A x = A
More informationLecture 14 Ellipsoid method
S. Boyd EE364 Lecture 14 Ellipsoid method idea of localization methods bisection on R center of gravity algorithm ellipsoid method 14 1 Localization f : R n R convex (and for now, differentiable) problem:
More informationChapter 1 Linear Programming. Paragraph 5 Duality
Chapter 1 Linear Programming Paragraph 5 Duality What we did so far We developed the 2-Phase Simplex Algorithm: Hop (reasonably) from basic solution (bs) to bs until you find a basic feasible solution
More informationMath Matrix Algebra
Math 44 - Matrix Algebra Review notes - (Alberto Bressan, Spring 7) sec: Orthogonal diagonalization of symmetric matrices When we seek to diagonalize a general n n matrix A, two difficulties may arise:
More informationTRANSPORTATION PROBLEMS
Chapter 6 TRANSPORTATION PROBLEMS 61 Transportation Model Transportation models deal with the determination of a minimum-cost plan for transporting a commodity from a number of sources to a number of destinations
More informationx 1 + x 2 2 x 1 x 2 1 x 2 2 min 3x 1 + 2x 2
Lecture 1 LPs: Algebraic View 1.1 Introduction to Linear Programming Linear programs began to get a lot of attention in 1940 s, when people were interested in minimizing costs of various systems while
More informationLinear Programming, Lecture 4
Linear Programming, Lecture 4 Corbett Redden October 3, 2016 Simplex Form Conventions Examples Simplex Method To run the simplex method, we start from a Linear Program (LP) in the following standard simplex
More informationLecture notes on the ellipsoid algorithm
Massachusetts Institute of Technology Handout 1 18.433: Combinatorial Optimization May 14th, 007 Michel X. Goemans Lecture notes on the ellipsoid algorithm The simplex algorithm was the first algorithm
More informationApproximation Basics
Approximation Basics, Concepts, and Examples Xiaofeng Gao Department of Computer Science and Engineering Shanghai Jiao Tong University, P.R.China Fall 2012 Special thanks is given to Dr. Guoqiang Li for
More information1 Locally computable randomized encodings
CSG399: Gems of Theoretical Computer Science Lectures 3-4 Feb 20-24, 2009 Instructor: Emanuele Viola Scribe: Eric Miles Cryptography in constant depth: II & III Locally computable randomized encodings
More informationIE 5531: Engineering Optimization I
IE 5531: Engineering Optimization I Lecture 15: Nonlinear optimization Prof. John Gunnar Carlsson November 1, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I November 1, 2010 1 / 24
More informationStochastic Design Criteria in Linear Models
AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework
More informationThe Shortest Vector Problem (Lattice Reduction Algorithms)
The Shortest Vector Problem (Lattice Reduction Algorithms) Approximation Algorithms by V. Vazirani, Chapter 27 - Problem statement, general discussion - Lattices: brief introduction - The Gauss algorithm
More informationOn the Readability of Monotone Boolean Formulae
On the Readability of Monotone Boolean Formulae Khaled Elbassioni 1, Kazuhisa Makino 2, Imran Rauf 1 1 Max-Planck-Institut für Informatik, Saarbrücken, Germany {elbassio,irauf}@mpi-inf.mpg.de 2 Department
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More informationLecture 5: Computational Complexity
Lecture 5: Computational Complexity (3 units) Outline Computational complexity Decision problem, Classes N P and P. Polynomial reduction and Class N PC P = N P or P = N P? 1 / 22 The Goal of Computational
More informationDO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO
QUESTION BOOKLET EECS 227A Fall 2009 Midterm Tuesday, Ocotober 20, 11:10-12:30pm DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO You have 80 minutes to complete the midterm. The midterm consists
More informationLecture 4: Algebra, Geometry, and Complexity of the Simplex Method. Reading: Sections 2.6.4, 3.5,
Lecture 4: Algebra, Geometry, and Complexity of the Simplex Method Reading: Sections 2.6.4, 3.5, 10.2 10.5 1 Summary of the Phase I/Phase II Simplex Method We write a typical simplex tableau as z x 1 x
More information3. Linear Programming and Polyhedral Combinatorics
Massachusetts Institute of Technology 18.453: Combinatorial Optimization Michel X. Goemans April 5, 2017 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the introductory
More informationInput: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function
Linear programming Input: System of inequalities or equalities over the reals R A linear cost function Output: Value for variables that minimizes cost function Example: Minimize 6x+4y Subject to 3x + 2y
More informationSMSTC (2017/18) Geometry and Topology 2.
SMSTC (2017/18) Geometry and Topology 2 Lecture 1: Differentiable Functions and Manifolds in R n Lecturer: Diletta Martinelli (Notes by Bernd Schroers) a wwwsmstcacuk 11 General remarks In this lecture
More informationGeneralized incidence theorems, homogeneous forms and sum-product estimates in finite fields arxiv: v2 [math.
Generalized incidence theorems, homogeneous forms and sum-product estimates in finite fields arxiv:0801.0728v2 [math.co] 31 Mar 2008 David Covert, Derrick Hart, Alex Iosevich, Doowon Koh, and Misha Rudnev
More informationDecision Procedures An Algorithmic Point of View
An Algorithmic Point of View ILP References: Integer Programming / Laurence Wolsey Deciding ILPs with Branch & Bound Intro. To mathematical programming / Hillier, Lieberman Daniel Kroening and Ofer Strichman
More informationSupplementary lecture notes on linear programming. We will present an algorithm to solve linear programs of the form. maximize.
Cornell University, Fall 2016 Supplementary lecture notes on linear programming CS 6820: Algorithms 26 Sep 28 Sep 1 The Simplex Method We will present an algorithm to solve linear programs of the form
More informationLinear programming: theory, algorithms and applications
Linear programming: theory, algorithms and applications illes@math.bme.hu Department of Differential Equations Budapest 2014/2015 Fall Vector spaces A nonempty set L equipped with addition and multiplication
More information15-850: Advanced Algorithms CMU, Fall 2018 HW #4 (out October 17, 2018) Due: October 28, 2018
15-850: Advanced Algorithms CMU, Fall 2018 HW #4 (out October 17, 2018) Due: October 28, 2018 Usual rules. :) Exercises 1. Lots of Flows. Suppose you wanted to find an approximate solution to the following
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More information