Interior Point Methods for LP

Similar documents
Duality Theory, Optimality Conditions

Simplex Algorithm Using Canonical Tableaus

Numerical Optimization

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

CS711008Z Algorithm Design and Analysis

Interior-Point Methods

The Primal-Dual Algorithm P&S Chapter 5 Last Revised October 30, 2006

10 Karmarkars Algorithm

CO350 Linear Programming Chapter 8: Degeneracy and Finite Termination

COMPUTATIONAL COMPLEXITY OF PARAMETRIC LINEAR PROGRAMMING +

2.098/6.255/ Optimization Methods Practice True/False Questions

min 4x 1 5x 2 + 3x 3 s.t. x 1 + 2x 2 + x 3 = 10 x 1 x 2 6 x 1 + 3x 2 + x 3 14

Linear and Combinatorial Optimization

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

Lecture: Algorithms for LP, SOCP and SDP

(P ) Minimize 4x 1 + 6x 2 + 5x 3 s.t. 2x 1 3x 3 3 3x 2 2x 3 6

Simplex method(s) for solving LPs in standard form

Operations Research Lecture 4: Linear Programming Interior Point Method

10 Numerical methods for constrained problems

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

7. Lecture notes on the ellipsoid algorithm

5. Duality. Lagrangian

Relation of Pure Minimum Cost Flow Model to Linear Programming

LP. Kap. 17: Interior-point methods

OPERATIONS RESEARCH. Linear Programming Problem

4. The Dual Simplex Method

Chapter 6 Interior-Point Approach to Linear Programming

1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations

CSCI : Optimization and Control of Networks. Review on Convex Optimization

A Brief Review on Convex Optimization

Slack Variable. Max Z= 3x 1 + 4x 2 + 5X 3. Subject to: X 1 + X 2 + X x 1 + 4x 2 + X X 1 + X 2 + 4X 3 10 X 1 0, X 2 0, X 3 0

Introduce the idea of a nondegenerate tableau and its analogy with nondenegerate vertices.

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

Linear Programming Duality

Review Solutions, Exam 2, Operations Research

Convex Optimization M2

Sensitivity Analysis

Discrete Optimization

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

LINEAR AND NONLINEAR PROGRAMMING

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Interior Point Algorithms for Constrained Convex Optimization

1 Review Session. 1.1 Lecture 2

8. Geometric problems

Standard Form An LP is in standard form when: All variables are non-negativenegative All constraints are equalities Putting an LP formulation into sta

A Strongly Polynomial Simplex Method for Totally Unimodular LP

Linear Programming and the Simplex method

Primal-Dual Interior-Point Methods

Lecture notes on the ellipsoid algorithm

Contents. 4.5 The(Primal)SimplexMethod NumericalExamplesoftheSimplexMethod

12. Interior-point methods

Conic Linear Optimization and its Dual. yyye

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin

Duality in Linear Programming

Dual Basic Solutions. Observation 5.7. Consider LP in standard form with A 2 R m n,rank(a) =m, and dual LP:

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Constraint Reduction for Linear Programs with Many Constraints

Algorithm for QP thro LCP

Linear Programming Duality P&S Chapter 3 Last Revised Nov 1, 2004

Lecture 24: August 28

Systems Analysis in Construction

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. xx xxxx 2017 xx:xx xx.

Constrained Optimization and Lagrangian Duality

January 29, Introduction to optimization and complexity. Outline. Introduction. Problem formulation. Convexity reminder. Optimality Conditions

Lecture 5. x 1,x 2,x 3 0 (1)

Convex Optimization Boyd & Vandenberghe. 5. Duality

Optimization (168) Lecture 7-8-9

Optimisation and Operations Research

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

Chapter 1. Preliminaries

The use of shadow price is an example of sensitivity analysis. Duality theory can be applied to do other kind of sensitivity analysis:

3 The Simplex Method. 3.1 Basic Solutions

Lecture 2: The Simplex method

Farkas Lemma, Dual Simplex and Sensitivity Analysis

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Written Examination

Interior Point Methods for Mathematical Programming

Developing an Algorithm for LP Preamble to Section 3 (Simplex Method)

minimize x subject to (x 2)(x 4) u,

Linear programs Optimization Geoff Gordon Ryan Tibshirani

Lecture: Duality.

Chapter 1 Linear Programming. Paragraph 5 Duality

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization /36-725

Lecture: Duality of LP, SOCP and SDP

Dr. Maddah ENMG 500 Engineering Management I 10/21/07


4.6 Linear Programming duality

Motivating examples Introduction to algorithms Simplex algorithm. On a particular example General algorithm. Duality An application to game theory

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

Optimization: Then and Now

Understanding the Simplex algorithm. Standard Optimization Problems.

8. Geometric problems

Introduction to Nonlinear Stochastic Programming

Lecture 7 Monotonicity. September 21, 2008

9. Geometric problems

Linear Programming Redux

Transcription:

11.1 Interior Point Methods for LP Katta G. Murty, IOE 510, LP, U. Of Michigan, Ann Arbor, Winter 1997. Simplex Method - A Boundary Method: Starting at an extreme point of the feasible set, the simplex method walks along its edges, until it either finds an optimum extreme point or an unbounded edge along which the objective function diverges and then it terminates. Since all its action takes place on the boundary of the feasible set, it is labeled as a boundary method. Interior Point Methods: start with a point in the (relative) interior of the feasible set, continue in the interior until they reach a near optimum solution. Each iteration here consists of 2 steps: Step 1: Determine the search direction, i.e., the direction to move at the current interior solution. 196

Step 2: Determine the step length of the move. There are 2 classes of interior point methods. Class 1: In these (affine scaling methods, Karmarkar s projective scaling method, etc.) search direction determined by the solution of a modified (approximating) problem constructed around the current interior feasible solution. Class 2: These methods apply variants of Newton s method (for solving systems of nonlinear eqs.) to the optimality conds. consisting of primal and dual feasibility and complementary slackness conds. 197

Minimizing a Linear Function Over a Ball or an Ellipsoid These problems are easy, the answer can be explicitly written down directly. The Class 1 interior point methods use these results. 1. Consider min z(x) =cx s. to (x x 0 ) T (x x 0 ) ρ 2. Case 1: c = 0. Every point in the ball is optimal. Case 2: c 0. Optimum solution is x 0 + ρ( c T / c ), obtained by moving from the center x 0, a step length of the radius ρ, in the direction of the negative gradient of the objective function. 2. Consider min z(x) =cx s. to (x x 0 ) T D 2 (x x 0 ) ρ 2 where x 0 =(x 0 j) is such that x 0 j 0 for all j, and D = diag(x 0 1,...,x 0 n). Set of feasible solutions of this is an ellipsoid with x 0 as center. An affine scaling transformation y = D 1 x 198

converts the ellipsoid into a ball with e =(1,...,1) T as center and ρ as radius. Assuming c 0 and using this transformation, it can be verified that the opt. sol. of this problem is:. x 0 ρ D2 c T Dc T 3. Consider min z(x) =cx s. to Ax = b, and (x x 0 ) T (x x 0 ) ρ 2 where A m n has rank m, andx 0 is a point in the affine space H = {x : Ax = b}. Let B = {x : (x x 0 ) T (x x 0 ) ρ 2 } be the ball. Since the center x 0 H, the intersection H B is another ball in H with radius ρ and center x 0. Optimum sol. of this problem is obtained by following procedure. Project c T = gradient of z(x)intotheaffinespaceh. This gives Pc T where P is the projection matrix corresponding to the affine space H, P = I A T (AA T ) 1 A 199

Optimum sol. is obtained by moving a step length of ρ from the center x 0 in the direction of the negative projected gradient, Pc T. Hence the opt. sol. is x 0 ρ PcT Pc T. Actually if Pc T = 0, then c is a linear combination of row vwctors of A, in this case every feasible sol. is optimal. 200

The Primal Affine Scaling Method We consider the LP in standard form: min z = cx, subject to Ax = b, x 0 where A m n and rank m. LetK denote the set of feasible slutions of this problem. A feasible sol. x is said to be an interior feasible solution, or strict feasible solution if x>0. Let x =( x j ) > 0 be the current interior feasible sol. Let X = diag{ x 1,..., x n }. The method now looks at the approximating problem obtained by replacing the constraints x 0 in original LP by the constraint x E = {x :(x x) T X 2 (x x) 1} So the approximating problem is: min z = cx s. to Ax = b, x E = {x :(x x) T X 2 (x x) ρ 2 } where ρ =1. THEOREM: If 0 <ρ 1, set of feasible solutions of approximating problem is K. The optimum solution of the approximating problem is: 201

x T XP Xc = x ρ P Xc T where P = I XA T (A X 2 A T ) 1 A X is the projection matrix. THEOREM: With ρ =1,ifx is on the boundary of K, i.e., if x j =0forsomej, then x is an opt. sol. of original LP and y T is an opt. dual sol, where y =(A X 2 A T ) 1 A X 2 c T and s = c T A T y is the dual slack vector at y. The method essentially consists of starting at an interior feasible sol. x, movingtox with ρ = 1 (or moving from x in the direction x x with a step length that is a certain percentage (typically 95%) of the maximum step length while maintaining feasibility), and repeating the whole process with the new interior feasible solution. PRIMAL AFFINE SCALING METHOD INPUT NEEDED: Problem in standard form, and an interior feasible solution. Let α be the step length fraction parameter 202

(α =0.95 typically). GENERAL STEP: Let x >0 be the current interior feasible sol. and X = diag{ x 1,..., x n }. Compute ȳ = tentative dual sol. = (A X 2 A T ) 1 A X 2 c T. Compute tentative dual slack vector s = c T A T ȳ. If s 0, z is unbounded below in original LP, terminate. Compute opt. sol. of approximating problem x = x X 2 s X s If x j =0forsomej, x optimal to the LP and ȳ opt. to the dual, terminate. Otherwise, compute step length θ =min{ X s x j s j : over j s. th. s j > 0} Takethenextpointtobe X 2 s ˆx = x 0.95θ X s Verify that cˆx = c x 0.95θ X s. Go to the next step with the new point ˆx. Convergence Results 203

Let {x k }, {y k }, {s k }, be the sequences generated by the affine scaling method. Assume that A m n has rank m, the LP has an optimum solution, and that c is not in the linear hull of the set of row vwctors of A. THEOREM: The primal objective value cx k is strictly monotone decreasing. THEOREM: The sequence {x k } converges to an optimum solution, x of the LP. THEOREM: If the LP is nondegenerate, all three sequences converge to x,y,s say, where x is optimum to primal, and y is optimal to the dual, and s is the dual slack vector corresponding to y. If the LP is degenerate, the dual sequence may not converge, counterexamples are known. However, if α 2/3, then the dual sequence has been shown to converge to an optimum dual solution even under degeneracy. 204

Primal-Dual Path Following Interior Point Methods We consider the LP in standard form: min z = cx, subject to Ax = b, x 0 where A m n and rank m. Let y =(y 1,...,y m ) T be the column vector of dual variables, and s =(s 1,...,s n ) T the column vector of slack variables. Let e =(1,...,1) T R n. Define X = diag(x 1,...,x n ), S = diag(s 1,...,s n ). From optimality conds., solving the LP is equivalent to finding a solution (x, y, s) satisfying (x, s) 0, to the system of 2n + m equations in 2n + m unknowns: F (x, y, s) = Let A T y + s c Ax b XSe =0 F = {(x, y, s) :Ax = b, A T y + s = c, (x, s) 0} F 0 = {(x, y, s) :Ax = b, A T y + s = c, (x, s) > 0} The primal-dual interior point methods generate iterates (x k,y k,s k ) 205

F 0 based on modified Newton methods for solving the square system of equations F (x, y, s) =0. 206

The Central Path This path, C is an arc in F 0 parametrized by a positive parameter τ>0. For each τ>0, the point (x τ,y τ,s τ ) Csatisfies: (x τ,s τ ) > 0and A T y τ + s τ Ax τ = c T = b x τ j sτ j = τ, j =1,...,n If τ = 0, the above eqs. define the optimality conditions for the LP. For each τ > 0, the solution (x τ,y τ,s τ ) is unique, and as τ 0 the central path converges to the center of the optimum face. Starting at an interior feasible solution (x, y, s) (i.e., a feasible solution with (x, s) > 0, these methods take steps in modified Newton directions towards points on C. For any interior feasible solution, define: Centering parameter σ [0, 1] Duality measure µ = x T s/n 207

The direction for the move is: ( x, y, s) obtained as the solution of the system of linear equations 0 A T I A 0 0 S 0 X, x y s = p where p =(0, 0, XSe + σµe). If σ = 1, the direction obtained will be a centering direction, which is a Newton direction towards the point (x µ,y µ,s µ )onc at which all pairwise products x j s j are = µ. Many algorithms choose σ from open interval (0, 1) to trade off between twin goals of reducing µ and improving centrality. General Primal-Dual Path Following Method INPUT NEEDED: Problem in standard form, initial interior feasible solution (x, y, s), i.e., one with (x, s) > 0. When (x, y, s) is current interior feasible solution, compute direction ( x, y, s) as above. Take the next point to be (ˆx, ŷ, ŝ) = (x, y, s) + α( x, y, s), where α is step length selected so that (ˆx, ŝ) remains > 0. 208