LP. Kap. 17: Interior-point methods

Similar documents
Operations Research Lecture 4: Linear Programming Interior Point Method

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 10: Interior methods. Anders Forsgren. 1. Try to solve theory question 7.

Introduction to optimization

Linear and Combinatorial Optimization

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

4TE3/6TE3. Algorithms for. Continuous Optimization

Primal-Dual Interior-Point Methods

Lecture 3 Interior Point Methods and Nonlinear Optimization

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

IE 5531 Midterm #2 Solutions

CS711008Z Algorithm Design and Analysis

On well definedness of the Central Path

min 4x 1 5x 2 + 3x 3 s.t. x 1 + 2x 2 + x 3 = 10 x 1 x 2 6 x 1 + 3x 2 + x 3 14

Optimisation and Operations Research

Introduction to Nonlinear Stochastic Programming

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Interior-Point Methods

Written Examination

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

Applications of Linear Programming

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

LECTURE 10 LECTURE OUTLINE

Interior Point Methods for Convex Quadratic and Convex Nonlinear Programming

1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

Semidefinite Programming

Interior Point Methods for LP

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

10 Numerical methods for constrained problems

Lecture 3. Optimization Problems and Iterative Algorithms

Linear programming II

Interior Point Methods in Mathematical Programming

5 Handling Constraints

Summary of the simplex method

Lecture 10. Primal-Dual Interior Point Method for LP

EE364a Review Session 5

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Solving Obstacle Problems by Using a New Interior Point Algorithm. Abstract

2 Two-Point Boundary Value Problems

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization /36-725

Additional Homework Problems

Nonlinear optimization

Lecture: Algorithms for LP, SOCP and SDP

Yinyu Ye, MS&E, Stanford MS&E310 Lecture Note #06. The Simplex Method

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Math 5593 Linear Programming Week 1

Penalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way.

Lecture 11: Post-Optimal Analysis. September 23, 2009

Convex Optimization and SVM

2.098/6.255/ Optimization Methods Practice True/False Questions

MAT016: Optimization

Algorithms for nonlinear programming problems II

LINEAR AND NONLINEAR PROGRAMMING

Interior Point Methods for Mathematical Programming

IE 5531: Engineering Optimization I

Following The Central Trajectory Using The Monomial Method Rather Than Newton's Method

Optimality, Duality, Complementarity for Constrained Optimization

Linear Programming Inverse Projection Theory Chapter 3

Numerical Optimization

A new primal-dual path-following method for convex quadratic programming

Nonlinear Optimization Solvers

3.10 Lagrangian relaxation

Conic Linear Optimization and its Dual. yyye

Nonlinear Optimization: What s important?

4.6 Linear Programming duality

SF2822 Applied nonlinear optimization, final exam Saturday December

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin

Summer School: Semidefinite Optimization

Constrained Optimization and Lagrangian Duality

8. Geometric problems

TMA947/MAN280 APPLIED OPTIMIZATION

9. Geometric problems

Lecture 6: Conic Optimization September 8

CONSTRAINED NONLINEAR PROGRAMMING

IMPLEMENTING THE NEW SELF-REGULAR PROXIMITY BASED IPMS

y Ray of Half-line or ray through in the direction of y

Algorithms for constrained local optimization

More First-Order Optimization Algorithms

A Review of Linear Programming

III. Applications in convex optimization

5. Duality. Lagrangian

The Primal-Dual Algorithm P&S Chapter 5 Last Revised October 30, 2006

Lecture 18: Optimization Programming

(P ) Minimize 4x 1 + 6x 2 + 5x 3 s.t. 2x 1 3x 3 3 3x 2 2x 3 6

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Sensitivity Analysis and Duality in LP

Part IB Optimisation

Duality in Linear Programming

4TE3/6TE3. Algorithms for. Continuous Optimization

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: COMPLEMENTARITY CONSTRAINTS

Optimization (168) Lecture 7-8-9

Interior Point Methods for Nonlinear Optimization

Transcription:

LP. Kap. 17: Interior-point methods the simplex algorithm moves along the boundary of the polyhedron P of feasible solutions an alternative is interior-point methods they find a path in the interior of P, from a starting point to an optimal solution for large-scale problems interior-point methods are usually faster we consider the main idea in these methods 1 / 16

1. The barrier problem Consider the LP problem and its dual max s.t. min s.t. c T x Ax b, x O b T y A T y c, y O Introduce slack variables w in the primal and (negative) slack z in the dual, which gives 2 / 16

Primal (P): Dual (D): max s.t. min s.t. c T x Ax + w = b, x, w O b T y A T y z = c, y, z O We want to rewrite the problems such that we eliminate the constraints x, w O og y, z O, but still avoid negative values (and 0) on the variables!! This is achieved by a logarithmic barrier function, and we get the following modified primal problem 3 / 16

The barrier problem: max (P µ ) : s.t. c T x + µ j log x j + µ i log w i Ax + w = b (P µ ) is not equivalent to the original problem (P), but it is an approximation it contains a parameter µ > 0. remember: x j 0 + implies that log x j. (P µ ) is a nonlinear optimization problem interpretation/geometry: see Figure 17.1 in Vanderbei: level curves for f µ, polyhedron P, central path when µ 0. Goal: shall see that (P µ ) has a unique optimal solution x(µ) for each µ > 0, and that x(µ) x when µ 0 +, where x is the unique optimal solution of (P). (Note: w is uniquely determined by x) 4 / 16

2. Lagrange multiplier From (for instance) T. Lindstrøm, Optimering av funksjoner av flere variable, MAT1110, multivariable calculus) we have the following Lagrange multiplier rule: Theorem Assume U IR n is open, and that f, g i : U IR are functions with continuous partial derivatives (i m), and let b 1,..., b m IR. Assume that x is a local maximum (or minimum) for f on the set S = {x IR n : g i (x) = b i (i m)}, and that g 1 (x ),..., g m (x ) are linearly independent. Then there are constants λ 1,..., λ m such that m ( ) f (x ) = λ i g i (x ). λ i s are called Lagrange multipliers This is a necessary optimality condition and leads to n + m equations for finding x and λ (n + m variables). i=1 5 / 16

This can also be expressed by the Lagrange function (we redefine the function g i by g i := g i b i, such that we now consider g i (x) = 0): Then ( ) says that L(x, y) = f (x) m y i g i (x). i=1 x L(x, y) = O while the constraints g i (x ) = 0 (i m) become (where y = λ) y L(x, y) = O. These equations are called the first-order optimality conditions and a solution x is called a critical point. Are these conditions also sufficient for optimality? Consider the Hessian matrix H f (x) = [ 2 f (x) x i x j ] IR n n Note: f (x) = o(g(x) when x 0 means lim x 0 f (x)/g(x) = 0 6 / 16

Theorem 17.1 Let the g i s be linear functions, and assume x is a critical point. Then x is a local maximum if z T H f (x )z < 0 for each z O satisfying z T g i (x ) = 0 (i m). Proof: Second order Taylor formula gives f (x + z) = f (x ) + f (x ) T z + (1/2)z T H f (x )z + o( z 2 ) where z is a perturbation from the point x. To preserve feasibility z must be chosen such that x + z satisfies the constraints, i.e., z T g i (x ) = 0 (i m). But, since x is a critical point m f (x ) T z = ( λ i g i (x )) T z = 0 i=1 so the assumption (z T H f (x )z < 0 for...) and Taylor s formula give that f (x + z) f (x ), so x is a local maximum. 7 / 16

3. Lagrange applied to the barrier problem The barrier problem (P µ ): max c T x + µ j log x j + µ i log w i s.t. Ax + w = b Introduce the Lagrange function L(x, w, y) = c T x + µ j log x j + µ i log w i + y T (b Ax w) First-order optimality condition becomes L x j = c j + µ 1 x j i y ia ij = 0 (j n) L w i = µ 1 w i y i = 0 (i m) L y i = b i j a ijx j w i = 0 (i m) 8 / 16

Notation: write X for the diagonal matrix with the vector x on the diagonal. e is the vector with only 1 s. Then first order optimality conditions become, in matrix form: A T y µx 1 e y Ax + w = c = µw 1 e = b Introduce z = µx 1 e and we obtain (1.OPT) Ax + w A T y z z y = b = c = µx 1 e = µw 1 e 9 / 16

We had: Ax + w A T y z z y = b = c = µx 1 e = µw 1 e Multiply the third equation by X and the fourth with W and we get Ax + w = b ( ) A T y z = c XZe = µe YWe = µe The last two equations say: x j z j = µ (j n) and y i w i = µ (i m) which is µ-complementarity (approximative complementary slack). These are nonlinear. In total we have 2(n + m) equations and the same number of variables. 10 / 16

It is quite simple: Interior-point methods (at least this type) consist in solving the equations ( ) approximately using Newton s method for a sequence of µ s (converging to 0). 11 / 16

4. Second order information We show: if there is a solution of the opt. condition ( ), then it must be unique! We use Theorem 17.1 and consider the barrier function f (x, w) = c T x + µ j log x j + µ i log w i First derivative: f x j = c j + µ x j = 0 (j n) Second derivative: f w i = µ w i (i m) 2 f x 2 j = µ x 2 j (j n) 2 f w 2 i = µ w 2 i (i m) So the Hessian matrix is a diagonal matrix with negative diagonal elements: this matrix is negative definite. Uniqueness then follows from Teorem 17.1. 12 / 16

5. Existence Theorem 17.2 There is a solution of the barrier problem if and only if both the primal feasible region and the dual feasible region have a nonempty interior. Proof: Shall show the if -part. Assume there is a ( x, w) > O such that A x + w = b (relative interior point in the (x, w)-space), and (ȳ, z) > O with A T ȳ z = c. Let (x, w) be primal feasible. Then so z T x + ȳ T w = (A T y c) T x + ȳ T (b Ax) = b T ȳ c T x. c T x = z T x ȳ T w + b T ȳ The barrier function f becomes f (x, w) = c T x + µ j log x j + µ i log w i = j ( z jx j + µ log x j ) + i ( ȳ iw i + µ log w i ) + b T ȳ 13 / 16

The terms in each sum has the form h(v) = av + µ log v where a > 0 and 0 < v < and this function has a unique maximum in µ/a and tends to as v. This implies that the set {(x, w) : f (x, w) δ} is bounded for each δ. Let now δ = f = f ( x, w) and define the set P = {(x, w) : Ax + w = b, x O, w O, } {(x, w) : x > O, w > O, f (x, w) f }. Then P is closed. Because: P is an intersection between two closed sets; the last set is closed as f is continuous (that the domain {(x, w) : x > O, w > O} is not closed does not matter here.) Therefore P is closed and bounded, i.e., compact. P is also nonempty (it contains ( x, w)). By the extreme value theorem f attains its supremum on P, and therefore also on {(x, w) : Ax + w = b, x > O, w > O} as desired. 14 / 16

We then obtain (using an exercise saying that the dual has an interior point when the primal feasible region is bounded): Corollary 17.3 If the primal feasible region has interior points and is bounded, then for each µ > 0 there exists a unique solution of ( ). (x(µ), w(µ), y(µ), z(µ)) We then get a path (curve) p(µ) := {(x(µ), w(µ), y(µ), z(µ)) : µ > 0} in IR 2(m+n) which is called the primal-dual central path. In the primal-dual path following method one computes a sequence µ (1), µ (2),... converging to 0, and for each µ (k) one approximately solves the nonlinear system of equations ( ) using Newton s method. The corresponding sequence p(µ (k) ) will then converge towards an optimal optimal primal-dual solution. A more precise result on this convergence, and more details, are found in Chapter 18 and 19 (not syllabus). 15 / 16

Example: A problem with m = 40 and n = 100. We show l 2 -norm of the the residuals for each iteration: (primal) ρ = b Ax w; (dual) σ = c A T y + z; (compl.slack.) γ = z T x + y T w. We find an optimal solution. Iter. primal dual KS 2 189.61190 124.81236 103.89923 4 117.87500 77.59142 49.26126 6 81.95498 53.94701 30.11503 8 55.11458 36.27926 18.64561 10 30.92967 20.35951 9.75917 12 10.05169 6.61654 3.24588 14 4.37507 2.87990 1.52481 16 1.62442 1.06928 0.59844 18 0.64896 0.42718 0.25285 20 0.37908 0.24953 0.15657 22 0.14284 0.09402 0.06359 24 0.11378 0.07490 0.05191 26 0.00145 0.00095 0.00413 28 0.00000 0.00000 0.00005 30 0.00000 0.00000 0.00000 16 / 16