Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Similar documents
Interior-Point Methods

Lecture 17: Primal-dual interior-point methods part II

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Primal-Dual Interior-Point Methods

Interior Point Algorithms for Constrained Convex Optimization

Linear Programming: Simplex

Operations Research Lecture 4: Linear Programming Interior Point Method

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization /36-725

12. Interior-point methods

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Interior Point Methods for LP

CS711008Z Algorithm Design and Analysis

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Lecture 14: October 17

Duality revisited. Javier Peña Convex Optimization /36-725

Barrier Method. Javier Peña Convex Optimization /36-725

Lecture 15: October 15

Lecture 24: August 28

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method


A Simpler and Tighter Redundant Klee-Minty Construction

Primal/Dual Decomposition Methods

Lecture: Introduction to LP, SDP and SOCP

CSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization

10 Numerical methods for constrained problems

2.098/6.255/ Optimization Methods Practice True/False Questions

Interior-Point Methods for Linear Optimization

An interior-point gradient method for large-scale totally nonnegative least squares problems

Newton s Method. Javier Peña Convex Optimization /36-725

Convex Optimization and Support Vector Machine

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

Linear programming II

Lecture 9 Sequential unconstrained minimization

Convex Optimization Lecture 13

Interior Point Methods for Linear Programming: Motivation & Theory

Interior Point Methods in Mathematical Programming

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

A priori bounds on the condition numbers in interior-point methods

Lecture 8. Strong Duality Results. September 22, 2008

On Generalized Primal-Dual Interior-Point Methods with Non-uniform Complementarity Perturbations for Quadratic Programming

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Constrained Optimization Theory

Lecture 14 Barrier method

Convex Optimization and SVM

Second-order cone programming

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Lecture 16: October 22

Interior-Point and Augmented Lagrangian Algorithms for Optimization and Control

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

Conic Linear Optimization and its Dual. yyye

Numerical Optimization

CS-E4830 Kernel Methods in Machine Learning

Nonlinear Optimization: What s important?

A Review of Linear Programming

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Nonlinear Programming

Applications of Linear Programming

Lecture 10. Primal-Dual Interior Point Method for LP

Nonlinear Optimization for Optimal Control

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

LP. Kap. 17: Interior-point methods

Part 1. The Review of Linear Programming

Interior Point Methods for Convex Quadratic and Convex Nonlinear Programming

Active-set prediction for interior point methods using controlled perturbations

Approximate Farkas Lemmas in Convex Optimization

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

A Distributed Newton Method for Network Utility Maximization, II: Convergence

Linear Programming. Operations Research. Anthony Papavasiliou 1 / 21

5 Handling Constraints

Support Vector Machines for Regression

Primal-Dual Interior-Point Methods. Javier Peña Convex Optimization /36-725

The use of shadow price is an example of sensitivity analysis. Duality theory can be applied to do other kind of sensitivity analysis:

Lagrangian Duality Theory

Algorithms for nonlinear programming problems II

Review Solutions, Exam 2, Operations Research

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Constrained Optimization and Lagrangian Duality

Lecture: Algorithms for LP, SOCP and SDP

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Sensitivity Analysis

Massachusetts Institute of Technology 6.854J/18.415J: Advanced Algorithms Friday, March 18, 2016 Ankur Moitra. Problem Set 6

The Primal-Dual Algorithm P&S Chapter 5 Last Revised October 30, 2006

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc.

Linear Programming Duality

12. Interior-point methods

Optimality, Duality, Complementarity for Constrained Optimization

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Linear and Combinatorial Optimization

Introduction to optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

Lecture 11: Post-Optimal Analysis. September 23, 2009

Summer School: Semidefinite Optimization

Optimization and Root Finding. Kurt Hornik

Convex Optimization. Prof. Nati Srebro. Lecture 12: Infeasible-Start Newton s Method Interior Point Methods

Support Vector Machines: Maximum Margin Classifiers

Transcription:

Course Notes for EE227C (Spring 218): Convex Optimiation and Approximation Instructor: Morit Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowit Email: msimchow+ee227c@berkeley.edu October 15, 218 26 Primal-dual interior point methods Previously, we discussed barrier methods and the so-called short-step method for solving constrained LPs, and proved that convergence is guaranteed (albeit slow). Herein, we study a primal-dual interior-point method (the so-called "long-step" path following algorithm), which similarly seeks to approximate points on the central path. Unlike the short-step method, the long-step method considers iterates of primal-dual variables and seeks a more aggressive step sie so long as it lies within a neighborhood of the central path. 26.1 Deriving the dual problem Let x R n be the decision variable, and A R m n, b R m and c R n min c x s.t. Ax = b, x Here, is pointwise. 1

Observe that we can always write the constraint as min x c x + max (b Ax) = min max x c x + max (b Ax) max b min (c x A ) x = max b I(A > c) Hence, the dual problem is as follows This is equivalent to max b s.t. A c max b s.t. A + s = c, s where we introduce the slack variable s R n. If (x,, s) are only feasible, then Ax = b A + s = c x, s Moreover, we can compute that for feasible (x,, s), x, s = x (c A ) = x, c Ax, = x, c b,. This is a proof of weak duality, namely that for any feasible x and, and therefore x, c b, x, c b, Moreover, if there exists an feasible (x,, s ), with x, s = then we have x, c = b, which is strong duality. Duality is also useful to bound the suboptimality gap, since in fact if (x,, s) is feasible, then x, s = x, c b, x, c x, c = x x, c 2

x 2 s2 1 C 2 3 N 8 (γ) x s 1 1 Figure 1: Require iterates to stay within a certain neighborhood of the central path. We want the pairwise products x i s i to be not too different for i = 1, 2,..., n. 26.2 Primal-dual iterates along the central path The above suggests the following approach. Consider iterates (x k, k, s k ), and define Define the strictly feasible set µ k := 1 n x k, s k = x k, c b, k n x k x, c n F o := {Ax = b A + s = c x, s > } Minimiing µ k thus amounts to minimiing a bilinear objective over a linear constraint set. The goal is to generate iterates (x k+1, k+1, s k+1 ) such that µ k+1 (1 Cn ρ )µ k This implies that x k x, c ɛ in k = O (n ρ log(n/ɛ)) steps. The goal is to find a tuple (x,, s) such that µ. We consider the following 3

approach. Define Ax b F τ (x,, s) := A + s c x s τ1 Then the goal is to approx solve F (x,, s ) = over F. We see that this can be obtained by computing the solutions (x τ, τ, τ ) to F τ (x,, s) =. We call the curve τ (x τ, τ, τ ) the central path. Note that, on the central path, x i s i = τ for some τ >. To ensure we stay close to the central path, we consider N (γ) := {(x,, s) F : min x i s i γµ(x, s)} i What we would like to do is take iterates (x k, k, s k ) such that µ k decreases, and (x k, k, s k ) N (γ) for appropriate constants γ. N (γ) ensures the nonnegativity contraints. This is portrayed in Figure 1. 1 Input: Parameters γ (, 1), < σ min < σ max < 1, and initialiation (x,, t ) N (γ). for t =, 1, 2,... do 2 Choose σ k [σ min, σ max ]; 3 Run Newton step on F σk µ k (to be defined). Let ( x k, k, s k ) denote the Newton step ( x k, k, s k ) = 2 F τk (w k ) 1 F τk (w k ), where τ k = σ k µ k and w k = (x k, k, s k ). Let α k (, 1] be the largest step such that α k = max{α (, 1] : (x k, k, s k ) + α( x k, k, s k ) N (γ)} Set (x k+1, k+1, s k+1 ) (x k, k, s k ) + α k ( x k, k, s k ). 4 end Algorithm 1: Long-step Path Following method 26.3 Generating Iterates with the Newton Step The Newton Step for solving fixed point equations F(w) =. Indeed F(w + d) = F(w) + J(w) d + o( d ) 4

Newton s Method in 1D This suggests an iterative scheme: Starting from some w 2 R N : d k = J(w k ) 1 F (w k ), w k+1 = w k + d k, k =, 1, 2,.... F(w) w * w 2 w 1 w w Wright (UW-Madison) Interior-Point Methods August 217 11 / 48 Figure 2: Recall that Newton s method iteratively finds better approximations to the roots (or eroes) of a real-valued function. The Newton s method then chooses w w + d, J(w)d = F(w) Which implies that F(w + d) = o( d ) for w sufficiently closed to the fixed point. This gives you the quick converge. Note that, if F is a linear map, then in fact one Newton step suffices. This can be seen from the Taylor expansion. Our function F τk is nearly linear, but not quite. Let s compute the Newton Step. We observe that the Jacobian is the linear operator A A I Diag(S) Diag(X) Moreover, since (x k, k, s k ) F, we have that Ax k b F τk (x k, k, s k ) = A k + s k c = x k s k τ k 1 x k s k τ k 1 Let s drop subscripts. Then, on one can verify that the Newton satisfies A x A I = Diag(S) Diag(X) s x s + τ1 Some remarks 1. A x =, and that A + s =. 5

2. This implies that (x +, +, s + ) := (x + x, +, s + s) satisfies Ax + b = and A + + s c = 3. We also have and thus 4. Thus, s x + x s = x s + τ1 x + s + = x s + ( x + x s) + x s = x s x s + τ1 + x s F τ (x + s + ) = x s In other words, if we can argue that x s is negligible, then the Newton step produces an almost exact solution. A more concrete analysis would be to study the term nµ(x + α x, s + α x) = x + x, s + s = x, s + α ( x, s + s, x ) + α 2 s, x The last term in the above display vanishes, as shown by the above = x (A + s) = (A x) + x, s = x, Moreover, since s x + x s = x s + τ1, we have by summing that x, s + s, x = x, x + τn = (1 σ) x, s where the last line uses nτ = nσµ = σnτ = σ x, s. Hence, nµ(x + α x, s + α x) = nµ(x, s)(1 α(1 σ)) Hence, if one can show that (1 α)α C(n) > for some constant depending on the dimension, then we see that nµ(x k+1 ) (1 C(n)) k nµ(x ) giving us the rate of decrease. One can then show with more technical work that α = Ω(1/n) while maintaining the N (γ) invariant. References 6