Primal-Dual Interior-Point Methods

Similar documents
Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization /36-725

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization

Primal-Dual Interior-Point Methods. Javier Peña Convex Optimization /36-725

CS711008Z Algorithm Design and Analysis

Convex Optimization Lecture 13

Interior Point Algorithms for Constrained Convex Optimization

12. Interior-point methods

Barrier Method. Javier Peña Convex Optimization /36-725

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

10 Numerical methods for constrained problems

Numerical Optimization

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725

CSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods

Lecture 6: Conic Optimization September 8

Homework 4. Convex Optimization /36-725

Lecture 9 Sequential unconstrained minimization

On Generalized Primal-Dual Interior-Point Methods with Non-uniform Complementarity Perturbations for Quadratic Programming

12. Interior-point methods

Convex Optimization. Lecture 12 - Equality Constrained Optimization. Instructor: Yuanzhang Xiao. Fall University of Hawaii at Manoa

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Applications of Linear Programming

minimize x subject to (x 2)(x 4) u,

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Linear programming II

Lecture 8. Strong Duality Results. September 22, 2008

LP. Kap. 17: Interior-point methods

ICS-E4030 Kernel Methods in Machine Learning

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Convex Optimization and l 1 -minimization

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Constrained Optimization and Lagrangian Duality

Lecture 15: October 15

Lecture 16: October 22

Convex Optimization and SVM

Lecture: Algorithms for LP, SOCP and SDP

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

2.098/6.255/ Optimization Methods Practice True/False Questions

Operations Research Lecture 4: Linear Programming Interior Point Method

Lecture 10: Linear programming duality and sensitivity 0-0

Introduction to optimization

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Generalization to inequality constrained problem. Maximize

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Interior-Point Methods for Linear Optimization

Nonlinear Optimization for Optimal Control

Newton s Method. Javier Peña Convex Optimization /36-725

5 Handling Constraints

Lecture 18: Optimization Programming

Conditional Gradient (Frank-Wolfe) Method

Linear and Combinatorial Optimization

Duality revisited. Javier Peña Convex Optimization /36-725

Computational Finance

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

10-725/ Optimization Midterm Exam

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

Support Vector Machines for Regression

Inequality constrained minimization: log-barrier method

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Lagrangian Duality and Convex Optimization

Lecture 14: Optimality Conditions for Conic Problems

Analytic Center Cutting-Plane Method

Lecture 11: Post-Optimal Analysis. September 23, 2009

SF2822 Applied nonlinear optimization, final exam Wednesday June

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Nonlinear Optimization: What s important?

Convex Optimization. Prof. Nati Srebro. Lecture 12: Infeasible-Start Newton s Method Interior Point Methods

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Optimization. Yuh-Jye Lee. March 28, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 40

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

Algorithms for constrained local optimization

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Algorithms for Constrained Optimization

Written Examination

LINEAR AND NONLINEAR PROGRAMMING

CS-E4830 Kernel Methods in Machine Learning

Lecture 17: Primal-dual interior-point methods part II

Duality in Linear Programs. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Kernel Machines. Pradeep Ravikumar Co-instructor: Manuela Veloso. Machine Learning

More First-Order Optimization Algorithms

Solving Dual Problems

Lecture 7: Convex Optimizations

Interior Point Methods for Mathematical Programming

Linear Programming Duality

Homework Set #6 - Solutions

Quasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization

Convex Optimization & Lagrange Duality

5. Duality. Lagrangian

Truncated Newton Method

Algorithms for nonlinear programming problems II

subject to (x 2)(x 4) u,

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 10: Interior methods. Anders Forsgren. 1. Try to solve theory question 7.

Duality Uses and Correspondences. Ryan Tibshirani Convex Optimization

Convex Optimization M2

Transcription:

Primal-Dual Interior-Point Methods Lecturer: Aarti Singh Co-instructor: Pradeep Ravikumar Convex Optimization 10-725/36-725

Outline Today: Primal-dual interior-point method Special case: linear programming 2

Barrier method versus primal-dual method Like the barrier method, primal-dual interior-point methods aim to compute (approximately) points on the central path. Main differences between primal-dual and barrier methods: Both can be motivated by perturbed KKT conditions, but as the name suggests primal-dual methods update both primal and dual variables Primal-dual interior-point methods usually take one Newton step per iteration (no additional loop for the centering step). Primal-dual interior-point methods are not necessarily feasible. Primal-dual interior-point methods are typically more efficient. Under suitable conditions they have better than linear convergence. 3

Constrained Optimization Consider the problem min x subject to f(x) Ax = b g(x) 0 where the equality constraints are linear. Lagrangian L(x, u, v) = f(x) + u T g(x) + v T (Ax b) 4

KKT conditions KKT conditions f(x) + g(x)u + A T v = 0 Ug(x) = 0 Ax = b u, g(x) 0. Here U = Diag(u), g(x) = [ g 1 (x) g r (x) ] 5

Barrier problem where KKT conditions for Barrier problem min x φ(x) = f(x) + ɛφ(x) Ax = b r log( g j (x)). j=1 KKT conditions for barrier problem f(x) + g(x)u + A T v = 0 Ug(x) = ɛ1 Ax = b u, g(x) > 0. Same as before, except complementary slackness condition is perturbed. 6

We didn t cover this, but Newton updates for log barrier problem can be seen as Newton step for solving these nonlinear equations, after eliminating u (i.e., taking u j = ɛ/g j (x), j = 1,..., r). Primal-dual interior-point updates are also motivated by a Newton step for solving these nonlinear equations, but without eliminating u. Write the KKT conditions as a set of nonlinear equations r(x; u; v) = 0, where f(x) + g(x)u + A T v r(x, u, v) := Ug(x) + ɛ1 Ax b 7

This is a nonlinear equation is (x; u; v), and hard to solve; so let s linearize, and approximately solve Let y = (x; u; v) be the current iterate, and y = ( x; u; v) be the update direction. Define r dual = f(x) + g(x)u + A v r cent = Ug(x) + ɛ1 r prim = Ax b the dual, central, and primal residuals at current y = (x; u; v). Now we make our first-order approximation 0 = r(y + y) r(y) + r(y) y and we want to solve for y in the above. 8

I.e., we solve 2 f(x) + r j=1 u j 2 g j (x) g(x) T A T U g(x) diag(g(x)) 0 A 0 0 r dual = r cent r prim x u v Solution y = ( x, u, v) is our primal-dual update direction Note that the update directions for the primal and dual variables are inexorably linked together (Also, these are different updates than those from barrier method) 9

10 Primal-dual interior-point method Putting it all together, we now have our primal-dual interior-point method. Start with a strictly feasible point x (0) and u (0) > 0, v (0). Define η (0) = g(x (0) ) T u (0) and let σ (0, 1), then we repeat for k = 1, 2, 3...

10 Primal-dual interior-point method Putting it all together, we now have our primal-dual interior-point method. Start with a strictly feasible point x (0) and u (0) > 0, v (0). Define η (0) = g(x (0) ) T u (0) and let σ (0, 1), then we repeat for k = 1, 2, 3... Define ɛ = ση (k 1) /m Compute primal-dual update direction y Determine step size s Update y (k) = y (k 1) + s y Compute η (k) = g(x (k) ) T u (k) Stop if η (k) δ and ( r prim 2 2 + r dual 2 2 )1/2 δ

10 Primal-dual interior-point method Putting it all together, we now have our primal-dual interior-point method. Start with a strictly feasible point x (0) and u (0) > 0, v (0). Define η (0) = g(x (0) ) T u (0) and let σ (0, 1), then we repeat for k = 1, 2, 3... Define ɛ = ση (k 1) /m Compute primal-dual update direction y Determine step size s Update y (k) = y (k 1) + s y Compute η (k) = g(x (k) ) T u (k) Stop if η (k) δ and ( r prim 2 2 + r dual 2 2 )1/2 δ Note the stopping criterion checks both the central residual via η, and (approximate) primal and dual feasibility

11 Backtracking line search At each step, we need to find s and set x + = x + s x, u + = u + s u, v + = v + s v. Two main goals: Maintain g(x) < 0, u > 0 Reduce r(x, u, v) Use a multi-stage backtracking line search for this purpose: start with largest step size s max 1 that makes u + s u 0: { } s max = min 1, min{ u i / u i : u i < 0} Then, with parameters α, β (0, 1), we set s = 0.99s max, and Update s = βs, until g j (x + ) < 0, j = 1,... r Update s = βs, until r(x +, u +, v + ) (1 αs) r(x, u, v)

12 Consider Special case: linear programming min x subject to c T x Ax = b x 0 for c R n, A R m n, b R m. Some history: Dantzig (1940s): the simplex method, still today is one of the most well-known/well-studied algorithms for LPs Karmarkar (1984): interior-point polynomial-time method for LPs. Fairly efficient (US Patent 4,744,026, expired in 2006) Modern state-of-the-art LP solvers typically use both simplex and interior-point methods

13 KKT conditions for standard form LP The points x and (u, v ) are respectively primal and dual optimal LP solutions if and only if they solve: A T v + u = c x i u i = 0, i = 1,..., n Ax = b x, u 0 (Neat fact: the simplex method maintains the first three conditions and aims for the fourth one... interior-point methods maintain the first and last two, and aim for the second)

14 The perturbed KKT conditions for standard form LP are hence: A T v + u = c x i u i = ɛ, i = 1,..., n Ax = b x, u 0 Let s work through the barrier method, and the primal-dual interior point method, to get a sense of these two Barrier method (after elim u): Primal-dual method: 0 = r br (x, v) ( A = T v + diag(x) 1 ɛ c Ax b ) 0 = r pd (x, u, v) A T v + u c = diag(x)u ɛ Ax b

15 Barrier method: 0 = r br (y + y) r br (y) + r br (y) y, i.e., we solve [ diag(x) 2 ɛ A T ] ( ) x = r A 0 v br (x, v) and take a step y + = y + s y (with line search for s > 0), and iterate until convergence. Then update ɛ = σɛ Primal-dual method: 0 = r pd (y + y) r pd (y) + r pd (y) y, i.e., we solve 0 I A T x diag(u) diag(x) 0 u = r pd (x, u, v) A 0 0 v and take a step y + = y + s y (with line search for s > 0), but only once. Then update ɛ = σɛ

Example: barrier versus primal-dual Example from B & V 11.3.2 and 11.7.4: standard LP with n = 50 variables and m = 100 equality constraints Barrier method uses various values of σ, primal-dual method uses σ = 0.1. Both use α = 0.01, β = 0.5 duality gap 10 2 10 0 10 2 10 4 10 6 µ =50 µ =150 µ =2 0 20 40 60 80 Newton iterations Barrier central residual (µ = 1/σ) 10 2 10 2 10 0 10 0 10 2 10 2 10 4 10 4 ˆη ˆη r rfeas 10 5 10 6 10 6 10 10 10 8 10 8 10 10 10 10 10 15 0 0 5 5 10 1015 1520 2025 2530 30 0 5 10 15 20 25 30 iteration iteration number number iteration number Figure 11.21 Progress of the primal-dual interior-point method for an LP, showing surrogate duality gap ˆη and the norm of the primal and dual residuals, versus iteration central number. The residual Primal-dual converges rapidly tofeasibility zero within Primal-dual 24 iterations; the surrogate gap also converges to a very small number in residual about 28 iterations. The primal-dual interior-point gap, methodr convergesfaster than the barrier method, especially if high accuracy is required. feas = 10 5 10 0 ( r prim 2 2 + r dual 2 2) 1/2 Can see that primal-dual is faster to converge to high accuracy 10 2 10 5 16

17 Now a sequence of problems with n = 2m, and n growing. Barrier method uses µ = 100, runs just two outer loops (decreases central residual by 10 4 ); primal-dual method uses σ = 0.1, stops when central residual and feasibility gap are at most 10 8 35 50 Newton iterations 30 25 20 iterations 40 30 20 15 10 1 10 2 10 3 m Barrier method 10 10 1 10 2 10 3 m Primal-dual method Primal-dual method require only slightly more iterations, despite the fact that it is producing higher accuracy solutions