Interior-point methods Optimization Geoff Gordon Ryan Tibshirani

Similar documents
Interior-point methods Optimization Geoff Gordon Ryan Tibshirani

Interior-point methods Optimization Geoff Gordon Ryan Tibshirani

Lecture 24: August 28

CS711008Z Algorithm Design and Analysis

Lecture 18: Optimization Programming

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization /36-725

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Support Vector Machines: Maximum Margin Classifiers

5. Duality. Lagrangian

Duality. Geoff Gordon & Ryan Tibshirani Optimization /

Interior Point Methods in Mathematical Programming

Interior Point Algorithms for Constrained Convex Optimization

Convex Optimization and SVM

Barrier Method. Javier Peña Convex Optimization /36-725

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Numerical Optimization

Interior Point Methods for Mathematical Programming

10 Numerical methods for constrained problems

Convex Optimization in Classification Problems

An E cient A ne-scaling Algorithm for Hyperbolic Programming

Nonlinear Optimization for Optimal Control

Convex Optimization and l 1 -minimization

Duality revisited. Javier Peña Convex Optimization /36-725

10701 Recitation 5 Duality and SVM. Ahmed Hefny

Semidefinite Programming

Support Vector Machine (SVM) and Kernel Methods

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Lecture 9 Sequential unconstrained minimization

12. Interior-point methods

Convex Optimization Lecture 13

10 Karmarkars Algorithm

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Lecture 5. The Dual Cone and Dual Problem

Machine Learning. Support Vector Machines. Manfred Huber

Convex Optimization M2

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Convex Optimization Boyd & Vandenberghe. 5. Duality

Linear programs Optimization Geoff Gordon Ryan Tibshirani

2.098/6.255/ Optimization Methods Practice True/False Questions

Lecture: Duality.

Optimization: Then and Now

Lecture: Duality of LP, SOCP and SDP

Applications of Linear Programming

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Lecture 8. Strong Duality Results. September 22, 2008

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Optimality, Duality, Complementarity for Constrained Optimization

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Lecture 16: October 22

Machine Learning and Data Mining. Support Vector Machines. Kalev Kask

A Brief Review on Convex Optimization

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Linear and non-linear programming

Nonlinear Optimization: What s important?

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Support Vector Machines and Kernel Methods

EE/AA 578, Univ of Washington, Fall Duality

CSCI : Optimization and Control of Networks. Review on Convex Optimization

Lagrangian Duality Theory

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

WHY DUALITY? Gradient descent Newton s method Quasi-newton Conjugate gradients. No constraints. Non-differentiable ???? Constrained problems? ????

8. Geometric problems

Primal-Dual Interior-Point Methods. Javier Peña Convex Optimization /36-725

Homework 4. Convex Optimization /36-725

Lecture 14: Optimality Conditions for Conic Problems

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Convex Optimization M2

12. Interior-point methods

Analytic Center Cutting-Plane Method

Interior-Point Methods for Linear Optimization

Bellman s Curse of Dimensionality

Optimisation in Higher Dimensions

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Lecture 15: October 15

Primal-Dual Interior-Point Methods

Algorithms for constrained local optimization

Interior Point Methods for LP

Kernel Machines. Pradeep Ravikumar Co-instructor: Manuela Veloso. Machine Learning

Primal/Dual Decomposition Methods

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725

On Generalized Primal-Dual Interior-Point Methods with Non-uniform Complementarity Perturbations for Quadratic Programming

Convex Optimization and Support Vector Machine

Lecture 17: Primal-dual interior-point methods part II

Lecture: Introduction to LP, SDP and SOCP

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Nonsymmetric potential-reduction methods for general cones

Learning From Data Lecture 25 The Kernel Trick

Newton s Method. Javier Peña Convex Optimization /36-725

Support Vector Machines and Kernel Methods

Duality in Linear Programs. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Transcription:

Interior-point methods 10-725 Optimization Geoff Gordon Ryan Tibshirani

SVM duality Review min v T v/2 + 1 T s s.t. Av yd + s 1 0 s 0 max 1 T α α T Kα/2 s.t. y T α = 0 0 α 1 Gram matrix K Interpretation support vectors & complementarity reconstruct primal solution from dual 2

Review Kernel trick high-dim feature spaces, fast positive definite function Examples polynomial homogeneous polynomial linear Gaussian RBF 2 1 0 1 2 2 1 0 1 2 3

Review: LF problem Ax + b 0 Ball center bad summary of LF problem Max-volume ellipsoid / ellipsoid center good summary (1/n of volume), but expensive Analytic center of LF problem maximize product of distances to constraints min ln(ai T x + bi) Dikin ellipsoid @ analytic center: not quite as good (just 1/m < 1/n), but much cheaper 4

Force-field interpretation of analytic center Pretend constraints are repelling a particle normal force for each constraint force 1/distance Analytic center = equilibrium = where forces balance 5

Newton for analytic center f(x) = ln(a i T x + bi) df/dx = d 2 f/df 2 = 6

Dikin ellipsoid E(x 0) = { x (x x0) T H(x x0) 1 } H = Hessian of log barrier at x0 unit ball of Hessian norm at x0 E(x 0) X for any strictly feasible x0 affine constraints can be just feasible E(x0): as above, but intersected w/ affine constraints vol(e(x ac)) vol(x)/m weaker than ellipsoid center, but still very useful 7

E(x0) X E(x 0) = { x (x x0) T H(x x0) 1 } H = A T S -2 A S = diag(s) = diag(ax0 + b) 8

me(x0) X Feasible point x: Ax + b 0 Analytic center x ac: A T y = 0 Let Y = diag(y ac), H = A T Y 2 A; show: (x xac) T H(x xac) m 2 [+ m] y = 1./(Axac+b) 9

Combinatorics v. analysis Two ways to find a feasible point of Ax+b 0 find analytic center minimize a smooth function find a feasible basis combinatorial search 10

Bad conditioning? No problem. Analytic center & Dikin ellipsoids invariant to affine xforms w = Mx+q W = { w AM -1 (w q) + b 0 } Can always xform so that a ball takes up vol(y)/m Dikin ellipsoid @ac sphere 11

LF LP: the central path Analytic center was for: find x st Ax + b 0 Now: min ct x st Ax + b 0 Same trick: min ft(x) = c T x (1/t) ln(ai T x + bi) parameter t > 0 central path = t 0: t : 12

Force-field interpretation of central path Force along objective; normal forces for each constraint ments c PSfrag replacements 3c t=1 t=3 13

Newton for central path min f t(x) = c T x (1/t) ln(ai T x + bi) df/dx = d 2 f/dx 2 = 14

Central path example objective t 0 t 15

New LP algorithm? Set t=1012. Find corresponding point on central path by Newton s method. worked for example on previous slide! but has convergence problems in general Alternatives? 16

Constraint form of central path min ln si st Ax + b 0 c T x λ a 1-1 mapping λ(t) w/ x(λ(t)) = x(t) t>0 but this form is slightly less convenient since we don t know minimal feasible value of λ or maximal nontrivial value of λ 17

Dual of central path min ct x (1/t) ln si st Ax + b = s 0 minx,s maxy L(x,s,y) = c T x (1/t) ln si + y T (s Ax b) 18

Primal-dual correspondence Primal and dual for central path: min c T x (1/t) ln si st Ax + b = s 0 max (m ln t)/t + m/t + (1/t) ln yi y T b st A T y = c y 0 L(x,s,y) = ct x (1/t) ln si + y T (s Ax b) grad wrt s: to get x: 19

At optimum: Duality gap primal value c T x (1/t) ln si = dual value (m ln t)/t + m/t + (1/t) ln yi y T b s y = te 20

Primal-dual constraint form Primal-dual pair: min c T x st Ax + b 0 max b T y st A T y = c y 0 KKT: Ax + b 0 (primal feasibility) y 0 A T y = c (dual feasibility) c T x + b T y 0 (strong duality) or, c T x + b T y λ (relaxed strong duality) 21

Analytic center of relaxed KKT Relaxed KKT conditions: Ax + b = s 0 y 0 A T y = c c T x + b T y λ Central path = {analytic centers of relaxed KKT} 22

Algorithm t := 1, y := 1m, x := 0 n [s := 1 m ] Repeat Use infeasible-start Newton to find point on dual central path Recover primal (s,x); gap c T x + b T y = m/t s = 1./ty x = A\(s b) [have already (Newton)] t := αt (α > 1) 23

Example ty gap 10 4 m/t 10 2 duality gap 10 0 10 2 m = 50 m = 500 m = 1000 10 4 0 10 20 30 40 50 Newton iterations 24