Lecture 17: Lee-Sidford Barrier

Similar documents
Feature Selection: Part 1

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )

Solutions to exam in SF1811 Optimization, Jan 14, 2015

Edge Isoperimetric Inequalities

Lecture 20: Lift and Project, SDP Duality. Today we will study the Lift and Project method. Then we will prove the SDP duality theorem.

MMA and GCMMA two methods for nonlinear optimization

1 Convex Optimization

COS 521: Advanced Algorithms Game Theory and Linear Programming

Lecture 10 Support Vector Machines II

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

U.C. Berkeley CS278: Computational Complexity Professor Luca Trevisan 2/21/2008. Notes for Lecture 8

Lecture 12: Discrete Laplacian

Singular Value Decomposition: Theory and Applications

Some modelling aspects for the Matlab implementation of MMA

Assortment Optimization under MNL

PHYS 705: Classical Mechanics. Calculus of Variations II

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.

6.854J / J Advanced Algorithms Fall 2008

Generalized Linear Methods

Eigenvalues of Random Graphs

Calculation of time complexity (3%)

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

SELECTED SOLUTIONS, SECTION (Weak duality) Prove that the primal and dual values p and d defined by equations (4.3.2) and (4.3.3) satisfy p d.

10-701/ Machine Learning, Fall 2005 Homework 3

Lecture 10 Support Vector Machines. Oct

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lecture 21: Numerical methods for pricing American type derivatives

Errors for Linear Systems

CS : Algorithms and Uncertainty Lecture 17 Date: October 26, 2016

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

Communication Complexity 16:198: February Lecture 4. x ij y ij

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

Lecture 4: Universal Hash Functions/Streaming Cont d

10-801: Advanced Optimization and Randomized Methods Lecture 2: Convex functions (Jan 15, 2014)

Lecture 2: Gram-Schmidt Vectors and the LLL Algorithm

Linear Approximation with Regularization and Moving Least Squares

Case A. P k = Ni ( 2L i k 1 ) + (# big cells) 10d 2 P k.

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Module 9. Lecture 6. Duality in Assignment Problems

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

APPENDIX A Some Linear Algebra

MA 323 Geometric Modelling Course Notes: Day 13 Bezier Curves & Bernstein Polynomials

Lecture 20: November 7

Learning Theory: Lecture Notes

Computing Correlated Equilibria in Multi-Player Games

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture #16 Scribe: Yannan Wang April 3, 2014

Spectral Graph Theory and its Applications September 16, Lecture 5

Moments of Inertia. and reminds us of the analogous equation for linear momentum p= mv, which is of the form. The kinetic energy of the body is.

2.3 Nilpotent endomorphisms

The Geometry of Logit and Probit

Linear Feature Engineering 11

CALCULUS CLASSROOM CAPSULES

Lecture Notes on Linear Regression

Lecture 10: May 6, 2013

Support Vector Machines CS434

Supplement: Proofs and Technical Details for The Solution Path of the Generalized Lasso

THE CHINESE REMAINDER THEOREM. We should thank the Chinese for their wonderful remainder theorem. Glenn Stevens

Lecture 4. Instructor: Haipeng Luo

1 (1 + ( )) = 1 8 ( ) = (c) Carrying out the Taylor expansion, in this case, the series truncates at second order:

Random Walks on Digraphs

Chapter Newton s Method

Convex Optimization. Optimality conditions. (EE227BT: UC Berkeley) Lecture 9 (Optimality; Conic duality) 9/25/14. Laurent El Ghaoui.

Difference Equations

Example: (13320, 22140) =? Solution #1: The divisors of are 1, 2, 3, 4, 5, 6, 9, 10, 12, 15, 18, 20, 27, 30, 36, 41,

Exercises. 18 Algorithms

Integrals and Invariants of Euler-Lagrange Equations

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

Bezier curves. Michael S. Floater. August 25, These notes provide an introduction to Bezier curves. i=0

Design and Analysis of Algorithms

Design and Optimization of Fuzzy Controller for Inverse Pendulum System Using Genetic Algorithm

Bézier curves. Michael S. Floater. September 10, These notes provide an introduction to Bézier curves. i=0

Salmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2

Grover s Algorithm + Quantum Zeno Effect + Vaidman

Problem Set 9 Solutions

The Second Eigenvalue of Planar Graphs

Mechanics Physics 151

LECTURE 9 CANONICAL CORRELATION ANALYSIS

The exam is closed book, closed notes except your one-page cheat sheet.

The Expectation-Maximization Algorithm

Outline and Reading. Dynamic Programming. Dynamic Programming revealed. Computing Fibonacci. The General Dynamic Programming Technique

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture # 15 Scribe: Jieming Mao April 1, 2013

BOUNDEDNESS OF THE RIESZ TRANSFORM WITH MATRIX A 2 WEIGHTS

PROBLEM SET 7 GENERAL EQUILIBRIUM

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

Kernel Methods and SVMs Extension

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

Stanford University Graph Partitioning and Expanders Handout 3 Luca Trevisan May 8, 2013

Lecture 5 Decoding Binary BCH Codes

A 2D Bounded Linear Program (H,c) 2D Linear Programming

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Chapter 12. Ordinary Differential Equation Boundary Value (BV) Problems

Foundations of Arithmetic

Stanford University CS254: Computational Complexity Notes 7 Luca Trevisan January 29, Notes for Lecture 7

Dynamic Programming. Preview. Dynamic Programming. Dynamic Programming. Dynamic Programming (Example: Fibonacci Sequence)

Section 8.3 Polar Form of Complex Numbers

Transcription:

CSE 599: Interplay between Convex Optmzaton and Geometry Wnter 2018 Lecturer: Yn Tat Lee Lecture 17: Lee-Sdford Barrer Dsclamer: Please tell me any mstake you notced. In ths lecture, we talk about the Lee-Sdford barrer (apparently, I have no talent n namng). My orgnal motve to study lnear programmng and convex optmzaton s to solve the maxmum flow problem faster. When we use the standard logarthmc barrer functon to solve the maxmum flow problem, we get a O ( m) teratons algorthm where m s the number of edges. Each teraton nvolves solvng a Laplacan system and hence takes nearly lnear tme. Therefore, we get a O (m 1.5 ) tme algorthm, matchng the prevous best algorthm by Goldberg and Rao [91]. Problem 17.0.2. What s the relaton between nteror pont methods and Goldberg and Rao algorthm? Last lecture, we showed that n general one can get an O ( n) teratons algorthm for lnear programs and hence naturally one conjectured that a O (m n) tme algorthm for the maxmum flow problem s possble where n s the number of vertces. To get the O (m n) tme algorthm, we need to overcome the followng hurdles: 1. Desgn a O (n)-self concordant barrer that can be computed very effcently. 2. Modfy the algorthm to work on the lnear program of the form mn c x. B f=d,0 f u Note that ths lnear program does not have O (n)-self concordant barrer. However, ths s not a proof that we cannot desgn an O ( n) teratons algorthm. In partcular, f there s no the upper constrant f u, the dual lnear program has n varables and hence we can run the nteror pont method on the dual. Wth the upper constrant, what we dd s smply move the dual algorthm to prmal and modfed t such that t works wth the upper constrant. For smplcty, we wll only focus on gettng a O (n)-self concordant barrer that can be computed pretty effcently. 17.1 Problem of Logarthmc barrer Before we talk about our barrer, let us understand the problem of the logarthm barrer. It s known that nteror pont methods on logarthmc barrer can take O( m) even f m s exponental to n. (See for example [92]). One reason s that we can perturb the barrer arbtrary by repeatng constrants. Suppose we repeat the th constrant w many tmes, essentally our barrer functon becomes w ln(a x b ). By choosng w approprately, one can make the central path nearly every vertces of cube lke the followng For these polytopes, the easest way to fx t s to remove redundant constrants. However, when there are many constrants wth smlar but not dentcal drecton, then t s less clear how to handle. 17-1

17-2 Lecture 17: Lee-Sdford Barrer 17.2 Volumetrc Barrer Fgure 17.1: Coped from [92]. Please gnore the labels. Vadya proposed the volumetrc barrer [94]. The self-concordance of ths barrer has a better dependence on m compared to the log-barrer. The volumetrc barrer s φ 2 (x) = lndet(a S 2 A) where s = a x b and S = dag(s). By smple calculatons, we have the followng: Lemma 17.2.1. We have that φ 2 (x) = A xσ x, 2 φ 2 (x) = A x(6σ x 4P (2) x ))A x. where A x = S 1 A, P x = A x(a xa x ) 1 A x, P x (2) s the Schur product of P, σ x s the dagonal of P x as a vector, and Σ x s the dagonal of P x as a dagonal matrx. Furthermore, we have that 2A x Σ xa x 2 φ 2 (x) 6A x Σ xa x. Intutvely, one can thnk φ 2 (x) σ x, ln(a x b ). Instead of a formal proof, we show how to estmate the self-concordance of φ 2 : Lemma 17.2.2. Consder φ(x) = w ln(a x b ). Then, φ(x) ( 2 φ(x)) 1 φ(x) w and that D 3 φ(x)[h,h,h] 2max σ ( WA x ) w ( D 2 φ(x)[h,h] ) 3/2. where σ ( WA x ) s the leverage score of A x W, namely σ ( WA x ) = ( WA x (A xwa x ) 1 Ax W). Remark. Geometrcally, σ( WA x) w = max h x 1 a h, s thedstancefromxtotheboundaryof th constrant. Proof. Note that φ(x) = A xw and 2 φ(x) = A xwa x. Hence, we have φ(x) ( 2 φ(x)) 1 φ(x) = w A x (A x WA x) 1 A x w w w = w

Lecture 17: Lee-Sdford Barrer 17-3 where we used that WA x (A x WA x) 1 A x W s an orthogonal projecton matrx. For the second condton, we have D 3 φ(x)[h,h,h] = 2 w ( a h ) 3 2 w ( a h ) 2 max a h s s s. Note that Hence, we have a h s = a (A xwa x ) 1 2(A xwa x ) 2h 1 s ( A x (A xwa x ) 1 A ) x h x. D 3 φ(x)[h,h,h] 2max σ ( WA x ) w h 3 x By rescalng the φ by constant, we obtan a barrer wth the self-concordance at x σ ( WA x ) w max w. Now, we analyze the case w = σ (A x ). Lemma 17.2.3. We have that σ (A x ) max σ ( Σ (A x )A x ) σ (A x ) 2n m. Proof. By the property of leverage score, we have that σ (A x ) n. We bound the maxmum term n two ways. Frst, we note that Second, we have that whch mples that Therefore, we have Hence, we have σ ( Σ (A x )A x ) σ (A x ) 1 σ (A x ). tr((a x A x) 1 A x dag(1 σ 1 2m )A x) σ 1 2m A x dag(1 σ 1 2m )A x 1 2 A x A x. σ 1 2 A xa x 2A xdag(1 σ 1 2m )A x 4m A xσa x. σ ( Σ A x ) ( ΣA x (A xσa x ) 1 Ax Σ) 4m( ΣA x (A xa x ) 1 Ax Σ) = 4mσ (A x ) 2. Combnng both cases, we have σ ( Σ (A x )A x ) σ (A x ) 1 mn( σ (A x ),4mσ (A x )) 2 m.

17-4 Lecture 17: Lee-Sdford Barrer Formally, the weghng depends on x and hence one need to compute the dervatves formally. By some calculaton, we ndeed can prove that φ 2 has self concordance O(n m). Note that ths barrer has less dependence on m. In the same paper, Vadya consder the barrer φ 2 n m lns and proved that t has self-concordance O( mn). You can agan estmate the self-concordance of ths hybrd barrer by the calculaton above. 17.3 Lee-Sdford Barrer Intutvely, we can thnk the volumetrc barrer rewegh the constrants by leverage score. Lemma 17.2.2 suggests that the best barrer should have σ( WA x) w all roughly the same. Ideally, one would set w = σ ( WA x ). Ths s exactly the condton of the l Lews weght. Recall from Defnton 11.1.3 that l p Lews weght w p (A) be the unque vector w R m + such that ) w = σ (W 1 2 1 p A. Now, the problem s smply to fnd a barrer functon φ such that at every x, we have φ p (x) w p (A x )ln(a x b ). Note that f we ndeed has such functon for p =, then ths barrer should have self-concordant O(n). Recall from Lemma 11.1.5 that w p (A x ) s the maxmzer of ( ) φ p (x) = max logdet A 2 x w W1 p Ax 0 (1 2 m q ) w. Note that ths functon looks very close to volumetrc barrer, except that t optmzes over a famly of volumetrc barrer wth dfferent reweghng. Therefore, t s natural to ask f φ p (x) s self-concordant. Note that φ 2 (x) s exactly the volumetrc barrer. Also, one can prove that w 0 (A x ) s constant (f A s n general poston) and hence φ 0 (x) looks lke the log barrer functon. Therefore, φ p can thnk as a famly of barrer that ncludes log barrer and volumetrc barrer. By some calculatons, t turns out φ p s self-concordance for all 0 < p <. Theorem 17.3.1. Gven a polytope Ω = {Ax > b}. Suppose that Ω s bounded and non-empty and that every row of A s non-zero. For any p > 0, after rescalng, the φ p (x) s a O p (n m 1 p+2 ) self-concordant barrer. Take p = Θ(logm), we have a barrer wth self concordance s O(nlog O(1) m). Unfortunately, the proof of ths s a lttle bt long. We left t as a 12-pages long calculus exercse. =1 17.4 Open Problems Ideally, one would ask f the LS barrer s practcal. Snce the current best approach n practce s the prmal dual central path, we have the followng queston: Problem 17.4.1. What s the prmal dual analogy of the LS result?

Lecture 17: Lee-Sdford Barrer 17-5 Besde lnear programs, another mportant class of problems s sem-defnte programmng. Nesterov and Nemrovsk generalzed the volumetrc barrer to semdefnte programmng settng [93]. Naturally, can we generalze the LS barrer to that settng? Note that the current fastest algorthm for sem-defnte programmng s cuttng plane method nstead of nteror pont method. Ths s very unusual for structured convex programmng. Therefore, there are a lot of opportuntes for mprovng algorthms on sem-defnte programmng. Problem 17.4.2. Improve the convergence rate of nteror pont methods for sem-defnte programmng. If we vew the LS barrer as a weghted combnaton of log barrer, then t s natural to ask: Problem 17.4.3. Gven barrers φ for K, can we come wth a barrer for K that s better than φ? Fnally, gong back to my orgnal purpose of solvng lnear programs: Problem 17.4.4. Can you solve the maxmum flow problem faster than O (m n)? References [91] Andrew V Goldberg and Satsh Rao. Beyond the flow decomposton barrer. Journal of the ACM (JACM), 45(5):783 797, 1998. [92] Murat Mut and Tamás Terlaky. A tght teraton-complexty upper bound for the mty predctorcorrector algorthm va redundant klee-mnty cubes. 2014. [93] Yur Nesterov and Arkad Nemrovsk. Interor-pont polynomal algorthms n convex programmng. SIAM, 1994. [94] Pravn M Vadya. A new algorthm for mnmzng convex functons over convex sets. In Foundatons of Computer Scence, 1989., 30th Annual Symposum on, pages 338 343. IEEE, 1989.