Lecture 6: Conic Optimization September 8

Similar documents
I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lecture 18: Optimization Programming

Lecture: Duality of LP, SOCP and SDP

Constrained Optimization and Lagrangian Duality

Convex Optimization M2

5. Duality. Lagrangian

Convex Optimization & Lagrange Duality

Lecture Note 5: Semidefinite Programming for Stability Analysis

CS711008Z Algorithm Design and Analysis

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Lecture 7: Convex Optimizations

Semidefinite Programming Basics and Applications

Lecture: Duality.

Convex Optimization Boyd & Vandenberghe. 5. Duality

Optimization for Machine Learning

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Primal-Dual Interior-Point Methods. Javier Peña Convex Optimization /36-725

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization

ICS-E4030 Kernel Methods in Machine Learning

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Convex Optimization and SVM

Math 273a: Optimization Subgradients of convex functions

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization /36-725

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Lecture: Cone programming. Approximating the Lorentz cone.

Lecture 9 Sequential unconstrained minimization

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Optimality Conditions for Constrained Optimization

Agenda. 1 Cone programming. 2 Convex cones. 3 Generalized inequalities. 4 Linear programming (LP) 5 Second-order cone programming (SOCP)

12. Interior-point methods

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

Lecture 14: Optimality Conditions for Conic Problems

Lecture 8. Strong Duality Results. September 22, 2008

Convex Optimization and l 1 -minimization

LECTURE 10 LECTURE OUTLINE

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

January 29, Introduction to optimization and complexity. Outline. Introduction. Problem formulation. Convexity reminder. Optimality Conditions

GEORGIA INSTITUTE OF TECHNOLOGY H. MILTON STEWART SCHOOL OF INDUSTRIAL AND SYSTEMS ENGINEERING LECTURE NOTES OPTIMIZATION III

Lecture 15: October 15

Lecture 5. The Dual Cone and Dual Problem

Chap 2. Optimality conditions

Part IB Optimisation

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Agenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization

Optimisation convexe: performance, complexité et applications.

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

EE 227A: Convex Optimization and Applications October 14, 2008

Canonical Problem Forms. Ryan Tibshirani Convex Optimization

10 Numerical methods for constrained problems

Optimality, Duality, Complementarity for Constrained Optimization

Lagrangian Duality Theory

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

CS-E4830 Kernel Methods in Machine Learning

Support Vector Machines: Maximum Margin Classifiers

CO 250 Final Exam Guide

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Relaxations and Randomized Methods for Nonconvex QCQPs

Tutorial on Convex Optimization for Engineers Part II

4TE3/6TE3. Algorithms for. Continuous Optimization

Lagrangian Duality and Convex Optimization

Primal-Dual Interior-Point Methods

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Generalization to inequality constrained problem. Maximize

Lecture: Examples of LP, SOCP and SDP

Advances in Convex Optimization: Theory, Algorithms, and Applications

IE 521 Convex Optimization

Lecture: Introduction to LP, SDP and SOCP

Course Outline. FRTN10 Multivariable Control, Lecture 13. General idea for Lectures Lecture 13 Outline. Example 1 (Doyle Stein, 1979)

IE 521 Convex Optimization

Barrier Method. Javier Peña Convex Optimization /36-725

Inner approximation of convex cones via primal-dual ellipsoidal norms

EE/AA 578, Univ of Washington, Fall Duality

Linear and non-linear programming

Optimization. Yuh-Jye Lee. March 28, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 40

CSCI : Optimization and Control of Networks. Review on Convex Optimization

Convex Optimization and Modeling

Tutorial on Convex Optimization: Part II

Module 04 Optimization Problems KKT Conditions & Solvers

Exact SDP Relaxations for Classes of Nonlinear Semidefinite Programming Problems

Convexification of Mixed-Integer Quadratically Constrained Quadratic Programs

The Q-parametrization (Youla) Lecture 13: Synthesis by Convex Optimization. Lecture 13: Synthesis by Convex Optimization. Example: Spring-mass System

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

A Brief Review on Convex Optimization

Interior Point Algorithms for Constrained Convex Optimization

Primal/Dual Decomposition Methods

Lecture 3 January 28

On the Method of Lagrange Multipliers

Largest dual ellipsoids inscribed in dual cones

4. Convex optimization problems

Global Optimization of Polynomials

15. Conic optimization

Nonlinear Optimization: What s important?

Transcription:

IE 598: Big Data Optimization Fall 2016 Lecture 6: Conic Optimization September 8 Lecturer: Niao He Scriber: Juan Xu Overview In this lecture, we finish up our previous discussion on optimality conditions for constrained convex programs. We will then introduce an important class of well-structured convex optimization problems conic programming. 6.1 Recap In the previous lecture, we have discussed the optimality condition for simple constrained or unconstrained convex problems If f(x) is a convex and differentiable function defined on a convex set X, then x arg min x X f(x) f(x ) (x x ) 0 x X, If f(x) is a convex and differentiable function defined on R n, then What about constrained convex problems? x arg min x R n f(x) f(x ) = 0. Primal: Opt(P) = min f(x) s.t. g i (x) 0 i = 1,..., m, x X. Lagrangian dual: Opt(D) = sup L(λ) Recall the Lagrangian duality theorem, Weak duality: Opt(D) Opt(P); Strong duality: Under slater condition, Opt(P)=Opt(D). where L(λ) = inf {f(x) + m λ i g i (x)}. x X Solving the primal problem is essentially the same as solving its Lagrangian dual problem. 6-1

Lecture 6: Conic Optimization September 8 6-2 6.2 Constrained Convex Optimization (cont d) 6.2.1 Saddle Point Formulation Recall the Lagrangian function: L(x, λ) = f(x) + λ i g i (x). We can rewrite the primal and its Lagrangian dual problems in the following form. Opt(D) = sup inf L(x, λ) x X }{{} :=L(λ) Opt(P) = inf sup x X L(x, λ) }{{} :=L(x) Note that L(x) = { f(x) gi (x) 0, i + otherwise. Hence, inf x X L(x) = Opt(P). Definition 6.1 (Saddle Point) We call (x, λ ), where x X and λ 0 a saddle point of L(x, λ) if and only if L(x, λ ) L(x, λ ) L(x, λ), x X, λ 0. Figure 6.1: Saddle point (Source: https://en.wikipedia.org/wiki/saddle_point) Remark. From Figure (6.1), we can see that the saddle point (the red point) is the local minimum point in the x-axis direction, and the local maximum point in the y-axis direction.

Lecture 6: Conic Optimization September 8 6-3 Theorem 6.2 (x, λ ) is a saddle point of L(x, λ) if and only if x is optimal for (P), λ is optimal for (D), and Opt(P)=Opt(D). Proof: ( ): On the one hand, we have Opt(D) = sup inf x X L(x, λ) inf x X L(x, λ ) = L(x, λ ). The last equality is due to the fact that L(x, λ ) L(x, λ ), x X. On the other hand, we have Opt(D) = inf sup x X L(x, λ) sup L(x, λ) = L(x, λ ). The last equality is due to the fact that L(x, λ ) L(x, λ), λ 0. We arrive at Opt(D) L(x, λ ) Opt(P), and by weak duality, Opt(D) Opt(P)); it can then be shown that Opt(P)=Opt(D), x is optimal for (P), and λ is optimal for (D). ( ): Since x is optimal for (P), λ is optimal for (D), we have Opt(D) = L(λ ) = inf x X L(x, λ ) L(x, λ ) Opt(P ) = L(x ) = sup L(x, λ) L(x, λ ) Now that Opt(P ) = Opt(D), we must have Opt(P ) = L(x, λ ) = Opt(D). That is, which implies that (x, λ ) is a saddle point. inf L(x, x X λ ) = L(x, λ ) = sup L(x, λ) 6.2.2 Optimality conditions for constrained convex problem Theorem 6.3 (Saddle point condition) If (P) is a convex optimization problem, and x is a feasible solution to (P), then x is optimal for (P) under slater condition (x, λ ) is a saddle point of L(x, λ) (6.1) Proof: ( ): Under the slater condition, if x is the optimal solution of (P), we can find an optimal solution λ for the Lagrangian dual problem such that Opt(P)=Opt(D). By Theorem (6.2), (x, λ ) is a saddle point of L(x, λ). ( ): By Theorem (6.2), if (x, λ ) is a saddle point of L(x, λ), then x is optimal for (P). Theorem 6.4 (Karush-Kuhn-Tucker (KKT) condition) If (P) is convex, x is a feasible solution to (P), and f(x), g i (x) s are all differentiable at x, then x is optimal for (P) under slater condition λ 0 s.t. (a) (b) f(x ) + m λ i g i(x ) N X (x ); λ i g i(x ) = 0, i = 1,..., m (6.2) where N X (x ) represents the normal cone of x (N X (x ) = {g : g (x x ) 0, x X}).

Lecture 6: Conic Optimization September 8 6-4 We refer to (a) as the Lagrangian stationarity condition, (b) the complementary slackness condition. Both (a) and (b) consist of the KKT condition. Proof: The proof follows directly from the previous theorem. First, L(x, λ ) L(x, λ ), x X L x (x, λ ) (x x ) 0, x X condition (a) This is because L x (x, λ ) = f(x ) + m λ i g i(x )). Second, L(x, λ ) L(x, λ), λ 0 f(x ) + λ i g i (x ) f(x ) + λ i g i (x ) λ 0 λ i g i (x ) λ i g i (x ) λ 0 condition (b) (because g i (x ) 0, i = 1,..., m, and λ 0). Therefore, (x, λ ) is a saddle point if and only if (a) and (b) hold true. 6.3 Conic Optimization 6.3.1 A bit of history Mid-1940s Mid-1970s Mid-1980s Dantzig Simplex method for solving linear programming Worst cast exponential time (however, fast when implemented) Shor, Khachiyan, Nemirovski, Yudin Ellipsoid method for linear programming and convex optimization Polynomial time (however, very slow for large problems) Complexity: O(n 2 log( 1 ɛ )) O(n2 ) Karmarkar, Nemirovski, Nesterov Interior point method for linear programming and well-structured convex problem Polynomial time Complexity: O(ν log( ν ɛ )) O(n3 ) The poor performance of some algorithms (e.g. Ellipsoid method) despite of polynomial time stems from their black box oriented nature - these algorithms do not adjust themselves to the structure of the problem and only utilize local information, e.g. the values and (sub)gradients of the objective and the constraints at query points. While in contrast, interior point method is designed to take advantage of the problem s structure and is much more efficient. 6.3.2 From linear programming to conic programming When passing from a linear program to a nonlinear convex program, we can act as follows

Lecture 6: Conic Optimization September 8 6-5 (Linear Programming) min{c x : Ax b 0} ( is component-wise) = min{c x : Ax b R n +}. (Conic Programming) min{c x : Ax b K 0} ( K is some order) = min{c x : Ax b K}, where K is a regular (closed, pointed, nonempty interior) cone. The order K is defined as x K y if and only if x y K. It is easy to see that linear programming is a special case of conic programming which will be shown below. 6.3.3 Three typical conic problems 1. Linear programming (LP) K = R n + = R + R +. (LP): min{c x : a i x b i 0, i = 1,..., k} 2. Second-order cone programming (SOCP, also called conic quadratic programming) K = L n1 L n k, where L n = {(ν 1,..., ν n ), ν n ν1 2 +... ν2 n 1 }. (SOCP): min{c x : A i x b i 2 c i x d i, i = 1,..., k} One can treat Ax b = [ A 1 x b 1, c 1 x d 1 ; ; A k x b k, c k x d k]. 3. Semi-definite programming (SDP) k = S n1 + S n k +, where S n + = {X : X 0, X = X }. (SDP): min{c x : A i x b i 0, i = 1,..., k}; This can be simplified as (SDP): min{c x : Ax b 0}, because Ax b := A 1 x b 1... A k x b k 0 A i x b i 0, i = 1,..., k. Lemma. [ (Schur ] complement for symmetric positive semidefinite matrices) P R If S = is symmetric with Q 0, then S 0 P R R Q Q 1 R 0. Remark. We show that (LP) (SOCP) (SDP).

Lecture 6: Conic Optimization September 8 6-6 (LP) (SOCP) (SOCP) (SDP) x 1. x m L m [ ] a 0 i x b i 0 a i x b L 2 0. i x m x 1... x m 1 x 1 x m.... x m 1 x m by Schur lemma 0 ========== x m x 2 1 + + x2 m 1. Note that the above matrix is linear in x and can be written as i x ia i with properly defined A i. 6.3.4 Calculus of Conic Programming Let K be a family of regular cones, e.g., R n +, L n, S n +, or their inner points. Definition 6.5 X is K-representable if K K, s.t. X = {x : u, Ax + Bu b K}. Definition 6.6 f(x) is K-representable if epi(f) is K-representable, i.e., K K, s.t., epi(f) = {(x, t) : v, Cx + dt + Dv e K}. Remark It can be shown that many convexity-preserving operations preserve K-representability, e. g taking intersections, direct product, affine image of K-representable sets; taking conic combination, affine composition, partial minimization of K-representable functions. Here we provide some examples: Examples (SOCP-r functions) f(x) = x 2 [ x epi(f) = {(x, t) : t x 2 } {(x, t) : L t] n+1 }. f(x) = x x 2x epi(f) = {(x, t) : t x x} {(x, t) : (t + 1) 2 (t 1) 2 + 4x x} {(x, t) : t 1 L n+2 }. t + 1 f(x) = x Qx + q x + t, where Q = L L 2L x epi(f) = {(x, t) : t x Qx + q x + t} {(x, t) : t q x r 1 L n+2 }. t q x r + 1

Lecture 6: Conic Optimization September 8 6-7 Figure 6.2: Relationships among LP, QCQP, SOCP, SDP (QCQP refers to quadratically constrained quadratic programming.) References [BV04] [DN01] Boyd, S. and Vandenberghe, L. (2004). Convex optimization. Cambridge university press. Ben-Tal, A. and Nemirovski, A. (2001). Lectures on modern convex optimization: analysis, algorithms, and engineering applications, (Vol. 2). SIAM.