Second-order cone programming

Similar documents
Approximate Farkas Lemmas in Convex Optimization

The Q Method for Symmetric Cone Programmin

2.1. Jordan algebras. In this subsection, we introduce Jordan algebras as well as some of their basic properties.

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

15. Conic optimization

A new primal-dual path-following method for convex quadratic programming

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

Introduction to Nonlinear Stochastic Programming

Semidefinite Programming, Combinatorial Optimization and Real Algebraic Geometry

Continuous Optimisation, Chpt 9: Semidefinite Problems

Lecture 17: Primal-dual interior-point methods part II

18. Primal-dual interior-point methods

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Apolynomialtimeinteriorpointmethodforproblemswith nonconvex constraints

Lecture: Algorithms for LP, SOCP and SDP

A PRIMAL-DUAL INTERIOR POINT ALGORITHM FOR CONVEX QUADRATIC PROGRAMS. 1. Introduction Consider the quadratic program (PQ) in standard format:

IMPLEMENTATION OF INTERIOR POINT METHODS

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

Tamás Terlaky George N. and Soteria Kledaras 87 Endowed Chair Professor. Chair, Department of Industrial and Systems Engineering Lehigh University

Nonlinear Optimization for Optimal Control

Interior Point Methods for Convex Quadratic and Convex Nonlinear Programming

Lecture: Introduction to LP, SDP and SOCP

Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma

Interior Point Algorithms for Constrained Convex Optimization

Primal-dual path-following algorithms for circular programming

Interior Point Methods in Mathematical Programming

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

12. Interior-point methods


EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Tamás Terlaky George N. and Soteria Kledaras 87 Endowed Chair Professor. Chair, Department of Industrial and Systems Engineering Lehigh University

1 Number Systems and Errors 1

Supplement: Universal Self-Concordant Barrier Functions

Duality revisited. Javier Peña Convex Optimization /36-725

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

Primal-Dual Geometry of Level Sets and their Explanatory Value of the Practical Performance of Interior-Point Methods for Conic Optimization

Research overview. Seminar September 4, Lehigh University Department of Industrial & Systems Engineering. Research overview.

CS711008Z Algorithm Design and Analysis

Continuous Optimisation, Chpt 9: Semidefinite Optimisation

Interior Point Methods for Linear Programming: Motivation & Theory

Primal-dual IPM with Asymmetric Barrier

Summer School: Semidefinite Optimization

The Q Method for Symmetric Cone Programming

Interior Point Methods for Nonlinear Optimization

Lecture 14: Optimality Conditions for Conic Problems

On implementing a primal-dual interior-point method for conic quadratic optimization

Semidefinite Programming Basics and Applications

Basic Concepts in Linear Algebra

Lecture 3. Optimization Problems and Iterative Algorithms

SEMIDEFINITE PROGRAM BASICS. Contents

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin

A smoothing Newton-type method for second-order cone programming problems based on a new smoothing Fischer-Burmeister function

New Infeasible Interior Point Algorithm Based on Monomial Method

Sparse Optimization Lecture: Basic Sparse Optimization Models

Interior-Point Methods for Linear Optimization

The Q Method for Second-Order Cone Programming

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

Projection methods to solve SDP

A Constraint-Reduced MPC Algorithm for Convex Quadratic Programming, with a Modified Active-Set Identification Scheme

Introduction to Applied Linear Algebra with MATLAB

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

We describe the generalization of Hazan s algorithm for symmetric programming

Interior-Point Methods

Using Schur Complement Theorem to prove convexity of some SOC-functions

Large Scale Portfolio Optimization with Piecewise Linear Transaction Costs

Following The Central Trajectory Using The Monomial Method Rather Than Newton's Method

Interior Point Methods for Mathematical Programming

Conic Linear Optimization and its Dual. yyye

A new Primal-Dual Interior-Point Algorithm for Second-Order Cone Optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

A Distributed Newton Method for Network Utility Maximization, II: Convergence

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

More First-Order Optimization Algorithms

Semidefinite Programming

III. Applications in convex optimization

Second-Order Cone Programming

A Full Newton Step Infeasible Interior Point Algorithm for Linear Optimization

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

An Infeasible Interior Point Method for the Monotone Linear Complementarity Problem

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

Lecture 14 Barrier method

LP. Kap. 17: Interior-point methods

ON POSITIVE SEMIDEFINITE PRESERVING STEIN TRANSFORMATION

The Solution of Euclidean Norm Trust Region SQP Subproblems via Second Order Cone Programs, an Overview and Elementary Introduction

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

SMO vs PDCO for SVM: Sequential Minimal Optimization vs Primal-Dual interior method for Convex Objectives for Support Vector Machines

A Smoothing Newton Method for Solving Absolute Value Equations

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Written Examination

Review of Basic Concepts in Linear Algebra

Iterative Methods. Splitting Methods

4TE3/6TE3. Algorithms for. Continuous Optimization

Introduction to Semidefinite Programs

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture 9 Sequential unconstrained minimization

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Transcription:

Outline Second-order cone programming, PhD Lehigh University Department of Industrial and Systems Engineering February 10, 2009

Outline 1 Basic properties Spectral decomposition The cone of squares The arrowhead operator 2 Notation Optimality conditions barrier functions perturbed optimality The Newton system 3 4

Basic properties Spectral decomposition The cone of squares The arrowhead operator A new For u, v R n define: u v = u T v; u 1 v 2:n + v 1 u 2:n ). Theorem Properties of ) 1 Distributive law: u v + w) = u v + u w. 2 Commutative law: u v = v u. 3 The unit element is ι = 1; 0), i.e., u ι = ι u = u. 4 Using the notation u 2 = u u we have u u 2 v) = u 2 u v). 5 Power associativity: u p = u u is well-defined. 6 Associativity does not hold in general.

Basic properties Spectral decomposition The cone of squares The arrowhead operator Spectral decomposition Every vector u R n can be written as u = λ 1 c 1) + λ 2 c 2), where c 1) and c 2) are on the boundary of the cone, and c 1)T c 2) = 0 c 1) c 2) = 0 c 1) c 1) = c 1) c 2) c 2) = c 2) c 1) + c 2) = ι c 1), c 2) : Jordan frame λ 1, λ 2 : eigenvalues or spectral values: λ 1,2 u) = u 1 ± u 2:n 2 Naturally: u L λ 1,2 u) 0

Basic properties Spectral decomposition The cone of squares The arrowhead operator The cone of squares Theorem A vector x is in a second order cone i.e., x 1 x 2:n 2 ) if and only if it can be written as the square of a vector under the multiplication, i.e., x = u u. u F = λ 2 1 + λ2 2 = 2 u 2, u 2 = max { λ 1, λ 2 } = u 1 + u 2:n 2, u 1 = λ 1 1 c1) + λ 1 2 c2), u 1 2 = λ 1 2 1 c1) + λ 1 2 2 c 2), where u u 1 = u 1 u = ι and u 1 2 u 1 2 = u.

Basic properties Spectral decomposition The cone of squares The arrowhead operator The arrowhead operator Since the mapping v u v is linear, it can be represented with a matrix. u 1 u 2... u n u 2 u 1 Arr u) =...., u n u 1 Now we have u v = Arr u) v = Arr u) Arr v) ι. Quadratic representation: Q u = 2 Arr u) 2 Arr u 2), thus Q u v) = 2u u v) u 2 v is a quadratic function.

Primal-dual interior-point methods: notation Outline Notation Optimality conditions barrier functions perturbed optimality The Newton system K = L n1 L nk, A = A 1,..., A k), x = x 1 ;... ; x k), s = s 1 ;... ; s k), c = c 1 ;... ; c k). With this notation we can write k Ax = A i x i, i=1 ) A T y = A 1 T y;,... ; A k T y. Arr u) and Q u are block diagonal matrices built from the blocks Arr u i) and Q u i, respectively.

Notation Optimality conditions barrier functions perturbed optimality The Newton system Duality and optimality Weak duality always holds Primal dual) strict feasibility implies strong duality and dual primal) solvability Under strong duality, the optimality conditions for second order conic optimization are Ax = b, x K A T y + s = c, s K x s = 0. An equivalent form of the complementarity condition is c T x b T y = x T s = 0.

Notation Optimality conditions barrier functions perturbed optimality The Newton system The central path using barrier functions If x int L, consider ) φx) = ln x 2 1 x 2:n 2 2 = ln λ 1 x) ln λ 2 x), Goes to if x is getting close to the boundary of the cone. Derivatives: φx) = 2 x 1; x 2:n ) T x 2 1 x 2:n 2 2 = 2 x 1) T, where the inverse is taken in the Jordan algebra.

Notation Optimality conditions barrier functions perturbed optimality The Newton system The central path Perturbed optimality conditions: Ax = b, x K A T y + s = c, s K where ι i = 1; 0;... ; 0) R n i. Newton system: x i s i = 2µι i, i = 1,..., k, A x = 0 A T y + s = 0, x i s i + x i s i = 2µι i x i s i, i = 1,..., k, where x = x 1 ;... ; x k ) and s = s 1 ;... ; s k ).

Newton system - rewritten Outline Notation Optimality conditions barrier functions perturbed optimality The Newton system A T A Arr s) I Arr x) y x s = 0 0 2µι x s, where ι = ι 1 ;... ; ι k ). Eliminating x and s: A Arr s) 1 Arr x) A T ) y = A Arr s) 1 2µι x s). Problems: Not symmetric May be singular Solution: symmetrization!

Outline min Q p 1c ) T Qp x) ) AQp 1 Qp x) = b Q p x K max b T y ) T AQp 1 y + Qp 1s = Q p 1c Q p 1s K Lemma If p int K, then 1 Q p Q p 1 = I. 2 The cone K is invariant, i.e., Q p K) = K. 3 The scaled and the original problems are equivalent.

Scaled optimality conditions Outline ) AQp 1 Qp x) = 0 ) T AQp 1 y + Qp 1 s = 0, Q p x) Q p 1 s ) + Q p x) Q p 1s ) = 2µι Q p x) Q p 1s ). Simplifies to A x = 0 A T y + s = 0, Q p x) Q p 1 s ) + Q p x) Q p 1s ) = 2µι Q p x) Q p 1s ). The last equation cannot be simplified!

The choice of p AHO: p = ι: does not provide a nonsingular Newton system HKM: p = s 1/2 or p = x 1/2, in which case Q p 1s = ι or Q p x = ι. Implemented in SDPT3. NT: Most popular one. p = Q x 1/2 Q x 1/2s) 1/2) 1/2 = Q s 1/2 Q s 1/2x) 1/2) 1/2. Simplifies to Q p x = Q p 1s. Implemented in SeDuMi, MOSEK, SDPT3.

Centrality measures µx, s) = k i=1 x it s i n i. w = w 1 ;... ; w k ), where w i = Q 1/2 x s i. i δ F x, s) := Q x 1/2 s µι F := k λ 1 w i ) µ) 2 + λ 2 w i ) µ) 2 i=1 δ x, s) := Q x 1/2 s µι 2 := max { λ 1w i ) µ, λ 2 w i ) µ } i=1,...,k δ x, s) := Qx 1/2 s µι) := µ min {λ 1w i ), λ 2 w 2 )}, i=1,...,k Neighbourhoods δ x, s) δ x, s) δ F x, s). N γ) := {x, y, s) strictly feasible : δx, s) γµx, s)}. δx, s) = δ F x, s): narrow neighbourhood δx, s) = δ x, s) wide neighbourhood

IPM for SOCP Theorem Short-step IPM for SOCO) Choose γ = 0.088 and ζ = 0.06. Assume that we have a starting point x 0, y 0, s 0 ) N F γ). Compute the Newton step from the scaled Newton ) system. In every iteration, µ is decreased to 1 ζ k µ, i.e., θ = ζ k, and the stepsize is α = 1. This finds an ε-optimal solution for the second order conic optimization problem with k second order cones in at most ) k 1 O log ε iterations. Independent of m, n!) The cost of one iteration is ) k O m 3 + m 2 n +. i=1 n 2 i