Convex Optimization : Conic Versus Functional Form

Similar documents
On implementing a primal-dual interior-point method for conic quadratic optimization

15. Conic optimization

Conic Linear Optimization and its Dual. yyye

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Lecture 6: Conic Optimization September 8

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

IMPLEMENTATION OF INTERIOR POINT METHODS

New stopping criteria for detecting infeasibility in conic optimization

Solving conic optimization problems with MOSEK

Interior-Point Methods

Convex Optimization M2

Warmstarting the Homogeneous and Self-Dual Interior Point Method for Linear and Conic Quadratic Problems

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994)

Interior Point Algorithms for Constrained Convex Optimization

5. Duality. Lagrangian

Interior Point Methods for Mathematical Programming

18. Primal-dual interior-point methods

4TE3/6TE3. Algorithms for. Continuous Optimization

Lecture 18: Optimization Programming

Solving conic optimization problems via self-dual embedding and facial reduction: a unified approach

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Semidefinite Optimization using MOSEK

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Nonlinear Optimization: What s important?

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization Boyd & Vandenberghe. 5. Duality

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Optimization & Lagrange Duality

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

12. Interior-point methods

Primal-Dual Interior-Point Methods. Javier Peña Convex Optimization /36-725

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lagrangian Duality Theory

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming

Interior Point Methods in Mathematical Programming

A Brief Review on Convex Optimization

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization

ORIE 6326: Convex Optimization. Quasi-Newton Methods

Constrained Optimization and Lagrangian Duality

A path following interior-point algorithm for semidefinite optimization problem based on new kernel function. djeffal

Barrier Method. Javier Peña Convex Optimization /36-725

Analytic Center Cutting-Plane Method

Newton s Method. Javier Peña Convex Optimization /36-725

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Optimization for Machine Learning

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Lecture: Duality.

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Advances in Convex Optimization: Theory, Algorithms, and Applications

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization /36-725

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Apolynomialtimeinteriorpointmethodforproblemswith nonconvex constraints

Linear Programming: Simplex

Additional Homework Problems

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization

10 Numerical methods for constrained problems

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION


CSCI : Optimization and Control of Networks. Review on Convex Optimization

Detecting Infeasibility in Infeasible-Interior-Point Methods for Optimization

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

1 A HOMOGENIZED CUTTING PLANE

Course Outline. FRTN10 Multivariable Control, Lecture 13. General idea for Lectures Lecture 13 Outline. Example 1 (Doyle Stein, 1979)

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Pattern Classification, and Quadratic Problems

Algorithms for nonlinear programming problems II

On Two Measures of Problem Instance Complexity and their Correlation with the Performance of SeDuMi on Second-Order Cone Problems

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

On self-concordant barriers for generalized power cones

10. Unconstrained minimization

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725

Numerical Methods for PDE-Constrained Optimization

Lecture: Algorithms for LP, SOCP and SDP

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO

Primal-dual Subgradient Method for Convex Problems with Functional Constraints

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Tutorial on Convex Optimization: Part II

Primal-Dual Interior-Point Methods

12. Interior-point methods

Computational Finance

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

Constraint Reduction for Linear Programs with Many Constraints

Interior Point Methods for Convex Quadratic and Convex Nonlinear Programming

Convex Optimization and l 1 -minimization

5 Handling Constraints

Optimisation in Higher Dimensions

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Lecture 14: Primal-Dual Interior-Point Method

Lecture 9 Sequential unconstrained minimization

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Descent methods. min x. f(x)

The Q Method for Symmetric Cone Programmin

Transcription:

Convex Optimization : Conic Versus Functional Form Erling D. Andersen MOSEK ApS, Fruebjergvej 3, Box 16, DK 2100 Copenhagen, Blog: http://erlingdandersen.blogspot.com Linkedin: http://dk.linkedin.com/in/edandersen Email: e.d.andersen@mosek.com WWW: http://www.mosek.com http://www.mosek.com

2 / 33

Topic Topic Considerations about the forms Conic optimization. A recap Basic cone types Comments Given a convex optimization problem then is the classical form min c(x) st a(x) 0 or the conic form min c T x st Ax = b, x K. preferable from an user perspective? Assumptions c(x) is convex anda(x) is concave. K is a convex cone. We can only solve convex problems efficiently. 3 / 33

Considerations about the forms Topic Considerations about the forms Conic optimization. A recap Basic cone types Comments Robustness. Accuracy of the solution. How sensitive is the optimizer to the problem data. Solution time. Does the form help the user to build a good. Does it prevent nonconvex problems because checking convexity is hard. Ease of use. 4 / 33

Conic optimization. A recap Topic Considerations about the forms Conic optimization. A recap Basic cone types Comments Primal problem: (CO P ) min c T x st Ax = b, x K 1 K 2 K k, where eachk k is a cone (closed, pointed and solid). Dual problem: (CO P ) max b T y st A T y +s = c, s K1 K 2 K k, where eachk k is the dual convex cone i.e. K k {s : x T s 0, x K}. 5 / 33

Basic cone types Topic Considerations about the forms Conic optimization. A recap Basic cone types Comments Linear: Quadratic: K q := K l := {x R : x 0} x Rn : x 1 n j=2 x 2 j 6 / 33

Comments Topic Considerations about the forms Conic optimization. A recap Basic cone types Comments We will only deal with the above 2 self-dual cones. A quote from Stephen Boyd of Stanford: Almost all convex optimization problem can be formulated using a combination of linear, quadratic, semi-definite and the exponential cones. See also the book of Ben-Tal and Nemirovski for conic quadratic representable sets. 7 / 33

8 / 33

Convex quadratic optimization Convex quadratic optimization continued Practical issues Functional formulation: min 1 2 xt Qx+c T x st Ax = b, x 0. Given convexity then there exists a H such that Q = HH T. Conic reformulation min 1 2 t+ct x st Ax = b, H T x y = 0, s = 1, y 2 2st, x 0. 9 / 33

continued Convex quadratic optimization continued Practical issues Conic reformulation requires more constraints and variables. The addition is highly structured and sparse. Conic form is convex by construction. Q is frequently negative semi-definite due to rounding errors or mistakes. How to deal with: [ 1 1 1 1 ǫ ]? In practiceqis frequently low rank and dense. Implying the conic reformulation requires much less space to store. 10 / 33

Practical issues Convex quadratic optimization continued Practical issues Many users of the nonlinear capabilities in MOSEK: Do believe their nonlinear problems are convex. Do believe convexity is not important. Almost always make mistakes. (Errors in function or gradient computations). Checking and verifying convexity is an issue even for. Checking convexity and smoothness for arbitrary black box functions are hard. Data for the conic form is very simple. Just vectors, matrices and list of indexes. 11 / 33

12 / 33

The conic s The available s determines which form is preferable. Only primal-dual interior-point s are considered. They are efficient in practice. Polynomial complexity holds: For some. All conic. Claim: The conic form is more efficient and robust than functional form. Justification of claim: Follows. 13 / 33

functional form The conic s Problem: Assumptions: 1. x R n min c(x) st a i (x) 0, i, x 0. 2. c is convex and twice differentiable. 3. a i is concave and twice differentiable. 14 / 33

The conic s Lagrange function wherey 0. The dual problem: s R n. Alternatively: L(x,y) c(x) y T a(x), max L(x,y) x T s st (x) L(x,y) T = s, x,y,s 0, max L(x,y) x L(x,y)x st x L(x,y) T = s, x,y,s 0. 15 / 33

The conic s (MCP) min (x;y) T (s;z) st a(x) = z, st x L(x,y) T = s, x,z,y,s 0. wherez R m is slack variables. Note: The KKT is a monotone complementarity problem. Primal-dual interior-point type. Difficulties with detecting infeasibility! Feasibility constraints are potentially highly nonlinear! 16 / 33

The conic s denoted (HMCP): min (x;τ) T (s;κ) st τ a(x/τ) = z, primal feasibility τ x L(x/τ,y/τ) T = s, dual feasibility, x T x L(x/τ,y/τ) T y T a(x/τ) = κ, zero gap, x,z,τ,s,y,κ 0. Has two additional variablesτ,κ and one additional constraint. (HMCP) is a homogeneous monotone complementarity problem. Always has solution (0). 17 / 33

The conic s Theorem: (Andersen and Ye 95) Let (x,z,y,τ,s,κ ) be a maximal complementarity solution to (HMCP). Then (MCP) has a solution if and only if τ > 0. In this case, (x,z,y,s )/τ is a complementarity solution for(mcp). Conclusion: Compute a maximal solution to (HMCP). 18 / 33

The conic s Algorithm: Apply a primal-dual interior-point method to (HMCP) i.e. apply Newtons method to the perturbed optimality : τ x L(x/τ,y/τ) T = s, τa(x/τ) = z, x T x L(x/τ,y/τ) T y T a(x/τ) = κ, Xs = µe, Zy = µe, τκ = µ, x,z,τ,s,y,κ > 0. µ > 0: Is a parameter. Primal feasibility cond. is nonlinear. Dual feasibility cond. is nonlinear. Perturbed complementarity cond. is nonlinear. 19 / 33

The Step 0: Choose (x;z;τ;y;s;κ) > 0. Step 1: Choose parameters. Step 2: Compute the residuals: The where r r D r G r P s τg κ+x T g y T a(x/τ) z τa(x/τ) w (x;z;τ;y;s;κ) J a(x/τ), g x L(x/τ,y/τ) T H 2 xl(x/τ,y/τ)., conic s 20 / 33

The conic s Step 3: Solve the Newton equation system: and where H v2 J T (v 1 ) T H G (v 3 ) T J v 3 0 v 1 v 2 v 3 H g d x d τ d y d s d κ d z Sd x +Xd s = Xs+γµe, Yd z +Zd y = Zy +γµe, κd τ +τd κ = τκ+γµ, c(x/τ) T Hx/τ, c(x/τ) T Hx/τ, a(x/τ) Jx/τ, x/τ T H(x/τ), µ (x;z;τ) T (s;y;κ)/(n+m+1). = ηr 21 / 33

The conic s η = (γ 1) whereγ [0,1] is an ic parameters. Step 4: Update. Update of slacks: or where (x + ;τ + ;y + ;) = (x;τ;y)+α(d x ;d τ ;d y ) (z + ;s + ;κ + ) = (z +αd z ;s+αd s ;κ+αd κ ) z + = ((1+αη)r P +τ + a(x + /τ + ), s + = ((1+αη)r D +g + ), κ + = ((1+αη)r G +g + κ, g + τ + x L(x + /τ +,y + /τ + ) T g κ + (x + ) T x L(x + /τ +,y + /τ + ) T (y + ) T a(x + /τ + ) 22 / 33

The conic s Step 5: Repeat step 2 to 4. Issues: Gap and residuals are reduced by the factor(1+αη) in every iteration (given the nonlinear ). Polynomial complexity holds under certain assumptions. OBSERVE: z + = ((1+αη)r P +τ + a(x + /τ + ), s + = ((1+αη)r D +g + ). κ + = ((1+αη)r G +g + κ). Does not always work well!!! 23 / 33

The conic s Alternative. A linear z + = x+αd z, s + = s+αd s κ + = κ+αd κ but then a merit function is required e.g. two norm of the residuals. Works most of the time. Can lead to very slow convergence and robustness problems. A major weakness. Nonlinear primal and dual feasibility constraints are the cause of the problems. 24 / 33

conic quadratic problems The conic s : (H) Ax bτ = 0, (primal feasibility) A T y +s cτ = 0, (dual feasibility) c T x+b T y κ = 0, (zero gap) (x;τ) K,(s;κ) K In linear case suggested suggested Goldman and Tucker. Later rediscovered by Mizuno, Todd and Ye. All solutions satisfies: x T s+τκ = 0. If and only ifτ > 0, then an optimal solution exists. Find a solution such that τ +κ > 0. Linear primal and dual feasibility!!! Algorithmic framework: The Nesterov-Todd direction. 25 / 33

The conic s NT : Ad x bd τ A T d y +d s cd τ c T d x +b T d y d κ XT(ΘW) 1 d s + STΘWd x τd κ +κd τ whereη := γ 1 andγ [0,1). New iterate: x + x τ + τ y + = y s + s κ + κ = η(ax bτ), = η(a T y +s cτ), = η( c T x+b T y κ), = X Se+γµe, = τκ+γµ. +α d x d τ d y d s d κ. 26 / 33

The conic s Fact: Ax + bτ + = (1+αη)(Ax bτ), A T y + +s + cτ + = (1+αη)(A T y +s cτ), c T x + +b T y + κ + = (1+αη)( c T x+b T y κ), d T xd T s +d τ d κ = 0, (x + ) T s + +τ + κ + = (1+αη)((x) T s+τκ). Observations: NT can be seen as Newtons method. The method is symmetric! The complementarity gap is reduced by a factor of (1+αη) [0,1). The infeasibility is reduced by the same factor. No merit function is required. 27 / 33

s The conic s Both s are called primal-dual interior-point. Only in the conic form are the feasibility constraints linear. The functional one is more like primal the KKT system The conic is truly symmetric i.e. you can flip the primal and dual problem. Due to issues with the merit function then the functional form is not as robust as conic one. 28 / 33

The conic s Conic form: Functional form: min st min st y x = y x y, x R y x = y x 2 y y, y 0, x,y R 29 / 33

MOSEK v6 on the conic formulation The conic s ITE PFEAS DFEAS KAP/TAU POBJ DOBJ MU 0 1.0e+000 0.0e+000 1.0e+000 1.000e+000 0.0000e+000 1.0e+000 1 5.0e-002 0.0e+000 5.0e-002 5.000e-002 0.0000e+000 5.0e-002 2 2.5e-003 0.0e+000 2.5e-003 2.500e-003 0.0000e+000 2.5e-003 3 1.3e-004 0.0e+000 1.3e-004 1.250e-004 0.0000e+000 1.3e-004 4 6.3e-006 2.2e-016 6.3e-006 6.250e-006 0.0000e+000 6.3e-006 5 3.1e-007 0.0e+000 3.1e-007 3.128e-007 0.0000e+000 3.1e-007 6 1.6e-008 0.0e+000 1.6e-008 1.564e-008 0.0000e+000 1.6e-008 7 7.8e-010 1.1e-016 7.8e-010 7.821e-010 0.0000e+000 7.8e-010 30 / 33

MOSEK v6 on the ulation The conic s ITE PFEAS DFEAS KAP/TAU POBJ DOBJ MU 0 1.0e+000 1.4e+000 1.0e+000 1.000e+000 0.000e+000 1.0e+000 1 1.2e-001 6.4e-001 3.9e-003 1.648e-001 0.000e+000 1.1e-001 2 2.2e-003 6.8e-001 1.8e-003 1.972e-003 0.000e+000 2.0e-003... 50 3.2e-009 3.5e-003 1.5e-009 2.040e-009-4.135e-025 2.4e-009 51 2.9e-009 3.1e-003 1.3e-009 1.836e-009 0.000e+000 2.1e-009 52 2.6e-009 2.8e-003 1.2e-009 1.653e-009 0.000e+000 1.9e-009... 122 1.6e-012 1.4e-006 7.4e-013 1.036e-012 0.0000e+000 1.2e-012 123 1.5e-012 1.2e-006 6.7e-013 9.327e-013 0.0000e+000 1.1e-012 Performance on the ulation is bad. Merit function forces the alg. to take small steps. Behavior seen occasionally on large instances. Other interior-point optimizers may do better of course. 31 / 33

Summary The conic s The (restricted) conic formulation of nonlinear problems is advantageous because The problem data is simpler to deal in a software sense and no blackbox functions are required. It is impossible to formulate nonconvex problems. The available is much better. It easy to dualize the conic formulation. Caveats: The conic formulation is usually bigger in terms of number of variables and constraints but also highly structured. Usually the storage and time spend per iteration will be same or less for the conic formulation. 32 / 33

Conclusion The conic form is better than the! The conic s 33 / 33