PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.1/39

Similar documents
PENNON A Generalized Augmented Lagrangian Method for Nonconvex NLP and SDP p.1/22

PENNON A Code for Convex Nonlinear and Semidefinite Programming

On the solution of large-scale SDP problems by the modified barrier method using iterative solvers

On the solution of large-scale SDP problems by the modified barrier method using iterative solvers

On the solution of large-scale SDP problems by the modified barrier method using iterative solvers

Projection methods to solve SDP

Introducing PENLAB a MATLAB code for NLP-SDP

Solving large Semidefinite Programs - Part 1 and 2

PENNON: Software for linear and nonlinear matrix inequalities

Parallel implementation of primal-dual interior-point methods for semidefinite programs

III. Applications in convex optimization

Solving linear and non-linear SDP by PENNON

The Ongoing Development of CSDP

Notes on PCG for Sparse Linear Systems

Primal-Dual Geometry of Level Sets and their Explanatory Value of the Practical Performance of Interior-Point Methods for Conic Optimization

Lift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints

A CONIC DANTZIG-WOLFE DECOMPOSITION APPROACH FOR LARGE SCALE SEMIDEFINITE PROGRAMMING

10 Numerical methods for constrained problems

SDPARA : SemiDefinite Programming Algorithm PARAllel version

Truss topology design with integer variables made easy

Solving Nonconvex SDP Problems of Structural Optimization with Stability Control

Applications of Linear Programming

Effective reformulations of the truss topology design problem

Conjugate Gradient Method

A direct formulation for sparse PCA using semidefinite programming

Algorithms for nonlinear programming problems II

A direct formulation for sparse PCA using semidefinite programming

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Numerical Methods I Non-Square and Sparse Linear Systems

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Barrier Method. Javier Peña Convex Optimization /36-725

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

4TE3/6TE3. Algorithms for. Continuous Optimization

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Introduction to Semidefinite Programs

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

1 Computing with constraints

Implementation and Evaluation of SDPA 6.0 (SemiDefinite Programming Algorithm 6.0)

Exploiting hyper-sparsity when computing preconditioners for conjugate gradients in interior point methods

Algorithms for nonlinear programming problems II

Nonlinear Optimization: What s important?

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey

Sparse PCA with applications in finance

Nonlinear Programming

Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO

Nonlinear Optimization for Optimal Control

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Truncated Newton Method

Advances in Convex Optimization: Theory, Algorithms, and Applications

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Lecture 13: Constrained optimization

Course Outline. FRTN10 Multivariable Control, Lecture 13. General idea for Lectures Lecture 13 Outline. Example 1 (Doyle Stein, 1979)

Optimisation in Higher Dimensions

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

IPAM Summer School Optimization methods for machine learning. Jorge Nocedal

IP-PCG An interior point algorithm for nonlinear constrained optimization

Copositive Programming and Combinatorial Optimization

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Written Examination

LINEAR AND NONLINEAR PROGRAMMING

Outline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems

15. Conic optimization

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE

MS&E 318 (CME 338) Large-Scale Numerical Optimization

A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization

Algorithms for constrained local optimization

Line Search Methods for Unconstrained Optimisation

Sparse Covariance Selection using Semidefinite Programming

A direct formulation for sparse PCA using semidefinite programming

DELFT UNIVERSITY OF TECHNOLOGY

SUCCESSIVE LINEARIZATION METHODS FOR NONLINEAR SEMIDEFINITE PROGRAMS 1. Preprint 252 August 2003

Interior Point Methods for Mathematical Programming

From Stationary Methods to Krylov Subspaces

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 3. Optimization Problems and Iterative Algorithms

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

Interior-Point Methods as Inexact Newton Methods. Silvia Bonettini Università di Modena e Reggio Emilia Italy

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control

5 Handling Constraints

Research Reports on Mathematical and Computing Sciences

Interior-Point Methods for Linear Optimization

Miscellaneous Nonlinear Programming Exercises

The moment-lp and moment-sos approaches

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING

minimize x subject to (x 2)(x 4) u,

Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma

Image restoration: numerical optimisation

Computational Science and Engineering (Int. Master s Program) Preconditioning for Hessian-Free Optimization

5. Duality. Lagrangian

Lecture V. Numerical Optimization

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

An inexact block-decomposition method for extra large-scale conic semidefinite programming

An Inexact Newton Method for Nonlinear Constrained Optimization

A CONIC INTERIOR POINT DECOMPOSITION APPROACH FOR LARGE SCALE SEMIDEFINITE PROGRAMMING

Lecture: Duality of LP, SOCP and SDP

An inexact block-decomposition method for extra large-scale conic semidefinite programming

Transcription:

PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University kocvara@utia.cas.cz http://www.utia.cas.cz/kocvara PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.1/39

PBM Method for convex NLP Ben-Tal, Zibulevsky, 92, 97 Combination of: (exterior) Penalty meth., (interior) Barrier meth., Method of Multipliers Problem: (CP) min {f(x) : g i(x) 0, i = 1,..., m} x R n Assume: 1. f, g i ( i = 1,..., m ) convex 2. X nonempty and compact (A1) 3. ˆx so that g i (ˆx) < 0 for all i = 1,..., m (A2) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.2/39

PBM Method for convex NLP ϕ possibly smooth, domϕ possibly large (ϕ 0 ) ϕ strictly convex, strictly monotone increasing and C 2 (ϕ 1 ) domϕ = (, b) with 0 < b (ϕ 2 ) ϕ(0) = 0, (ϕ 4 ) lim ϕ (t) = t b (ϕ 3 ) ϕ (0) = 1, (ϕ 5 ) lim t ϕ (t) = 0 1 ϕ(t) b ϕ (t) b PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.3/39

PBM Method for convex NLP Examples: ϕ r 1 (t) = { c1 1 2 t2 + c 2 t + c 3 t r c 4 log(t c 5 ) + c 6 t < r. PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.4/39

PBM Method for convex NLP Examples: ϕ r 1 (t) = { c1 1 2 t2 + c 2 t + c 3 t r c 4 log(t c 5 ) + c 6 t < r. ϕ r 2 (t) = { c 1 1 2 t2 + c 2 t + c 3 t r, c 4 t c 5 + c 6 t < r, r 1, 1 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.4/39

PBM Method for convex NLP Examples: ϕ r 1 (t) = { c1 1 2 t2 + c 2 t + c 3 t r c 4 log(t c 5 ) + c 6 t < r. ϕ r 2 (t) = { c 1 1 2 t2 + c 2 t + c 3 t r, c 4 t c 5 + c 6 t < r, r 1, 1 Properties: C 2, bounded second derivative = improved behaviour of Newton s method composition of barrier branch (logarithmic/reciprocal) and penalty branch (quadratic) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.4/39

PBM algorithm for convex problems With p i > 0 for i {1,..., m}, we have g i (x) 0 p i ϕ(g i (x)/p i ) 0, i = 1,..., m The corresponding augmented Lagrangian: F(x, u, p) := f(x) + m u i p i ϕ(g i (x)/p i ) i=1 PBM algorithm: = arg min x IR uk, p k ) n u k+1 i = u k i ϕ (g i (x k+1 )/p k i ) i = 1,..., m x k+1 p k+1 i = π p k i i = 1,..., m PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.5/39

Properties of the PBM method Theory: {u k } k generated by PBM is the same as for a Proximal Point algorithm applied to the dual problem ( convergence proof) any cluster point of {x k } k is an optimal solution to (CP) f(x k ) f without p k 0 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.6/39

Properties of the PBM method Theory: {u k } k generated by PBM is the same as for a Proximal Point algorithm applied to the dual problem ( convergence proof) any cluster point of {x k } k is an optimal solution to (CP) f(x k ) f without p k 0 Praxis: fast convergence thanks to the barrier branch of ϕ particularly suitable for large sparse problems robust, typically 10 15 outer iterations and 40 80 Newton steps PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.6/39

PBM in semidefinite programming Problem: { min b T x : A(x) 0 } x R n Question: How can the matrix constraint be treated by PBM approach? A(x) 0 (A : R n S d convex) Idea: Find an augmented Lagrangian as follows: F(x, U, p) = f(x) + U, Φ p (A(x)) Sd PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.7/39

PBM in semidefinite programming Problem: { min b T x : A(x) 0 } x R n Question: How can the matrix constraint be treated by PBM approach? A(x) 0 (A : R n S d convex) Idea: Find an augmented Lagrangian as follows: F(x, U, p) = f(x) + U, Φ p (A(x)) Sd Notation: A, B Sd S d+ U S d+ Φ p := tr ( A T B ) inner product on S d = {A S d A positive semidefinite} matrix multiplier (dual variable) penalty function on S d PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.7/39

Construction of the penalty function Φ p, first idea Given: scalar valued penalty function ϕ satisfying (ϕ 0 ) (ϕ 5 ) matrix A = S ΛS, where Λ = diag (λ 1, λ 2,..., λ d ) Define A Φ p S T pϕ ( λ1 p ) 0 pϕ ( λ2 p.. 0... 0 ).... 0 0... 0 pϕ ( λd p ) S any positive eigenvalue of A is penalized by ϕ PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.8/39

PBM algorithm for semidefinite problems We have A(x) 0 Φ p (A(x)) 0 and the corresponding augmented Lagrangian: PBM algorithm: F(x, U, p) := f(x) + U, Φ p (A(x)) Sd (i) x k+1 = argmin x R n F(x, Uk, p k ) (ii) U k+1 = D A Φ p (A(x); U k ) (iii) p k+1 < p k PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.9/39

PBM algorithm for semidefinite problems The first idea may not be the best one: The matrix function Φ p corresponding to ϕ is convex but may be nonmonotone on H d (r, ) (right branch) may be nonconvex. U, Φ p (A(x)) Sd PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.10/39

PBM algorithm for semidefinite problems The first idea may not be the best one: The matrix function Φ p corresponding to ϕ is convex but may be nonmonotone on H d (r, ) (right branch) may be nonconvex. U, Φ p (A(x)) Sd Complexity of Hessian assembling O(d 4 + d 3 n + d 2 n 2 ) Even for a very sparse structure the complexity can be O(d 4 )! n... number of variables d... size of matrix constraint PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.10/39

PBM algorithm for semidefinite problems Hessian: [ xx U, Φ p (A(x)) Sd ] d k=1 i,j = ( [ ( sk (x) A i S(x) [ 2 ϕ(λ l (x), λ m (x), λ k (x))] n l,m=1 ] ) [S(x) US(x)] S(x) A j s k (x) ) S : s k : decomposition matrix of A(x) k-th column of S i divided difference of i-th order A : S d R n conjugate operator to A PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.11/39

Construction of the penalty function, second idea Find a penalty function ϕ which allows direct computation of Φ, its gradient and Hessian. PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.12/39

Construction of the penalty function, second idea Find a penalty function ϕ which allows direct computation of Φ, its gradient and Hessian. Example: (A(x) = x i A i ) ϕ(x) = x 2 Φ(A) = A 2 Then and Φ(A(x)) = A(x)A i + A i A(x) x i 2 x i x j Φ(A(x)) = A j A i + A i A j PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.12/39

Construction of the penalty function The reciprocal barrier function in SDP: (A(x) = x i A i ) ϕ := 1 t 1 1 The corresponding matrix function is and we can show that and 2 Φ(A) = (A I) 1 I x i Φ(A(x)) = (A I) 1 A i (A I) 1 x i x j Φ(A(x)) = (A I) 1 A i (A I) 1 A j (A I) 1 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.13/39

Construction of the penalty function Complexity of Hessian assembling: O(d 3 n + d 2 n 2 ) for dense matrices O(n 2 K 2 ) for sparse matrices (K... max. number of nonzeros in A i, i = 1,..., n) Compare to O(d 4 + d 3 n + d 2 n 2 ) in the general case { min b T x : A(x) 0 } x R n A : R n S d PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.14/39

Handling sparsity... essential for code efficiency { min b T x : A(x) 0 } x R n A = x i A i Three basic sparsity types: many (small) blocks sparse Hessian (multi-load truss/material) (A I) 1 A i (A I) 1 A j (A I) 1 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.15/39

Handling sparsity... essential for code efficiency { min b T x : A(x) 0 } x R n A = x i A i Three basic sparsity types: many (small) blocks sparse Hessian (multi-load truss/material) few (large) blocks (A I) 1 A i (A I) 1 A j (A I) 1 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.15/39

Handling sparsity... essential for code efficiency { min b T x : A(x) 0 } x R n A = x i A i Three basic sparsity types: many (small) blocks sparse Hessian (multi-load truss/material) few (large) blocks A dense, A i sparse (most of SDPLIB examples) (A I) 1 A i (A I) 1 A j (A I) 1 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.15/39

Handling sparsity x + 1 x + 2 x +... 3 full version as inefficient as general sparse version Recently, 3 matrix-matrix multiplication routines: full full full sparse sparse sparse PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.16/39

Handling sparsity essential for code efficiency { min b T x : A(x) 0 } x R n A = x i A i Three basic sparsity types: many (small) blocks sparse Hessian (multi-load truss/material) few (large) blocks A dense, A i sparse (most of SDPLIB examples) A sparse (truss design with buckling/vibration, maxg,... ) (A I) 1 A i (A I) 1 A j (A I) 1 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.17/39

Handling sparsity Fast inverse computation of sparse matrices Z = M 1 N Explicite inverse of M: O(n 3 ) Assume M is sparse and Cholesky factor L of M is sparse Z i = (L 1 ) T L 1 N i, i = 1,..., n Complexity: n times nk O(n 2 K) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.18/39

Numerical results New code called PENNON Comparison with DSDP by Benson and Ye SDPT3 by Toh, Todd and Tütüncü SeDuMi by Jos Sturm SDPLIB problems: http://www.nmt.edu/ sdplib/ PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.19/39

Numerical results problem variables matrix DSDP SDPT3 PENNON arch8 174 335 4 7 6 control7 136 45 114 48 82 control11 1596 165 1236 288 974 gpp500-4 501 500 28 39 21 mcp500-1 500 500 2 18 7 qap10 1021 101 19 8 16 ss30 132 426 10 18 20 theta6 4375 300 551 287 797 equalg11 801 801 139 156 102 equalg51 1001 1001 351 350 391 maxg11 800 800 6 54 25 maxg32 2000 2000 72 650 259 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.20/39

Examples from Mechanics Multiple-load Free Material Optimization After reformulation, discretization, further reformulation: min α R,x (R n ) L { α } L (c l ) T x l A i (α, x) 0, i = 1,..., m l=1 Many ( 5000) small (11 19) matrices. Large dimension (nl 20 000) In a standard form: { min x (R n ) L a T x nl i=1 x i B i 0 } x 1 + x 2 +... PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.21/39

Examples from Mechanics no. of size of problem var. matrix DSDP SDPT3 SeDuMi PENNON mater3 1439 3588 146 35 20 6 mater4 4807 12498 6269 295 97 29 mater5 10143 26820 36000 m 202 78 mater6 20463 56311 m m 533 233 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.22/39

Truss design with free vibration control Lowest eigenfrequency of the optimal structure should be bigger than a prescribed value min t,u s.t. ti A(t)u = f σ σ l (g(u) c) t T min. eigenfrequency a given value PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.23/39

Truss design with free vibration control Formulated as SDP problem: ti min t subject to A(t) ( λm(t) ) 0 c f T 0 f A(t) t i 0, i = 1,..., n where A(t) = t i A i A i = E i l 2 i γ i γ T i M(t) = t i M i M i = c diag(γ i γ T i ) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.24/39

truss test problems trto: problems from single-load truss topology design. Normally formulated as LP, here reformulated as SDP for testing purposes. vibra: single load truss topology problems with a vibration constraint. The constraint guarantees that the minimal self-vibration frequency of the optimal structure is bigger than a given value. buck: single load truss topology problems with linearized global buckling constraint. Originally a nonlinear matrix inequality, the constraint should guarantee that the optimal structure is mechanically stable (does not buckle). All problems characterized by sparsity of the matrix operator A. PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.25/39

truss test problems problem n m DSDP SDPT3 PENNON trto3 544 321+544 11 19 17 trto4 1200 673+1200 134 124 106 trto5 3280 1761+3280 3125 1422 1484 buck3 544 641+544 44 43 39 buck4 1200 1345+1200 340 241 221 buck5 3280 3521+3280 10727 2766 3006 vibra3 544 641+544 52 45 34 vibra4 1200 1345+1200 596 294 191 vibra5 3280 3521+3280 25290 3601 2724 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.26/39

Benchmark tests by Hans Mittelmann: http://plato.la.asu.edu/bench.html Implemented on the NEOS server: http://www-neos.anl.gov Homepage: http://www2.am.uni-erlangen.de/ kocvara/pennon/ http://www.penopt.com/ Available with MATLAB interface through TOMLAB http://www.tomlab.biz PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.27/39

When PCG helps (SDP)? Linear SDP, dense Hessian Complexity of Hessian evaluation A = n A i, A i R m m i=1 O(m 3 A n + m2 A n2 ) for dense matrices O(m 2 A n + K2 n 2 ) for sparse matrices (K... max. number of nonzeros in A i, i = 1,..., n) Complexity of Cholesky algorithm - linear SDP O(n 3 ) (... from PCG we expect O(n 2 )) Problems with large n and small m: CG better than Cholesky (expected) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.28/39

Iterative algorithms Conjugate Gradient method for Hd = g, H S n +...... y = Hx complexity O(n 2 )...... Exact arithmetics: convergence in n steps overall complexity O(n 3 ) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.29/39

Iterative algorithms Conjugate Gradient method for Hd = g, H S n +...... y = Hx complexity O(n 2 )...... Exact arithmetics: convergence in n steps Praxis: may be much worse (ill-conditioned problems) overall complexity O(n 3 ) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.29/39

Iterative algorithms Conjugate Gradient method for Hd = g, H S n +...... y = Hx complexity O(n 2 )...... Exact arithmetics: convergence in n steps Praxis: may be much worse (ill-conditioned problems) Praxis: may be much better preconditioning overall complexity O(n 3 ) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.29/39

Iterative algorithms Conjugate Gradient method for Hd = g, H S n +...... y = Hx complexity O(n 2 )...... Exact arithmetics: convergence in n steps Praxis: may be much worse (ill-conditioned problems) Praxis: may be much better preconditioning Convergence theory: number of iterations depends on condition number distribution of eigenvalues overall complexity O(n 3 ) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.29/39

Iterative algorithms Conjugate Gradient method for Hd = g, H S n +...... y = Hx complexity O(n 2 )...... Exact arithmetics: convergence in n steps Praxis: may be much worse (ill-conditioned problems) Praxis: may be much better preconditioning Convergence theory: number of iterations depends on condition number distribution of eigenvalues overall complexity O(n 3 ) Preconditioning: solve M 1 Hd = M 1 g with M H 1 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.29/39

Conditioning of Hessian Solve Hd = g, H Hessian of F(x, u, U, p, P) = f(x) + U, Φ P (A(x)) SmA Condition number depends on P Example: problem Theta2 from SDPLIB (n = 498) 2.5 3 2 1 2 1.5 0-1 1-2 0.5-3 0-4 -5-0.5 0 50 100 150 200 250 300 350 400 450 500-6 0 50 100 150 200 250 300 350 400 450 κ 0 = 394 κ opt = 4.9 10 7 500 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.30/39

Theta2 from SDPLIB (n = 498) 2.5 3 2 1 2 1.5 0-1 1-2 0.5-3 0-4 -5-0.5 0 50 100 150 200 250 300 350 400 450 500-6 0 50 100 150 200 250 300 350 400 450 500 Behaviour of CG: testing Hd + g / g 2 1 0 0.5-2 0-0.5-4 -1-6 -1.5-2 -8-2.5-10 0 2 4 6 8 10 12 14 16-3 0 100 200 300 400 500 600 700 800 900 1000 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.31/39

Theta2 from SDPLIB (n = 498) 2.5 3 2 1 2 1.5 0-1 1-2 0.5-3 0-4 -5-0.5 0 50 100 150 200 250 300 350 400 450 500-6 0 50 100 150 200 250 300 350 400 450 500 Behaviour of QMR: testing Hd + g / g 2 0 0-0.5-2 -1-4 -1.5-2 -6-2.5-8 -3-10 0 2 4 6 8 10 12 14 16-3.5 0 100 200 300 400 500 600 700 800 900 1000 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.31/39

Theta2 from SDPLIB (n = 498) 2.5 3 2 1 2 1.5 0-1 1-2 0.5-3 0-4 -5-0.5 0 50 100 150 200 250 300 350 400 450 500-6 0 50 100 150 200 250 300 350 400 450 500 QMR: effect of preconditioning (for small P ) 0 0-0.5-1 -1-2 -1.5-3 -2-4 -2.5-5 -3-6 -3.5 0 100 200 300 400 500 600 700 800 900 1000-7 0 100 200 300 400 500 600 700 800 900 1000 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.31/39

Control3 from SDPLIB (n = 136) 11 6 10 4 9 2 8 7 0 6-2 5-4 4 3-6 2 0 20 40 60 80 100 120 140-8 0 20 40 60 80 100 120 140 κ 0 = 3.1 10 8 κ opt = 7.3 10 12 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.32/39

Control3 from SDPLIB (n = 136) 11 6 10 4 9 2 8 7 0 6-2 5-4 4 3-6 2 0 20 40 60 80 100 120 140-8 0 20 40 60 80 100 120 140 Behaviour of CG: testing Hd + g / g 4 4 2 3.5 0 3-2 2.5-4 2-6 1.5-8 1-10 0 50 100 150 200 250 300 0.5 0 100 200 300 400 500 600 700 800 900 1000 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.32/39

Control3 from SDPLIB (n = 136) 11 6 10 4 9 2 8 7 0 6-2 5-4 4 3-6 2 0 20 40 60 80 100 120 140-8 0 20 40 60 80 100 120 140 Behaviour of QMR: testing Hd + g / g 0-1 0-2 -0.02-3 -0.04-4 -0.06-5 -0.08-6 -0.1-7 -0.12-8 -0.14-9 -0.16-10 0 50 100 150 200 250 300-0.18 0 100 200 300 400 500 600 700 800 900 1000 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.32/39

Preconditioners Should be: efficient (obvious but often difficult to reach) simple (low complexity) only use Hessian-vector product (NOT Hessian elements) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.33/39

Preconditioners Should be: efficient (obvious but often difficult to reach) simple (low complexity) only use Hessian-vector product (NOT Hessian elements) Diagonal Symmetric Gauss-Seidel L-BFGS (Morales-Nocedal, SIOPT 2000) A-inv (approximate inverse) (Benzi-Collum-Tuma, SISC 2000) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.33/39

Preconditioners Monteiro, O Neil, Nemirovski (2004): Adaptive Ellipsoid Preconditioner Improves the CG performance on extremely ill-conditioned systems. preconditioner: M = C k C T k, C k+1 αc k + βc k p k p T k, C 1 = γi α, β, p k... by matrix-vector products VERY preliminary results (MATLAB implementation) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.34/39

Preconditioners example Example: problem Theta2 from SDPLIB (n = 498) 3 2 1 0 1 2 3 4 5 6 0 50 100 150 200 250 300 350 400 450 500 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.35/39

Preconditioners example Example: problem Theta2 from SDPLIB (n = 498) 0 1 2 3 CG CG-diag 4 5 6 CG-SGS 7 8 9 10 0 100 200 300 400 500 600 700 800 900 1000 CG-MON. PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.36/39

Hessian free methods Use finite difference formula for Hessian-vector products: with h = (1 + x k 2 ε) 2 F(x k )v F(x k + hv) F(x k ) h Complexity: Hessian-vector product = gradient evaluation Complexity: need for Hessian-vector-product type preconditioner Limited accuracy (4 5 digits) PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.37/39

Test results: linear SDP, dense Hessian Stopping criterium for PENNON Exact Hessian: 10 7 (7 8 digits in objective function) Approximate Hessian: 10 4 (4 5 digits in objective function) Stopping criterium for CG/QMR??? Hd = g, stop when Hd + g / g ǫ PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.38/39

Test results: linear SDP, dense Hessian Stopping criterium for PENNON Exact Hessian: 10 7 (7 8 digits in objective function) Approximate Hessian: 10 4 (4 5 digits in objective function) Stopping criterium for CG/QMR??? Hd = g, stop when Hd + g / g ǫ Experiments: ǫ = 10 2 sufficient. often very low (average) number of CG iterations Complexity: n 3 kn 2, k 4 8 Practice: effect not that strong, due to other complexity issues PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.38/39

Problems with large n and small m Library of examples with large n and small m (courtesy of Kim Toh) CG-exact much better than Cholesky CG-approx much better than CG-exact PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.39/39

problem n m PENSDP PENSDP (APCG) SDPLR CPU CPU CG CPU iter ham_7_5_6 1793 128 126 4 52 1 113 ham_9_8 2305 512 423 210 66 46 222 ham_8_3_4 16129 256 81274 104 52 21 195 ham_9_5_6 53761 512 1984 71 71 102 theta42 200 5986 4722 51 269 393 11548 theta6 4375 300 2327 108 308 1221 20781 theta62 13390 300 68374 196 240 1749 16784 theta8 7905 400 11947 263 311 1854 15257 theta82 23872 400 m 650 267 4650 20653 theta83 39862 400 m 1715 277 7301 23017 theta10 12470 500 57516 492 278 4636 18814 theta102 37467 500 m 1948 340 12275 29083 theta103 62516 500 m 6149 421 17687 29483 theta104 87845 500 m 8400 269 theta12 17979 600 t 843 240 8081 21338 keller4 5101 171 3264 52 432 244 8586 sanr200-0.7 6033 200 6664 52 278 405 12139 problem theta83 theta103 theta123 n m PENSDP (APCG) RENDL 39862 400 460 345 440 62516 500 1440 491 850 90020 600 5286 1062 1530