Iterative LP and SOCP-based. approximations to. sum of squares programs. Georgina Hall Princeton University. Joint work with:

Similar documents
DSOS/SDOS Programming: New Tools for Optimization over Nonnegative Polynomials

Optimization over Structured Subsets of Positive Semidefinite Matrices via Column Generation

Optimization over Nonnegative Polynomials: Algorithms and Applications. Amir Ali Ahmadi Princeton, ORFE

Semidefinite Programming Basics and Applications

Complexity of 10 Decision Problems in Continuous Time Dynamical Systems. Amir Ali Ahmadi IBM Watson Research Center

DC Decomposition of Nonconvex Polynomials with Algebraic Techniques

Dimension reduction for semidefinite programming

SOS hierarchies we could have had in the 1920s

Computation of the Joint Spectral Radius with Optimization Techniques

ORF 523 Lecture 9 Spring 2016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Thursday, March 10, 2016

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

Semidefinite Programming

Fast ADMM for Sum of Squares Programs Using Partial Orthogonality

III. Applications in convex optimization

Model Reduction for Linear Parameter Varying Systems Using Scaled Diagonal Dominance

4. Algebra and Duality

A Hierarchy of Polyhedral Approximations of Robust Semidefinite Programs

Some applications of polynomial optimization in operations research and real-time decision making

Copositive Programming and Combinatorial Optimization

Towards Scalable Algorithms with Formal Guarantees for Lyapunov Analysis of Control Systems via Algebraic Optimization

Lecture Note 5: Semidefinite Programming for Stability Analysis

Analysis of Control Systems on Symmetric Cones

Lecture: Introduction to LP, SDP and SOCP

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

Analysis and synthesis: a complexity perspective

OPTIMIZATION OVER NONNEGATIVE AND CONVEX POLYNOMIALS WITH AND WITHOUT SEMIDEFINITE PROGRAMMING GEORGINA HALL A DISSERTATION PRESENTED TO THE FACULTY

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

Lecture: Cone programming. Approximating the Lorentz cone.

Global Optimization of Polynomials

A solution approach for linear optimization with completely positive matrices

Control System Analysis on Symmetric Cones

Copositive Programming and Combinatorial Optimization

Convex computation of the region of attraction for polynomial control systems

SDP Relaxations for MAXCUT

9. Interpretations, Lifting, SOS and Moments

Semidefinite Programming

Convex computation of the region of attraction for polynomial control systems

Complexity of Deciding Convexity in Polynomial Optimization

m i=1 c ix i i=1 F ix i F 0, X O.

SEMIDEFINITE PROGRAM BASICS. Contents

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

1 The independent set problem

Real Symmetric Matrices and Semidefinite Programming

Convex sets, conic matrix factorizations and conic rank lower bounds

Lecture: Examples of LP, SOCP and SDP

The maximal stable set problem : Copositive programming and Semidefinite Relaxations

12. Interior-point methods

Linear Programming Inverse Projection Theory Chapter 3

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Robust and Optimal Control, Spring 2015

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Lec12p1, ORF363/COS323. The idea behind duality. This lecture

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Facial Reduction and Geometry on Conic Programming

Agenda. 1 Cone programming. 2 Convex cones. 3 Generalized inequalities. 4 Linear programming (LP) 5 Second-order cone programming (SOCP)

Chapter 1 Linear Programming. Paragraph 5 Duality

The Ongoing Development of CSDP

arxiv: v1 [math.oc] 31 Jan 2017

Reduction methods in semidefinite and conic optimization

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

Region of attraction approximations for polynomial dynamical systems

Positive semidefinite rank

Continuous Optimisation, Chpt 9: Semidefinite Optimisation

V&V MURI Overview Caltech, October 2008

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lecture 14: Optimality Conditions for Conic Problems

Convex Optimization M2

Conic Linear Optimization and its Dual. yyye

There are two things that are particularly nice about the first basis

Chapter 1: Systems of Linear Equations and Matrices

Lecture notes for quantum semidefinite programming (SDP) solvers

Lecture 20: November 1st

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Linear and non-linear programming

Module 8 Linear Programming. CS 886 Sequential Decision Making and Reinforcement Learning University of Waterloo

Linear Programming Redux

Agenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization

Introduction to Semidefinite Programming I: Basic properties a

On approximations, complexity, and applications for copositive programming Gijben, Luuk

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

Massachusetts Institute of Technology 6.854J/18.415J: Advanced Algorithms Friday, March 18, 2016 Ankur Moitra. Problem Set 6

The moment-lp and moment-sos approaches

Lecture 8: Semidefinite programs for fidelity and optimal measurements

The Simplest Semidefinite Programs are Trivial

Equivariant semidefinite lifts and sum of squares hierarchies

1 Robust optimization

Lecture 6: Conic Optimization September 8

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 12 Luca Trevisan October 3, 2017

Lecture 6 Verification of Hybrid Systems

Lecture 8 : Eigenvalues and Eigenvectors

Decentralized Control of Stochastic Systems

EE364 Review Session 4

Equations in Quadratic Form

IE 521 Convex Optimization

MIT Algebraic techniques and semidefinite optimization May 9, Lecture 21. Lecturer: Pablo A. Parrilo Scribe:???

Introduction and Math Preliminaries

Transcription:

Iterative LP and SOCP-based approximations to sum of squares programs Georgina Hall Princeton University Joint work with: Amir Ali Ahmadi (Princeton University) Sanjeeb Dash (IBM)

Sum of squares programs Problems of the type: Decision variables are the coefficients of the polynomial p p p x sos Linear objective and affine constraints in the coefficients of p (e.g., sum of coefs =1) Sum of squares condition Many applications for problems of this form 2

Semidefinite programming formulation Sum of squares program: p p sos Equivalent semidefinite programming formulation: min p,q C(p) p = z x T Qz x Q 0 But: Size of Q = n+d d n+d d 3

Alternatives to sum of squares: dsos and sdsos Sum of squares (sos) p x = z x T Qz x, Q 0 SDP PSD cone Q Q 0} DD cone Q Q ii j Q ij, i} SDD cone Q diagonal D with D ii > 0 s.t. DQD dd} Diagonally dominant sum of squares (dsos) p x = z x T Qz x, Q diagonally dominant (dd) LP Scaled diagonally dominant sum of squares (sdsos) p x = z x T Qz x, Q scaled diagonally dominant (sdd) SOCP [Ahmadi, Majumdar] 4

Alternatives to sum of squares: dsos and sdsos p sos scalability p dsos/sdsos b Example: For a parametric family of polynomials: p x 1, x 2 = 2x 1 4 + 2x 2 4 + ax 1 3 x 2 + (1 a)x 1 2 x 2 2 + bx 1 x 2 3 a 5

Alternatives to sum of squares: dsos and sdsos Example: Stabilizing the inverted N-link pendulum (2N states) N=1 Runtime: N=2 N=6 ROA volume ratio: [Ahmadi, Majumdar, Tedrake]

Improvements on dsos and sdsos Replacing sos polynomials by dsos/sdsos polynomials: +: fast bounds - : not always as good quality (compared to sos) Iteratively construct a sequence of improving LP/SOCPs Initialization: Start with the dsos/sdsos polynomials Method 1: Cholesky change of basis Method 2: Column Generation Method 3: r-s/dsos hierarchy 7

Method 1: Cholesky change of basis (1/3) dd in the right basis psd but not dd Goal: iteratively improve on basis. 8

Method 1: Cholesky change of basis (2/3) Sos problem p sos Initialize p = z x T Qz x, Q dd/sdd Step 1 Replace: T U U 1 = chol(q k = chol(u k 1 Q ) U k 1 ) k k + 1 Step 2 p = z x T U T k1 QU 1k z x, Q dd/sdd New basis One iteration of this method on a parametric family of polynomials: p x 1, x 2 = 2x 1 4 + 2x 2 4 + ax 1 3 x 2 + (1 a)x 1 2 x 2 2 + bx 1 x 2 3 9

Lower bound on optimal value CDC 2017 Method 1: Cholesky change of basis (3/3) Example: minimizing a degree-4 polynomial in 4 variables 10

Method 2: Column generation (1/4) Focus on LP-based version of this method (SOCP is similar). Sos problem p sos Replace with Dsos problem p dsos Dsos problem p(x) = z x T Qz x, Q dd Two different ways of characterizing Q dd: 1 Q dd Q ii j Q ij, i 2 Q dd α i 0 s.t. Q = i α i v i v i T, where v i fixed vector with at most two nonzero components = ±1 [In collaboration with Sanjeeb Dash, IBM Research] 11

Method 2: Column generation (2/4) Dsos problem p(x) = z x T Qz x, Q dd Using 2 Dsos problem p x = i α i v i T z x 2, α i 0 Idea behind the algorithm: Expand feasible space at each iteration by adding a new vector v and variable α p x = i α i v i T z x 2 + α v T z x 2 α i 0, a 0 Question: How to pick v? Use the dual! 12

Method 2: Column generation (3/4) PRIMAL DUAL A general SDP SDP Dual of SDP max y R m bt y m s. t. C y i A i 0 LP min tr(cx) X Sn s. t. tr A i X = b i X 0 SDP LP i=1 LP obtained with inner approximation of PSD by DD s. t. C max y R m bt y m i=1 y i A i = α i 0 α i v i v i T Dual of LP min tr(cx) X Sn s. t. tr A i X = b i v T i Xv i 0 Pick v s.t. v T Xv < 0. 13

Method 2: Column Generation (4/4) Example 2: minimizing a degree-4 polynomial n = 15 n = 20 n = 25 n = 30 n = 40 bd t(s) bd t(s) bd t(s) bd t(s) bd t(s) DSOS -10.96 0.38-18.01 0.74-26.45 15.51-36.52 7.88-62.30 10.68 DSOS k -5.57 31.19-9.02 471.39-20.08 600-32.28 600-35.14 600 SOS -3.26 5.60-3.58 82.22-3.71 1068.66 NA NA NA NA 14

Method 3: r-s/dsos hierarchy (1/3) A polynomial p is r-dsos if p x i x i 2 r is dsos. A polynomial p is r-sdsos if p x i x i 2 r is sdsos. Defines a hierarchy based on r. Theorem Any even positive definite form p is r-dsos for some r. Proof: Follows from a result by Polya. Proof of positivity using LP. [Ahmadi, Majumdar]

Method 3: r-s/dsos hierarchy (2/3) Example: certifying stability of a switched linear system x k+1 = A σ k x k where A σ k {A 1,, A m } Recall: Theorem 1: A switched linear system is stable if and only if ρ A 1, A m < 1. Theorem 2 [Parrilo, Jadbabaie]: ρ A 1,, A m < 1 a pd polynomial Lyapunov function V(x) such that V x V A i x > 0, x 0.

Method 3: r-s/dsos hierarchy (3/3) Theorem: For nonnegative A 1,, A m, ρ A 1,, A m < 1 r N and a polynomial Lyapunov function V(x) such that V x. 2 r-dsos and V x. 2 V A i x. 2 r-dsos. ( ) Proof: V x 0 and V x V A i x 0 for any x 0. Combined to A i 0, this implies that trajectories of x k+1 = A σ(k) x k starting from x 0 0 go to zero. This can be extended to any x 0 by noting that x 0 = x 0 + x 0, x 0 +, x 0 0. ( ) From Theorem 2, and using Polya s result as V(x. 2 ) and V x. 2 V A i x. 2 are even forms.

Main messages Can construct iterative inner approximations of the cone of nonnegative polynomials using LPs and SOCPs. Presented three methods: Cholesky change of basis Column Generation r-s/dsos hierarchies Initialization Initialize with dsos/sdsos polynomials Method Rotate existing atoms of the cone of dsos/sdsos polynomials Add new atoms to the extreme rays of the cone of dsos/sdsos polynomials Use multipliers to certify nonnegativity of more polynomials. Size of the LP/SOCPs obtained Does not grow (but possibly denser) Grows slowly Grows quickly Objective taken into consideration Yes Yes No Can beat the SOS bound No No Yes

Thank you for listening Questions? Want to learn more? http://scholar.princeton.edu/ghall/ 19