The moment-lp and moment-sos approaches

Similar documents
The moment-lp and moment-sos approaches in optimization

A new look at nonnegativity on closed sets

Moments and Positive Polynomials for Optimization II: LP- VERSUS SDP-relaxations

Moments and Positive Polynomials for Optimization II: LP- VERSUS SDP-relaxations

On Polynomial Optimization over Non-compact Semi-algebraic Sets

Optimization over Polynomials with Sums of Squares and Moment Matrices

Convex Optimization & Parsimony of L p-balls representation

Convergence rates of moment-sum-of-squares hierarchies for volume approximation of semialgebraic sets

Hilbert s 17th Problem to Semidefinite Programming & Convex Algebraic Geometry

CONVEXITY IN SEMI-ALGEBRAIC GEOMETRY AND POLYNOMIAL OPTIMIZATION

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

Minimum Ellipsoid Bounds for Solutions of Polynomial Systems via Sum of Squares

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Solving Global Optimization Problems with Sparse Polynomials and Unbounded Semialgebraic Feasible Sets

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

Strong duality in Lasserre s hierarchy for polynomial optimization

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

Introduction to Semidefinite Programming I: Basic properties a

Semidefinite Programming Basics and Applications

4. Algebra and Duality

Approximate Optimal Designs for Multivariate Polynomial Regression

arxiv: v1 [math.oc] 31 Jan 2017

A brief history of noncommutative Positivstellensätze. Jaka Cimprič, University of Ljubljana, Slovenia

An explicit construction of distinguished representations of polynomials nonnegative over finite sets

How to generate weakly infeasible semidefinite programs via Lasserre s relaxations for polynomial optimization

Semialgebraic Relaxations using Moment-SOS Hierarchies

Robust and Optimal Control, Spring 2015

Lecture 6: Conic Optimization September 8

Optimality, Duality, Complementarity for Constrained Optimization

c 2000 Society for Industrial and Applied Mathematics

Semidefinite approximations of projections and polynomial images of semialgebraic sets

15. Conic optimization

What can be expressed via Conic Quadratic and Semidefinite Programming?

Constrained Optimization Theory

Semidefinite programming and convex algebraic geometry

SEMIDEFINITE PROGRAMMING VS. LP RELAXATIONS FOR POLYNOMIAL PROGRAMMING

Representations of Positive Polynomials: Theory, Practice, and

Semidefinite programming relaxations for semialgebraic problems

Assignment 1: From the Definition of Convexity to Helley Theorem

arxiv: v1 [math.oc] 9 Sep 2015

A JOINT+MARGINAL APPROACH TO PARAMETRIC POLYNOMIAL OPTIMIZATION

that a broad class of conic convex polynomial optimization problems, called

Polyhedral Results for A Class of Cardinality Constrained Submodular Minimization Problems

Optimization over Nonnegative Polynomials: Algorithms and Applications. Amir Ali Ahmadi Princeton, ORFE

m i=1 c ix i i=1 F ix i F 0, X O.

Optimization based robust control

Global Optimization with Polynomials

An Exact Jacobian SDP Relaxation for Polynomial Optimization

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

Course Outline. FRTN10 Multivariable Control, Lecture 13. General idea for Lectures Lecture 13 Outline. Example 1 (Doyle Stein, 1979)

Convex Optimization in Classification Problems

Mean squared error minimization for inverse moment problems

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Numerical Optimization

Robust linear optimization under general norms

Conic Linear Programming. Yinyu Ye

Minimum volume semialgebraic sets for robust estimation

Lecture 5. The Dual Cone and Dual Problem

The Trust Region Subproblem with Non-Intersecting Linear Constraints

Convex computation of the region of attraction of polynomial control systems

Formal Proofs, Program Analysis and Moment-SOS Relaxations

Semidefinite Programming

Advanced SDPs Lecture 6: March 16, 2017

Lecture 18: Optimization Programming

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 2

Inner approximations of the region of attraction for polynomial dynamical systems

Relations between Semidefinite, Copositive, Semi-infinite and Integer Programming

Optimality Conditions for Constrained Optimization

Semidefinite Programming

Primal-Dual Geometry of Level Sets and their Explanatory Value of the Practical Performance of Interior-Point Methods for Conic Optimization

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Sum of Squares Relaxations for Polynomial Semi-definite Programming

Towards Solving Bilevel Optimization Problems in Quantum Information Theory

Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A.

DSOS/SDOS Programming: New Tools for Optimization over Nonnegative Polynomials

Copositive Programming and Combinatorial Optimization

Robust conic quadratic programming with ellipsoidal uncertainties

12. Interior-point methods

Module 04 Optimization Problems KKT Conditions & Solvers

Conic Linear Programming. Yinyu Ye

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

Network Utility Maximization With Nonconcave Utilities Using Sum-of-Squares Method

Non-commutative polynomial optimization

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

Linear conic optimization for nonlinear optimal control

A new approximation hierarchy for polynomial conic optimization

Analysis and synthesis: a complexity perspective

Lecture Note 5: Semidefinite Programming for Stability Analysis

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

on Semialgebraic Sets for Nonnnegative Polynomials Finding Representations

Non-Convex Optimization via Real Algebraic Geometry

Convex computation of the region of attraction of polynomial control systems

Tilburg University. Hidden Convexity in Partially Separable Optimization Ben-Tal, A.; den Hertog, Dick; Laurent, Monique. Publication date: 2011

Modern Optimal Control

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones

Mean squared error minimization for inverse moment problems

Transcription:

The moment-lp and moment-sos approaches LAAS-CNRS and Institute of Mathematics, Toulouse, France CIRM, November 2013

Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators

Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators

Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators

Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators

Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators

Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators

Semidefinite Programming The CONVEX optimization problem: P min x R n { c x n A i x i i=1 b}, where c R n and b, A i S m (m m symmetric matrices), is called a semidefinite program. The notation 0" means the real symmetric matrix " is positive semidefinite, i.e., all its (real) EIGENVALUES are nonnegative.

Example P : min {x 1 + x 2 : x [ 3 + 2x1 + x s.t. 2 x 1 5 x 1 5 x 1 2x 2 or, equivalently ] }, 0 P : min {x 1 + x 2 : x [ 3 5 s.t. 5 0 ] + x 1 [ 2 1 1 1 ] + x 2 [ 1 0 0 2 ] } 0

P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...

P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...

P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...

Why Polynomial Optimization? After all... the polynomial optimization problem: f = min{f (x) : g j (x) 0, j = 1,..., m} is just a particular case of Non Linear Programming (NLP)! True!... if one is interested with a LOCAL optimum only!!

Why Polynomial Optimization? After all... the polynomial optimization problem: f = min{f (x) : g j (x) 0, j = 1,..., m} is just a particular case of Non Linear Programming (NLP)! True!... if one is interested with a LOCAL optimum only!!

When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!

When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!

When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!

BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}.... and so to compute f one needs TRACTABLE CERTIFICATES of POSITIVITY on K!

BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}.... and so to compute f one needs TRACTABLE CERTIFICATES of POSITIVITY on K!

BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}.... and so to compute f one needs TRACTABLE CERTIFICATES of POSITIVITY on K!

REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)

REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)

REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)

SOS-based certificate K = {x : g j (x) 0, j = 1,..., m} Theorem (Putinar s Positivstellensatz) If K is compact (+ a technical Archimedean assumption) and f > 0 on K then: f (x) = σ 0 (x) + m σ j (x) g j (x), x R n, j=1 for some SOS polynomials (σ j ) R[x]. Testing whether holds for some SOS (σ j ) R[x] with a degree bound, is SOLVING an SDP!

SOS-based certificate K = {x : g j (x) 0, j = 1,..., m} Theorem (Putinar s Positivstellensatz) If K is compact (+ a technical Archimedean assumption) and f > 0 on K then: f (x) = σ 0 (x) + m σ j (x) g j (x), x R n, j=1 for some SOS polynomials (σ j ) R[x]. Testing whether holds for some SOS (σ j ) R[x] with a degree bound, is SOLVING an SDP!

LP-based certificate K = {x : g j (x) 0; (1 g j (x)) 0, j = 1,..., m} Theorem (Krivine-Vasilescu-Handelman s Positivstellensatz) Let K be compact and the family {g j, (1 g j )} generate R[x]. If f > 0 on K then: f (x) = α,β c αβ m j=1 g j (x) α j (1 g j (x)) β j,, x R n, for some NONNEGATIVE scalars (c αβ ). Testing whether holds for some NONNEGATIVE (c αβ ) with α + β M, is SOLVING an LP!

LP-based certificate K = {x : g j (x) 0; (1 g j (x)) 0, j = 1,..., m} Theorem (Krivine-Vasilescu-Handelman s Positivstellensatz) Let K be compact and the family {g j, (1 g j )} generate R[x]. If f > 0 on K then: f (x) = α,β c αβ m j=1 g j (x) α j (1 g j (x)) β j,, x R n, for some NONNEGATIVE scalars (c αβ ). Testing whether holds for some NONNEGATIVE (c αβ ) with α + β M, is SOLVING an LP!

SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY! Farkas Lemma Krivine-Stengle KKT-Optimality conditions Schmüdgen-Putinar

SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY! Farkas Lemma Krivine-Stengle KKT-Optimality conditions Schmüdgen-Putinar

SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY! Farkas Lemma Krivine-Stengle KKT-Optimality conditions Schmüdgen-Putinar

SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY! Farkas Lemma Krivine-Stengle KKT-Optimality conditions Schmüdgen-Putinar

In addition, polynomials NONNEGATIVE ON A SET K R n are ubiquitous. They appear in many important applications, and not only in global optimization! For instance, one may also want: To approximate sets defined with QUANTIFIERS, like.e.g., R f := {x B : f (x, y) 0 for all y such that (x, y) K} D f := {x B : f (x, y) 0 for some y such that (x, y) K} where f R[x, y], B is a simple set (box, ellipsoid). To compute convex polynomial underestimators p f of a polynomial f on a box B R n. (Very useful in MINLP.)

In addition, polynomials NONNEGATIVE ON A SET K R n are ubiquitous. They appear in many important applications, and not only in global optimization! For instance, one may also want: To approximate sets defined with QUANTIFIERS, like.e.g., R f := {x B : f (x, y) 0 for all y such that (x, y) K} D f := {x B : f (x, y) 0 for some y such that (x, y) K} where f R[x, y], B is a simple set (box, ellipsoid). To compute convex polynomial underestimators p f of a polynomial f on a box B R n. (Very useful in MINLP.)

The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Stengle s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In may situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.

The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Stengle s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In may situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.

The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Stengle s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In may situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.

The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Stengle s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In may situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.

LP- and SDP-hierarchies for optimization Replace f = sup λ,σj { λ : f (x) λ 0 x K} with: The SDP-hierarchy indexed by d N: m fd = sup { λ : f λ = σ 0 + σ }{{} j g j ; deg (σ j g j ) 2d } }{{} SOS j=1 SOS or, the LP-hierarchy indexed by d N: θ d = sup { λ : f λ = α,β c αβ }{{} 0 m α g j j (1 g j ) β j ; α + β 2d} j=1

LP- and SDP-hierarchies for optimization Replace f = sup λ,σj { λ : f (x) λ 0 x K} with: The SDP-hierarchy indexed by d N: m fd = sup { λ : f λ = σ 0 + σ }{{} j g j ; deg (σ j g j ) 2d } }{{} SOS j=1 SOS or, the LP-hierarchy indexed by d N: θ d = sup { λ : f λ = α,β c αβ }{{} 0 m α g j j (1 g j ) β j ; α + β 2d} j=1

Theorem Both sequence (f d ), and (θ d), d N, are MONOTONE NON DECREASING and when K is compact (and satisfies a technical Archimedean assumption) then: f = lim fd = lim θ d. d d

What makes this approach exciting is that it is at the crossroads of several disciplines/applications: Commutative, Non-commutative, and Non-linear ALGEBRA Real algebraic geometry, and Functional Analysis Optimization, Convex Analysis Computational Complexity in Computer Science, which BENEFIT from interactions! As mentioned... potential applications are ENDLESS!

What makes this approach exciting is that it is at the crossroads of several disciplines/applications: Commutative, Non-commutative, and Non-linear ALGEBRA Real algebraic geometry, and Functional Analysis Optimization, Convex Analysis Computational Complexity in Computer Science, which BENEFIT from interactions! As mentioned... potential applications are ENDLESS!

Has already been proved useful and successful in applications with modest problem size, notably in optimization, control, robust control, optimal control, estimation, computer vision, etc. HAS initiated and stimulated new research issues: in Convex Algebraic Geometry (e.g. semidefinite representation of convex sets, algebraic degree of semidefinite programming and polynomial optimization) in Computational algebra (e.g., for solving polynomial equations via SDP and Border bases) Computational Complexity where LP- and SDP-HIERARCHIES have become an important tool to analyze Hardness of Approximation for 0/1 combinatorial problems ( links with quantum computing)

A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.

A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.

A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.

Even though the moment-sos approach DOES NOT SPECIALIZES to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. (Finite convergence also occurs for general convex problems.) (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems!

Even though the moment-sos approach DOES NOT SPECIALIZES to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. (Finite convergence also occurs for general convex problems.) (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems!

Even though the moment-sos approach DOES NOT SPECIALIZES to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. (Finite convergence also occurs for general convex problems.) (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems!

A remarkable property: II FINITE CONVERGENCE of the SOS-hierarchy is GENERIC!... and provides a GLOBAL OPTIMALITY CERTIFICATE, the analogue for the NON CONVEX CASE of the KKT-OPTIMALITY conditions in the CONVEX CASE!

Theorem (Marshall, Nie) Let x K be a global minimizer of P : f = min {f (x) : g j (x) 0, j = 1,..., m}. and assume that: (i) The gradients { g j (x )} are linearly independent, (ii) Strict complementarity holds (λ j g j (x ) = 0 for all j.) (iii) Second-order sufficiency conditions hold at (x, λ ) K R m +. m Then f (x) f = σ0 (x) + σj (x)g j(x), SOS polynomials {σ j }. j=1 x R n, for some Moreover, the conditions (i)-(ii)-(iii) HOLD GENERICALLY!

Theorem (Marshall, Nie) Let x K be a global minimizer of P : f = min {f (x) : g j (x) 0, j = 1,..., m}. and assume that: (i) The gradients { g j (x )} are linearly independent, (ii) Strict complementarity holds (λ j g j (x ) = 0 for all j.) (iii) Second-order sufficiency conditions hold at (x, λ ) K R m +. m Then f (x) f = σ0 (x) + σj (x)g j(x), SOS polynomials {σ j }. j=1 x R n, for some Moreover, the conditions (i)-(ii)-(iii) HOLD GENERICALLY!

In summary: KKT-OPTIMALITY when f and g j are CONVEX m f (x ) λ j g j(x ) = 0 f (x) f j=1 m λ j g j(x) j=1 PUTINAR s CERTIFICATE in the non CONVEX CASE m f (x ) σ j (x ) g j (x ) = 0 f (x) f j=1 m σj (x)g j(x) j=1 0 for all x R n (= σ0 (x)) 0 for all x Rn. for some SOS {σj }, and σj (x ) = λ j.

In summary: KKT-OPTIMALITY when f and g j are CONVEX m f (x ) λ j g j(x ) = 0 f (x) f j=1 m λ j g j(x) j=1 PUTINAR s CERTIFICATE in the non CONVEX CASE m f (x ) σ j (x ) g j (x ) = 0 f (x) f j=1 m σj (x)g j(x) j=1 0 for all x R n (= σ0 (x)) 0 for all x Rn. for some SOS {σj }, and σj (x ) = λ j.

II. Approximation of sets with quantifiers Let f R[x, y] and let K R n R p be the semi-algebraic set: K := {(x, y) : g j (x, y) 0, j = 1,..., m}, and let B R n be the unit ball or the [ 1, 1] n. Suppose that one wants to approximate the set: R f := {x B : f (x, y) 0 for all y such that (x, y) K} as closely as desired by a sequence of sets of the form: Θ k := {x B : J k (x) 0 } for some polynomials J k.

With g 0 = 1 and with K R n R p and k N, let m Q k (g) := σ j (x, y) g j (x, y) : σ j Σ[x, y], deg σ j g j 2k j=0 Let x F(x) := max {f (x, y) : (x, y) K }, and for every integer k consider the optimization problem: ρ k = min J R[x] k { } (J F ) dx : J(x) f (x, y) Q k (g) B

1. The criterion B (J F) dx = F dx B }{{} + α unknown but constant J α x α dx B }{{} easy to compute is LINEAR in the coefficients J α of the unknown polynomial J R[x] k! 2. The constraint J(x) f (x, y) = m σ j (x, y) g j (x, y) j=0 is just LINEAR CONSTRAINTS + LMIs!

1. The criterion B (J F) dx = F dx B }{{} + α unknown but constant J α x α dx B }{{} easy to compute is LINEAR in the coefficients J α of the unknown polynomial J R[x] k! 2. The constraint J(x) f (x, y) = m σ j (x, y) g j (x, y) j=0 is just LINEAR CONSTRAINTS + LMIs!

ρ k = min J R[x] k Hence, the optimization problem { } (J F ) dx : J(x) f (x, y) Q k (g) B IS AN SDP! Moreover, it has an optimal solution J k R[x] k! Alternatively, if one uses LP-based positivity certificates for J(x) f (x, y), one ends up with solving an LP! From the definition of Jk, the sublevel sets Θ k := {x B : J k (x) 0} R f, k N, provide a nested sequence of INNNER approximations of R f.

ρ k = min J R[x] k Hence, the optimization problem { } (J F ) dx : J(x) f (x, y) Q k (g) B IS AN SDP! Moreover, it has an optimal solution J k R[x] k! Alternatively, if one uses LP-based positivity certificates for J(x) f (x, y), one ends up with solving an LP! From the definition of Jk, the sublevel sets Θ k := {x B : J k (x) 0} R f, k N, provide a nested sequence of INNNER approximations of R f.

ρ k = min J R[x] k Hence, the optimization problem { } (J F ) dx : J(x) f (x, y) Q k (g) B IS AN SDP! Moreover, it has an optimal solution J k R[x] k! Alternatively, if one uses LP-based positivity certificates for J(x) f (x, y), one ends up with solving an LP! From the definition of Jk, the sublevel sets Θ k := {x B : J k (x) 0} R f, k N, provide a nested sequence of INNNER approximations of R f.

Theorem (Lass) (Strong) convergence in L 1 (B)-norm takes place, that is: Jk F dx = 0 lim k B and, if in addition the set {x B : F(x) = 0} has Lebesgue measure zero, then lim VOL(R f \ Θ k ) = 0 k

Theorem (Lass) (Strong) convergence in L 1 (B)-norm takes place, that is: Jk F dx = 0 lim k B and, if in addition the set {x B : F(x) = 0} has Lebesgue measure zero, then lim VOL(R f \ Θ k ) = 0 k

Ex: Polynomial Matrix Inequalities: (with D. Henrion) Let x A(x) R p p where A(x) is the matrix-polynomial ) x A(x) = ( A α x α = A α x α 1 1 x n αn α N n α N n. for finitely many real symmetric matrices (A α ), α N n.... and suppose one wants to approximate the set R A := {x B : A(x) 0} = {x : λ min (A(x)) 0}. R A = Then: x B : y T A(x)y }{{} f (x,y) 0, y s.t. y 2 = 1

Ex: Polynomial Matrix Inequalities: (with D. Henrion) Let x A(x) R p p where A(x) is the matrix-polynomial ) x A(x) = ( A α x α = A α x α 1 1 x n αn α N n α N n. for finitely many real symmetric matrices (A α ), α N n.... and suppose one wants to approximate the set R A := {x B : A(x) 0} = {x : λ min (A(x)) 0}. R A = Then: x B : y T A(x)y }{{} f (x,y) 0, y s.t. y 2 = 1

Illustrative example (continued) Let B be the unit disk {x : x 1} and let: { ( [ 1 16x1 x R A := x B : A(x) = 2 x 1 x 1 1 x1 2 x 2 2 ]) } 0 Then by solving relatively simple semidefinite programs, one may approximate R A with sublevel sets of the form: Θ k := {x B : J k (x) 0 } for some polynomial J k of degree k = 2, 4,... and with VOL (R A \ Θ k ) 0 as k.

Illustrative example (continued) Let B be the unit disk {x : x 1} and let: { ( [ 1 16x1 x R A := x B : A(x) = 2 x 1 x 1 1 x1 2 x 2 2 ]) } 0 Then by solving relatively simple semidefinite programs, one may approximate R A with sublevel sets of the form: Θ k := {x B : J k (x) 0 } for some polynomial J k of degree k = 2, 4,... and with VOL (R A \ Θ k ) 0 as k.

1 1 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0.2 x 2 0 x 2 0 0.2 0.2 0.4 0.4 0.6 0.6 0.8 0.8 1 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 x 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 x 1 Θ 2 (left) and Θ 4 (right) inner approximations (light gray) of (dark gray) embedded in unit disk B (dashed).

1 1 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0.2 x 2 0 x 2 0 0.2 0.2 0.4 0.4 0.6 0.6 0.8 0.8 1 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 x 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 x 1 Θ 6 (left) and Θ 8 (right) inner approximations (light gray) of (dark gray) embedded in unit disk B (dashed).

Similarly, suppose that one wants to approximate the set: D f := {x B : f (x, y) 0 for some y such that (x, y) K} as closely as desired by a sequence of sets of the form: Θ k := {x B : J k (x) 0 } for some polynomials J k.

Let x F(x) := min {f (x, y) : (x, y) K }, and for every integer k the optimization problem: ρ k = max J R[x] k { } (F J) dx : J(x) f (x, y) Q k (g) B IS AN SDP with an optimal solution J k R[x] k. From the definition of Jk, the sublevel sets Θ k := {x B : J k (x) 0} D f, k N, provide a nested sequence of OUTER approximations of D f.

Theorem (Lass) (Strong) convergence in L 1 (B)-norm takes place, that is: F Jk dx = 0 lim k B and, if in addition the set {x B : F(x) = 0} has Lebesgue measure zero, then lim VOL(Θ k \ D f ) = 0 k

Theorem (Lass) (Strong) convergence in L 1 (B)-norm takes place, that is: F Jk dx = 0 lim k B and, if in addition the set {x B : F(x) = 0} has Lebesgue measure zero, then lim VOL(Θ k \ D f ) = 0 k

III. Convex underestimators of polynomials In large scale Mixed Integer Nonlinear Programming (MINLP), a popular method is to use B & B where LOWER BOUNDS at each node of the search tree must be computed EFFICIENTLY! In such a case... one needs CONVEX UNDERESTIMATORS of the objective function, say on a BOX B R n! Message: Good" CONVEX POLYNOMIAL UNDERESTIMATORS can be computed efficienty!

III. Convex underestimators of polynomials In large scale Mixed Integer Nonlinear Programming (MINLP), a popular method is to use B & B where LOWER BOUNDS at each node of the search tree must be computed EFFICIENTLY! In such a case... one needs CONVEX UNDERESTIMATORS of the objective function, say on a BOX B R n! Message: Good" CONVEX POLYNOMIAL UNDERESTIMATORS can be computed efficienty!

inf p R[x] d Solving { (f (x) p(x)) dx : B s.t. f p 0 on B and p convex on B} will provide a degree-d POLYNOMIAL CONVEX UNDERESTIMATOR p of f on B that minimizes the L 1 (B)-norm f p 1! Notice that: (f (x) p(x)) dx is LINEAR in the coefficients of p! B p convex on B y T 2 p(x) y }{{} R[xy] d 0 on B {y : y 2 = 1}!

Hence replace the positivity and convexity constraints f p 0 on B and p convex on B with the positivity certificates f (x) p(x) = y T 2 p(x) y = m σ j (x) g j (x) }{{} SOS m ψ(x, y) g }{{} j (x) + ψ m+1 (x, y) (1 y 2 ) SOS k=0 k=0

Hence replace the positivity and convexity constraints f p 0 on B and p convex on B with the positivity certificates f (x) p(x) = y T 2 p(x) y = m σ j (x) g j (x) }{{} SOS m ψ(x, y) g }{{} j (x) + ψ m+1 (x, y) (1 y 2 ) SOS k=0 k=0

and apply the moment-sos approach to obtain a sequence of polynomials p k R[x] d, k N, of degree d which converges to the BEST convex polynomial underestimator of degree d.

Conclusion The moment-sos hierarchy is a powerful general methodology. Works for problems of modest size (or larger size problem with sparsity and/or symmetries) f (x) = α,β Mixed LP-SOS positivity certificate c αβ }{{} 0 where k IS FIXED! g j (x) α j j An alternative for larger size problems? j (1 g j (x)) β j + σ 0 (x) }{{} SOS of degree k

Conclusion The moment-sos hierarchy is a powerful general methodology. Works for problems of modest size (or larger size problem with sparsity and/or symmetries) f (x) = α,β Mixed LP-SOS positivity certificate c αβ }{{} 0 where k IS FIXED! g j (x) α j j An alternative for larger size problems? j (1 g j (x)) β j + σ 0 (x) }{{} SOS of degree k

THANK YOU!!