The moment-lp and moment-sos approaches in optimization

Similar documents
The moment-lp and moment-sos approaches

Moments and Positive Polynomials for Optimization II: LP- VERSUS SDP-relaxations

Moments and Positive Polynomials for Optimization II: LP- VERSUS SDP-relaxations

Optimization over Polynomials with Sums of Squares and Moment Matrices

A new look at nonnegativity on closed sets

Convex Optimization & Parsimony of L p-balls representation

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

CONVEXITY IN SEMI-ALGEBRAIC GEOMETRY AND POLYNOMIAL OPTIMIZATION

Towards Solving Bilevel Optimization Problems in Quantum Information Theory

On Polynomial Optimization over Non-compact Semi-algebraic Sets

4. Algebra and Duality

Croatian Operational Research Review (CRORR), Vol. 3, 2012

arxiv: v1 [math.oc] 26 Sep 2015

Solving Global Optimization Problems with Sparse Polynomials and Unbounded Semialgebraic Feasible Sets

Global Optimization with Polynomials

Strong duality in Lasserre s hierarchy for polynomial optimization

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

MATHEMATICAL ENGINEERING TECHNICAL REPORTS. Exact SDP Relaxations with Truncated Moment Matrix for Binary Polynomial Optimization Problems

Degeneracy in Maximal Clique Decomposition for Semidefinite Programs

Convergence rates of moment-sum-of-squares hierarchies for volume approximation of semialgebraic sets

Hilbert s 17th Problem to Semidefinite Programming & Convex Algebraic Geometry

Research Reports on Mathematical and Computing Sciences

An explicit construction of distinguished representations of polynomials nonnegative over finite sets

Minimum Ellipsoid Bounds for Solutions of Polynomial Systems via Sum of Squares

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

Introduction to Semidefinite Programming I: Basic properties a

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Optimization over Nonnegative Polynomials: Algorithms and Applications. Amir Ali Ahmadi Princeton, ORFE

SEMIDEFINITE PROGRAMMING VS. LP RELAXATIONS FOR POLYNOMIAL PROGRAMMING

A PARALLEL INTERIOR POINT DECOMPOSITION ALGORITHM FOR BLOCK-ANGULAR SEMIDEFINITE PROGRAMS

Semidefinite Programming, Combinatorial Optimization and Real Algebraic Geometry

Robust and Optimal Control, Spring 2015

arxiv: v1 [math.oc] 31 Jan 2017

Representations of Positive Polynomials: Theory, Practice, and

How to generate weakly infeasible semidefinite programs via Lasserre s relaxations for polynomial optimization

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

A PARALLEL INTERIOR POINT DECOMPOSITION ALGORITHM FOR BLOCK-ANGULAR SEMIDEFINITE PROGRAMS IN POLYNOMIAL OPTIMIZATION

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones

Detecting global optimality and extracting solutions in GloptiPoly

The maximal stable set problem : Copositive programming and Semidefinite Relaxations

Semidefinite Programming Basics and Applications

BBCPOP: A Sparse Doubly Nonnegative Relaxation of Polynomial Optimization Problems with Binary, Box and Complementarity Constraints

A CONIC DANTZIG-WOLFE DECOMPOSITION APPROACH FOR LARGE SCALE SEMIDEFINITE PROGRAMMING

Global Optimization with Polynomials

III. Applications in convex optimization

Convexification of Mixed-Integer Quadratically Constrained Quadratic Programs

approximation algorithms I

Lecture 6: Conic Optimization September 8

Semidefinite programming and convex algebraic geometry

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

that a broad class of conic convex polynomial optimization problems, called

A new approximation hierarchy for polynomial conic optimization

Semidefinite programming lifts and sparse sums-of-squares

Copositive Programming and Combinatorial Optimization

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Semialgebraic Relaxations using Moment-SOS Hierarchies

Modeling with semidefinite and copositive matrices

Analysis and synthesis: a complexity perspective

DSOS/SDOS Programming: New Tools for Optimization over Nonnegative Polynomials

Sparsity in Sums of Squares of Polynomials

Relations between Semidefinite, Copositive, Semi-infinite and Integer Programming

Semidefinite Programming

c 2000 Society for Industrial and Applied Mathematics

m i=1 c ix i i=1 F ix i F 0, X O.

Lecture 3: Semidefinite Programming

Strange Behaviors of Interior-point Methods. for Solving Semidefinite Programming Problems. in Polynomial Optimization

Dimension reduction for semidefinite programming

Semidefinite approximations of projections and polynomial images of semialgebraic sets

Approximate Optimal Designs for Multivariate Polynomial Regression

Lift-and-Project Techniques and SDP Hierarchies

EE 227A: Convex Optimization and Applications October 14, 2008

Optimality Conditions for Constrained Optimization

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

Research Reports on Mathematical and Computing Sciences

Global Optimization of Polynomials

Research Reports on Mathematical and Computing Sciences

Non-commutative polynomial optimization

Sparse Matrix Theory and Semidefinite Optimization

An Introduction to Polynomial and Semi-Algebraic Optimization

15. Conic optimization

Optimality, Duality, Complementarity for Constrained Optimization

SDP Relaxations for MAXCUT

Real solving polynomial equations with semidefinite programming

A PARALLEL TWO-STAGE INTERIOR POINT DECOMPOSITION ALGORITHM FOR LARGE SCALE STRUCTURED SEMIDEFINITE PROGRAMS

The convex algebraic geometry of rank minimization

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

Lagrangian-Conic Relaxations, Part I: A Unified Framework and Its Applications to Quadratic Optimization Problems

A ten page introduction to conic optimization

Convex Optimization in Classification Problems

Copositive Programming and Combinatorial Optimization

Certified Roundoff Error Bounds using Semidefinite Programming

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

Lower bounds on the size of semidefinite relaxations. David Steurer Cornell

SPECTRAHEDRA. Bernd Sturmfels UC Berkeley

The Geometry of Semidefinite Programming. Bernd Sturmfels UC Berkeley

Advanced SDPs Lecture 6: March 16, 2017

Detecting global optimality and extracting solutions in GloptiPoly

Transcription:

The moment-lp and moment-sos approaches in optimization LAAS-CNRS and Institute of Mathematics, Toulouse, France Workshop Linear Matrix Inequalities, Semidefinite Programming and Quantum Information Theory 2016, Toulouse, January 2016

978-0-387-09685-8.tif 827 1 355 pixels http://images.springer.com/covers/978-0-387-09685-8.tif 13/09/12 10:09 Page 1 sur 1 See in particular the Chapter by M. Laurent

M.F. Anjos, Polytechnique Montréal, QC, Canada; J.B. Lasserre, LAAS, Toulouse Cedex 4, France (Eds.) Handbook on Semidefinite, Conic and Polynomial Optimization Series: International Series in Operations Research & Management Science Summarizes research and developments of last ten years and brings field up to date Individual sections covering theory, algorithms, software and applications Editors are quite prominent in the field. 2012, 2012, XI, 957 p. 57 illus. Printed book Hardcover 199,95 180.00 $279.00 *213,95 (D) 219,94 (A) CHF 287.00 ebook For individual purchases buy at a lower price on springer.com. A free preview is available on SpringerLink. Also available from libraries offering Springer s ebook Collection. springer.com/ebooks MyCopy Printed ebook exclusively available to patrons whose library offers Springer s ebook Collection.*** $ 24.95 springer.com/mycopy Semidefinite and conic optimization is a major and thriving research area within the optimization community. Although semidefinite optimization has been studied (under different names) since at least the 1940s, its importance grew immensely during the 1990s after polynomial-time interior-point methods for linear optimization were extended to solve semidefinite optimization problems. Since the beginning of the 21st century, not only has research into semidefinite and conic optimization continued unabated, but also a fruitful interaction has developed with algebraic geometry through the close connections between semidefinite matrices and polynomial optimization. This has brought about important new results and led to an even higher level of research activity. This Handbook on Semidefinite, Conic and Polynomial Optimization provides the reader with a snapshot of the state-of-the-art in the growing and mutually enriching areas of semidefinite optimization, conic optimization, and polynomial optimization. It contains a compendium of the recent research activity that has taken place in these thrilling areas, and will appeal to doctoral students, young graduates, and experienced researchers alike. The Handbook s thirty-one chapters are organized into four parts:theory, covering significant theoretical developments as well as the interactions between conic optimization and polynomial optimization;algorithms, documenting the directions of current algorithmic development;software, providing an overview of the state-ofthe-art;applications, dealing with the application areas where semidefinite and conic optimization has made a significant impact in recent years.

Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches

Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches

Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches

Consider the polynomial optimization problem: P : f = min{ f (x) : g j (x) 0, j = 1,..., m } for some polynomials f, g j R[x]. Why Polynomial Optimization? After all... P is just a particular case of Non Linear Programming (NLP)!

Consider the polynomial optimization problem: P : f = min{ f (x) : g j (x) 0, j = 1,..., m } for some polynomials f, g j R[x]. Why Polynomial Optimization? After all... P is just a particular case of Non Linear Programming (NLP)!

True!... if one is interested with a LOCAL optimum only!! When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!

True!... if one is interested with a LOCAL optimum only!! When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!

True!... if one is interested with a LOCAL optimum only!! When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!

True!... if one is interested with a LOCAL optimum only!! When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!

BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}. (Not true for a global minimum!)) and so to compute f... one needs to handle EFFICIENTLY the difficult constraint f (x) λ 0 x K, i.e. one needs TRACTABLE CERTIFICATES of POSITIVITY on K for the polynomial x f (x) λ!

BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}. (Not true for a global minimum!)) and so to compute f... one needs to handle EFFICIENTLY the difficult constraint f (x) λ 0 x K, i.e. one needs TRACTABLE CERTIFICATES of POSITIVITY on K for the polynomial x f (x) λ!

BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}. (Not true for a global minimum!)) and so to compute f... one needs to handle EFFICIENTLY the difficult constraint f (x) λ 0 x K, i.e. one needs TRACTABLE CERTIFICATES of POSITIVITY on K for the polynomial x f (x) λ!

REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)

REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)

REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)

SOS-based certificate K = { x : g j (x) 0, j = 1,..., m } Theorem (Putinar s Positivstellensatz) If K is compact (+ a technical Archimedean assumption) and f > 0 on K then: f (x) = σ 0 (x) + m σ j (x) g j (x), x R n, j=1 for some SOS polynomials (σ j ) R[x].

However... In Putinar s theorem... nothing is said on the DEGREE of the SOS polynomials (σ j )! BUT... GOOD news..!! Testing whether holds for some SOS (σ j ) R[x] with a degree bound, is SOLVING an SDP!

However... In Putinar s theorem... nothing is said on the DEGREE of the SOS polynomials (σ j )! BUT... GOOD news..!! Testing whether holds for some SOS (σ j ) R[x] with a degree bound, is SOLVING an SDP!

Semidefinite Programming The CONVEX optimization problem: P min x R n { c x n A i x i i=1 b}, where c R n and b, A i S m (m m symmetric matrices), is called a semidefinite program. The notation 0" means the real symmetric matrix " is positive semidefinite, i.e., all its (real) EIGENVALUES are nonnegative.

Example P : min {x 1 + x 2 : x [ 3 + 2x1 + x s.t. 2 x 1 5 x 1 5 x 1 2x 2 or, equivalently ] }, 0 P : min {x 1 + x 2 : x [ 3 5 s.t. 5 0 ] + x 1 [ 2 1 1 1 ] + x 2 [ 1 0 0 2 ] } 0

P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...

P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...

P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...

Dual side of Putinar s theorem: The K -moment problem Given a real sequence y = (y α ), α N n, does there exist a Borel measure µ on K such that y α = x α 1 1 x n αn dµ, α N n. K Introduce the so-called Riesz linear functional L y : R[x] R : ( f = ) f α x α L y (f ) = f α y α α α N n

Dual side of Putinar s theorem: The K -moment problem Given a real sequence y = (y α ), α N n, does there exist a Borel measure µ on K such that y α = x α 1 1 x n αn dµ, α N n. K Introduce the so-called Riesz linear functional L y : R[x] R : ( f = ) f α x α L y (f ) = f α y α α α N n

Theorem If K = {x : g j (x) 0, j = 1,..., m} is compact and satisfies an Archimedean assumption then holds if and only if for every h R[x] 2 : ( ) L y (h 2 ) 0; L y (h 2 g j ) 0, j = 1,..., m. The condition ( ) is equivalent to m + 1 positive semidefiniteness of some moment and localizing matrices, i.e., M(y) 0; M(g j y) 0, j = 1,..., m. whose rows & columns are indexed by N n, and entries are LINEAR in the y α s

Theorem If K = {x : g j (x) 0, j = 1,..., m} is compact and satisfies an Archimedean assumption then holds if and only if for every h R[x] 2 : ( ) L y (h 2 ) 0; L y (h 2 g j ) 0, j = 1,..., m. The condition ( ) is equivalent to m + 1 positive semidefiniteness of some moment and localizing matrices, i.e., M(y) 0; M(g j y) 0, j = 1,..., m. whose rows & columns are indexed by N n, and entries are LINEAR in the y α s

LP-based certificate K = {x : g j (x) 0; (1 g j (x)) 0, j = 1,..., m} Theorem (Krivine-Vasilescu-Handelman s Positivstellensatz) Let K be compact and the family {1, g j } generate R[x]. If f > 0 on K then: ( ) f (x) = α,β c αβ m j=1 g j (x) α j (1 g j (x)) β j,, x R n, for some NONNEGATIVE scalars (c αβ ).

However... Again in Krivine s theorem In ( )... nothing is said on how many nonnegative scalars c αβ are needed! BUT... GOOD news... again!! Testing whether ( ) holds for some nonnegative scalars (c αβ ) is SOLVING an LP!

However... Again in Krivine s theorem In ( )... nothing is said on how many nonnegative scalars c αβ are needed! BUT... GOOD news... again!! Testing whether ( ) holds for some nonnegative scalars (c αβ ) is SOLVING an LP!

Dual side of Krivine s theorem: The K -moment problem Theorem If K = {x : g j (x) 0, j = 1,..., m} is compact, 0 g j 1 on K, and {1, g j } generates R[x], then holds if and only if ( ) L y m α g j β j (1 g j j 0, α, β N m. j=1 The condition ( ) is equivalent to countably many LINEAR INEQUALITIES on the y α s

Dual side of Krivine s theorem: The K -moment problem Theorem If K = {x : g j (x) 0, j = 1,..., m} is compact, 0 g j 1 on K, and {1, g j } generates R[x], then holds if and only if ( ) L y m α g j β j (1 g j j 0, α, β N m. j=1 The condition ( ) is equivalent to countably many LINEAR INEQUALITIES on the y α s

HENCE... SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY!

HENCE... SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY!

In addition, polynomials NONNEGATIVE ON A SET K R n are ubiquitous. They also appear in many important applications (outside optimization),... modeled as particular instances of the so called Generalized Moment Problem, among which: Probability, Optimal and Robust Control, Game theory, Signal processing, multivariate integration, etc. (GMP) : s inf { f i dµ i : µ i M(K i ) i=1 K i s i=1 K i h ij dµ i = b j, j J} with M(K i ) space of Borel measures on K i R n i, i = 1,..., s.

In addition, polynomials NONNEGATIVE ON A SET K R n are ubiquitous. They also appear in many important applications (outside optimization),... modeled as particular instances of the so called Generalized Moment Problem, among which: Probability, Optimal and Robust Control, Game theory, Signal processing, multivariate integration, etc. (GMP) : s inf { f i dµ i : µ i M(K i ) i=1 K i s i=1 K i h ij dµ i = b j, j J} with M(K i ) space of Borel measures on K i R n i, i = 1,..., s.

The DUAL of the GMP is the linear program GMP : sup λ j { s λ j b j : j J f i j J λ j h ij 0 on K i, i = 1,..., s } And one can see that... the constraints of GMP state that some functions f i j J λ j h ij must be nonnegative on a certain set K i, i = 1,..., s.

A couple of examples I: Global OPTIM f = inf x { f (x) : x K } is the SIMPLEST example of the GMP because... f = inf { f dµ : µ M(K) K K 1 dµ = 1} Indeed if f (x) f for all x K and µ is a probability measure on K, then K f dµ f dµ = f. On the other hand, for every x K the probability measure µ := δ x is such that f dµ = f (x).

A couple of examples I: Global OPTIM f = inf x { f (x) : x K } is the SIMPLEST example of the GMP because... f = inf { f dµ : µ M(K) K K 1 dµ = 1} Indeed if f (x) f for all x K and µ is a probability measure on K, then K f dµ f dµ = f. On the other hand, for every x K the probability measure µ := δ x is such that f dµ = f (x).

A couple of examples I: Global OPTIM f = inf x { f (x) : x K } is the SIMPLEST example of the GMP because... f = inf { f dµ : µ M(K) K K 1 dµ = 1} Indeed if f (x) f for all x K and µ is a probability measure on K, then K f dµ f dµ = f. On the other hand, for every x K the probability measure µ := δ x is such that f dµ = f (x).

A couple of examples I: Global OPTIM f = inf x { f (x) : x K } is the SIMPLEST example of the GMP because... f = inf { f dµ : µ M(K) K K 1 dµ = 1} Indeed if f (x) f for all x K and µ is a probability measure on K, then K f dµ f dµ = f. On the other hand, for every x K the probability measure µ := δ x is such that f dµ = f (x).

II. Let K R n and S K be given, and let Γ N n be also given. BOUNDS on measures with moment conditions max { 1 S, µ : x α dµ = m α, α Γ } µ M(K) to compute an upper bound on µ(s) over all distributions µ M(K) with a certain fixed number of moments m α. K If Γ = N n then one may use this to compute the Lebesgue volume of a compact basic semi-algebraic set S K := [ 1, 1] n. Take m α := x α dx, α N n. [ 1,1] n

II. Let K R n and S K be given, and let Γ N n be also given. BOUNDS on measures with moment conditions max { 1 S, µ : x α dµ = m α, α Γ } µ M(K) to compute an upper bound on µ(s) over all distributions µ M(K) with a certain fixed number of moments m α. K If Γ = N n then one may use this to compute the Lebesgue volume of a compact basic semi-algebraic set S K := [ 1, 1] n. Take m α := x α dx, α N n. [ 1,1] n

III. For instance, one may also want: To approximate sets defined with QUANTIFIERS, like.e.g., R f := {x B : f (x, y) 0 for all y such that (x, y) K} D f := {x B : f (x, y) 0 for some y such that (x, y) K} where f R[x, y], B is a simple set (box, ellipsoid). To compute convex polynomial underestimators p f of a polynomial f on a box B R n. (Very useful in MINLP.)

III. For instance, one may also want: To approximate sets defined with QUANTIFIERS, like.e.g., R f := {x B : f (x, y) 0 for all y such that (x, y) K} D f := {x B : f (x, y) 0 for some y such that (x, y) K} where f R[x, y], B is a simple set (box, ellipsoid). To compute convex polynomial underestimators p f of a polynomial f on a box B R n. (Very useful in MINLP.)

The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Vasilescu-Handelman s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In many situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.

The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Vasilescu-Handelman s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In many situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.

The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Vasilescu-Handelman s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In many situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.

The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Vasilescu-Handelman s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In many situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.

LP- and SDP-hierarchies for optimization Replace f = sup λ,σj { λ : f (x) λ 0 x K} with: The SDP-hierarchy indexed by d N: m fd = sup { λ : f λ = σ 0 + σ }{{} j g j ; deg (σ j g j ) 2d } }{{} SOS j=1 SOS or, the LP-hierarchy indexed by d N: θ d = sup { λ : f λ = α,β c αβ }{{} 0 m α g j j (1 g j ) β j ; α + β 2d} j=1

LP- and SDP-hierarchies for optimization Replace f = sup λ,σj { λ : f (x) λ 0 x K} with: The SDP-hierarchy indexed by d N: m fd = sup { λ : f λ = σ 0 + σ }{{} j g j ; deg (σ j g j ) 2d } }{{} SOS j=1 SOS or, the LP-hierarchy indexed by d N: θ d = sup { λ : f λ = α,β c αβ }{{} 0 m α g j j (1 g j ) β j ; α + β 2d} j=1

Theorem Both sequence (f d ), and (θ d), d N, are MONOTONE NON DECREASING and when K is compact (and satisfies a technical Archimedean assumption) then: f = lim fd = lim θ d. d d

What makes this approach exciting is that it is at the crossroads of several disciplines/applications: Commutative, Non-commutative, and Non-linear ALGEBRA Real algebraic geometry, and Functional Analysis Optimization, Convex Analysis Computational Complexity in Computer Science, which BENEFIT from interactions! As mentioned... potential applications are ENDLESS!

What makes this approach exciting is that it is at the crossroads of several disciplines/applications: Commutative, Non-commutative, and Non-linear ALGEBRA Real algebraic geometry, and Functional Analysis Optimization, Convex Analysis Computational Complexity in Computer Science, which BENEFIT from interactions! As mentioned... potential applications are ENDLESS!

Has already been proved useful and successful in applications with modest problem size, notably in optimization, control, robust control, optimal control, estimation, computer vision, etc. (If sparsity then problems of larger size can be addressed) HAS initiated and stimulated new research issues: in Convex Algebraic Geometry (e.g. semidefinite representation of convex sets, algebraic degree of semidefinite programming and polynomial optimization) in Computational algebra (e.g., for solving polynomial equations via SDP and Border bases) Computational Complexity where LP- and SDP-HIERARCHIES have become an important tool to analyze Hardness of Approximation for 0/1 combinatorial problems ( links with quantum computing)

Has already been proved useful and successful in applications with modest problem size, notably in optimization, control, robust control, optimal control, estimation, computer vision, etc. (If sparsity then problems of larger size can be addressed) HAS initiated and stimulated new research issues: in Convex Algebraic Geometry (e.g. semidefinite representation of convex sets, algebraic degree of semidefinite programming and polynomial optimization) in Computational algebra (e.g., for solving polynomial equations via SDP and Border bases) Computational Complexity where LP- and SDP-HIERARCHIES have become an important tool to analyze Hardness of Approximation for 0/1 combinatorial problems ( links with quantum computing)

Recall that both LP- and SDP- hierarchies are GENERAL PURPOSE METHODS... NOT TAILORED to solving specific hard problems!!

Recall that both LP- and SDP- hierarchies are GENERAL PURPOSE METHODS... NOT TAILORED to solving specific hard problems!!

A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.

A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.

A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.

Even though the moment-sos approach DOES NOT SPECIALIZE to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. Finite convergence also occurs for general convex problems and generically for non convex problems (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems! The Computer Science community talks about a META-Algorithm.

Even though the moment-sos approach DOES NOT SPECIALIZE to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. Finite convergence also occurs for general convex problems and generically for non convex problems (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems! The Computer Science community talks about a META-Algorithm.

Even though the moment-sos approach DOES NOT SPECIALIZE to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. Finite convergence also occurs for general convex problems and generically for non convex problems (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems! The Computer Science community talks about a META-Algorithm.

Even though the moment-sos approach DOES NOT SPECIALIZE to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. Finite convergence also occurs for general convex problems and generically for non convex problems (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems! The Computer Science community talks about a META-Algorithm.

Even though the moment-sos approach DOES NOT SPECIALIZE to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. Finite convergence also occurs for general convex problems and generically for non convex problems (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems! The Computer Science community talks about a META-Algorithm.

A remarkable property: II FINITE CONVERGENCE of the SOS-hierarchy is GENERIC!... and provides a GLOBAL OPTIMALITY CERTIFICATE, the analogue for the NON CONVEX CASE of the KKT-OPTIMALITY conditions in the CONVEX CASE!

Theorem (Marshall, Nie) Let x K be a global minimizer of P : f = min {f (x) : g j (x) 0, j = 1,..., m}. and assume that: (i) The gradients { g j (x )} are linearly independent, (ii) Strict complementarity holds (λ j g j (x ) = 0 for all j.) (iii) Second-order sufficiency conditions hold at (x, λ ) K R m +. m Then f (x) f = σ0 (x) + σj (x)g j(x), SOS polynomials {σ j }. j=1 x R n, for some Moreover, the conditions (i)-(ii)-(iii) HOLD GENERICALLY!

Theorem (Marshall, Nie) Let x K be a global minimizer of P : f = min {f (x) : g j (x) 0, j = 1,..., m}. and assume that: (i) The gradients { g j (x )} are linearly independent, (ii) Strict complementarity holds (λ j g j (x ) = 0 for all j.) (iii) Second-order sufficiency conditions hold at (x, λ ) K R m +. m Then f (x) f = σ0 (x) + σj (x)g j(x), SOS polynomials {σ j }. j=1 x R n, for some Moreover, the conditions (i)-(ii)-(iii) HOLD GENERICALLY!

Certificates of positivity already exist in convex optimization f = f (x ) = min { f (x) : g j (x) 0, j = 1,..., m } when f and g j are CONVEX. Indeed if Slater s condition holds there exist nonnegative KKT-multipliers λ j R m + such that: f (x ) m λ j g j (x ) = 0; λ j g j (x ) = 0, j = 1,..., m. j=1... and so... the Lagrangian L λ (x) := f (x) f j=1 λ j g j (x), satisfies L λ (x ) = 0 and L λ (x) 0 for all x. Therefore: L λ (x) 0 f (x) f x K!

Certificates of positivity already exist in convex optimization f = f (x ) = min { f (x) : g j (x) 0, j = 1,..., m } when f and g j are CONVEX. Indeed if Slater s condition holds there exist nonnegative KKT-multipliers λ j R m + such that: f (x ) m λ j g j (x ) = 0; λ j g j (x ) = 0, j = 1,..., m. j=1... and so... the Lagrangian L λ (x) := f (x) f j=1 λ j g j (x), satisfies L λ (x ) = 0 and L λ (x) 0 for all x. Therefore: L λ (x) 0 f (x) f x K!

In summary: KKT-OPTIMALITY when f and g j are CONVEX m f (x ) λ j g j(x ) = 0 f (x) f j=1 m λ j g j(x) j=1 PUTINAR s CERTIFICATE in the non CONVEX CASE m f (x ) σ j (x ) g j (x ) = 0 f (x) f j=1 m σj (x)g j(x) j=1 0 for all x R n (= σ0 (x)) 0 for all x Rn. for some SOS {σj }, and σj (x ) = λ j.

In summary: KKT-OPTIMALITY when f and g j are CONVEX m f (x ) λ j g j(x ) = 0 f (x) f j=1 m λ j g j(x) j=1 PUTINAR s CERTIFICATE in the non CONVEX CASE m f (x ) σ j (x ) g j (x ) = 0 f (x) f j=1 m σj (x)g j(x) j=1 0 for all x R n (= σ0 (x)) 0 for all x Rn. for some SOS {σj }, and σj (x ) = λ j.

The no-free lunch rule... The size of SDP-relaxations grows rapidly with the original problem size... In particular: O(n 2d ) variables for the d th SDP-relaxation in the hierarchy O(n d ) matrix size for the LMIs In view of the present status of SDP-solvers... only small to medium size problems can be solved by "standard" SDP-relaxations...... How to handle larger size problems?

The no-free lunch rule... The size of SDP-relaxations grows rapidly with the original problem size... In particular: O(n 2d ) variables for the d th SDP-relaxation in the hierarchy O(n d ) matrix size for the LMIs In view of the present status of SDP-solvers... only small to medium size problems can be solved by "standard" SDP-relaxations...... How to handle larger size problems?

The no-free lunch rule... The size of SDP-relaxations grows rapidly with the original problem size... In particular: O(n 2d ) variables for the d th SDP-relaxation in the hierarchy O(n d ) matrix size for the LMIs In view of the present status of SDP-solvers... only small to medium size problems can be solved by "standard" SDP-relaxations...... How to handle larger size problems?

The no-free lunch rule... The size of SDP-relaxations grows rapidly with the original problem size... In particular: O(n 2d ) variables for the d th SDP-relaxation in the hierarchy O(n d ) matrix size for the LMIs In view of the present status of SDP-solvers... only small to medium size problems can be solved by "standard" SDP-relaxations...... How to handle larger size problems?

The no-free lunch rule... The size of SDP-relaxations grows rapidly with the original problem size... In particular: O(n 2d ) variables for the d th SDP-relaxation in the hierarchy O(n d ) matrix size for the LMIs In view of the present status of SDP-solvers... only small to medium size problems can be solved by "standard" SDP-relaxations...... How to handle larger size problems?

develop more efficient general purpose SDP-solvers... (limited impact)... or perhaps dedicated solvers...? exploit symmetries when present... Recent promising works by De Klerk, Gaterman, Gvozdenovic, Laurent, Pasechnick, Parrilo, Schrijver.. in particular for combinatorial optimization problems. Algebraic techniques permit to define an equivalent SDP of much smaller size. exploit sparsity in the data. In general, each constraint involves a small number of variables κ, and the cost criterion is a sum of polynomials involving also a small number of variables. Recent works by Kim, Kojima, Lasserre, Maramatsu and Waki

develop more efficient general purpose SDP-solvers... (limited impact)... or perhaps dedicated solvers...? exploit symmetries when present... Recent promising works by De Klerk, Gaterman, Gvozdenovic, Laurent, Pasechnick, Parrilo, Schrijver.. in particular for combinatorial optimization problems. Algebraic techniques permit to define an equivalent SDP of much smaller size. exploit sparsity in the data. In general, each constraint involves a small number of variables κ, and the cost criterion is a sum of polynomials involving also a small number of variables. Recent works by Kim, Kojima, Lasserre, Maramatsu and Waki

develop more efficient general purpose SDP-solvers... (limited impact)... or perhaps dedicated solvers...? exploit symmetries when present... Recent promising works by De Klerk, Gaterman, Gvozdenovic, Laurent, Pasechnick, Parrilo, Schrijver.. in particular for combinatorial optimization problems. Algebraic techniques permit to define an equivalent SDP of much smaller size. exploit sparsity in the data. In general, each constraint involves a small number of variables κ, and the cost criterion is a sum of polynomials involving also a small number of variables. Recent works by Kim, Kojima, Lasserre, Maramatsu and Waki

In recent works, Koijma s group has developed a systematic procedure to discover sparsity pattern p p I = {1, 2,..., n} = I j. J = {1, 2,..., m} = J k. j=1 k=1 Essentially one looks for maximal cliques {I j } in some chordal graph extension of a graph associated with the problem. P : f = min {f (x) : g j (x) 0, j = 1,..., m}.

Assumption I: Sparsity pattern f = p k=1 f k with f k R[X i ; i I k ]. For all j J k, g j R[X i ; i I k ]. Assumption II: The Running Intersection Property For every k = 2,... p k 1 I k I s for some s k 1, j=1 I j

Assumption I: Sparsity pattern f = p k=1 f k with f k R[X i ; i I k ]. For all j J k, g j R[X i ; i I k ]. Assumption II: The Running Intersection Property For every k = 2,... p k 1 I k I s for some s k 1, j=1 I j

Theorem (Lass (Sparse version of Putinar s Positivstellensatz)) Let Assumption I and II hold and an archimedean assumption holds for K. If f > 0 on K then: p f = σ 0k + σ jk g j, j Jk k=1 for some s.o.s. polynomials σ jk R[X i : i I k ]. Examples with n large (say with n = 1000) and small clique size κ (e.g. κ = 3, 4,..7) are solved relatively easily with Kojima s group software SparsePOP

Theorem (Lass (Sparse version of Putinar s Positivstellensatz)) Let Assumption I and II hold and an archimedean assumption holds for K. If f > 0 on K then: p f = σ 0k + σ jk g j, j Jk k=1 for some s.o.s. polynomials σ jk R[X i : i I k ]. Examples with n large (say with n = 1000) and small clique size κ (e.g. κ = 3, 4,..7) are solved relatively easily with Kojima s group software SparsePOP

There has been also recents attempts to use other types of algebraic certificates of positivity that try to avoid the size explosion due to the semidefinite matrices associated with the SOS weights in Putinar s positivity certificate Recent work by Ahmadi et al. and Lasserre, Toh and Zhang