Moments and Positive Polynomials for Optimization II: LP- VERSUS SDP-relaxations

Similar documents
Moments and Positive Polynomials for Optimization II: LP- VERSUS SDP-relaxations

A new look at nonnegativity on closed sets

The moment-lp and moment-sos approaches

The moment-lp and moment-sos approaches in optimization

SEMIDEFINITE PROGRAMMING VS. LP RELAXATIONS FOR POLYNOMIAL PROGRAMMING

CONVEXITY IN SEMI-ALGEBRAIC GEOMETRY AND POLYNOMIAL OPTIMIZATION

On Polynomial Optimization over Non-compact Semi-algebraic Sets

arxiv: v2 [math.oc] 31 May 2010

Hilbert s 17th Problem to Semidefinite Programming & Convex Algebraic Geometry

Matricial real algebraic geometry

Convex Optimization & Parsimony of L p-balls representation

Convergence rates of moment-sum-of-squares hierarchies for volume approximation of semialgebraic sets

Approximate Optimal Designs for Multivariate Polynomial Regression

c 2008 Society for Industrial and Applied Mathematics

Global Optimization of Polynomials

APPROXIMATE VOLUME AND INTEGRATION FOR BASIC SEMI-ALGEBRAIC SETS

Towards Solving Bilevel Optimization Problems in Quantum Information Theory

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

Global Optimization with Polynomials

Optimization over Polynomials with Sums of Squares and Moment Matrices

Convex computation of the region of attraction for polynomial control systems

Strong duality in Lasserre s hierarchy for polynomial optimization

Notes on the decomposition result of Karlin et al. [2] for the hierarchy of Lasserre by M. Laurent, December 13, 2012

arxiv:math/ v1 [math.oc] 13 Mar 2007

Convex computation of the region of attraction for polynomial control systems

A brief history of noncommutative Positivstellensätze. Jaka Cimprič, University of Ljubljana, Slovenia

Cutting Planes for First Level RLT Relaxations of Mixed 0-1 Programs

How to generate weakly infeasible semidefinite programs via Lasserre s relaxations for polynomial optimization

A new approximation hierarchy for polynomial conic optimization

Advanced SDPs Lecture 6: March 16, 2017

Mean squared error minimization for inverse moment problems

Region of attraction approximations for polynomial dynamical systems

Linear conic optimization for nonlinear optimal control

4. Algebra and Duality

Semidefinite Programming Basics and Applications

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Sum of Squares Relaxations for Polynomial Semi-definite Programming

arxiv: v1 [math.oc] 31 Jan 2017

Minimum Ellipsoid Bounds for Solutions of Polynomial Systems via Sum of Squares

Measures and LMIs for optimal control of piecewise-affine systems

LP Infeasibility - The Pursuit of Intelligent Discretization

2.098/6.255/ Optimization Methods Practice True/False Questions

Lecture 3: Semidefinite Programming

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

A PARALLEL INTERIOR POINT DECOMPOSITION ALGORITHM FOR BLOCK-ANGULAR SEMIDEFINITE PROGRAMS IN POLYNOMIAL OPTIMIZATION

Integer programming, Barvinok s counting algorithm and Gomory relaxations

Linear and conic programming relaxations: Graph structure and message-passing

Representations of Positive Polynomials: Theory, Practice, and

An Introduction to Polynomial and Semi-Algebraic Optimization

SEMIDEFINITE PROGRAMMING FOR MIN-MAX PROBLEMS AND GAMES

Mean squared error minimization for inverse moment problems

Global Minimization of Rational Functions and the Nearest GCDs

Semialgebraic Relaxations using Moment-SOS Hierarchies

Semi-definite representibility. For fun and profit

An explicit construction of distinguished representations of polynomials nonnegative over finite sets

Robust and Optimal Control, Spring 2015

Solving Global Optimization Problems with Sparse Polynomials and Unbounded Semialgebraic Feasible Sets

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

Computing uniform convex approximations for convex envelopes and convex hulls

An Exact Jacobian SDP Relaxation for Polynomial Optimization

Input: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function

A JOINT+MARGINAL APPROACH TO PARAMETRIC POLYNOMIAL OPTIMIZATION

Global Optimization with Polynomials

Rank-one LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the well-known Lyapunov's matrix inequality about

Convex computation of the region of attraction of polynomial control systems

Determinant maximization with linear. S. Boyd, L. Vandenberghe, S.-P. Wu. Information Systems Laboratory. Stanford University

Approximating Pareto Curves using Semidefinite Relaxations

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Introduction to LP and SDP Hierarchies

Analysis and synthesis: a complexity perspective

Sherali-Adams Relaxations of Graph Isomorphism Polytopes. Peter N. Malkin* and Mohammed Omar + UC Davis

Lecture 8. Strong Duality Results. September 22, 2008

Introduction to Semidefinite Programming I: Basic properties a

A PARALLEL INTERIOR POINT DECOMPOSITION ALGORITHM FOR BLOCK-ANGULAR SEMIDEFINITE PROGRAMS

A Review of Linear Programming

Semidefinite approximations of projections and polynomial images of semialgebraic sets

Modal occupation measures and LMI relaxations for nonlinear switched systems control

New Developments on OPF Problems. Daniel Bienstock, Columbia University. April 2016

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

A GENERALIZATION OF THE LÖWNER-JOHN S ELLIPSOID THEOREM

1 Overview. 2 Extreme Points. AM 221: Advanced Optimization Spring 2016

The Trust Region Subproblem with Non-Intersecting Linear Constraints

3.7 Cutting plane methods

Robust Stability. Robust stability against time-invariant and time-varying uncertainties. Parameter dependent Lyapunov functions

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

Lift-and-Project Techniques and SDP Hierarchies

BBM402-Lecture 20: LP Duality

Network Utility Maximization With Nonconcave Utilities Using Sum-of-Squares Method

A Note on Representations of Linear Inequalities in Non-Convex Mixed-Integer Quadratic Programs

Using real algebraic geometry to solve combinatorial problems with symmetries

Moments and convex optimization for analysis and control of nonlinear partial differential equations

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

Real solving polynomial equations with semidefinite programming

Lecture 5. Max-cut, Expansion and Grothendieck s Inequality

9. Interpretations, Lifting, SOS and Moments

Detecting global optimality and extracting solutions in GloptiPoly

16.1 L.P. Duality Applied to the Minimax Theorem

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

Transcription:

Moments and Positive Polynomials for Optimization II: LP- VERSUS SDP-relaxations LAAS-CNRS and Institute of Mathematics, Toulouse, France EECI Course: February 2016

LP-relaxations LP- VERSUS SDP-relaxations

LP-relaxations LP- VERSUS SDP-relaxations

LP-relaxations LP- VERSUS SDP-relaxations

Recall the Global Optimization problem P: f := min{ f (x) g j (x) 0, j = 1,..., m}, where f and g j are all POLYNOMIALS, and let K := { x R n g j (x) 0, j = 1,..., m} be the feasible set (a compact basic semi-algebraic set)

Putinar Positivstellensatz Assumption 1: For some M > 0, the quadratic polynomial M X 2 belongs to the quadratic module Q(g 1,..., g m ) Theorem (Putinar-Jacobi-Prestel) Let K be compact and Assumption 1 hold. Then [ f R[X] and f > 0 on K ] f Q(g 1,..., g m ), i.e., m f (x) = σ 0 (x) + σ j (x) g j (x), j=1 for some s.o.s. polynomials {σ j } m j=0. x R n

Putinar Positivstellensatz Assumption 1: For some M > 0, the quadratic polynomial M X 2 belongs to the quadratic module Q(g 1,..., g m ) Theorem (Putinar-Jacobi-Prestel) Let K be compact and Assumption 1 hold. Then [ f R[X] and f > 0 on K ] f Q(g 1,..., g m ), i.e., m f (x) = σ 0 (x) + σ j (x) g j (x), j=1 for some s.o.s. polynomials {σ j } m j=0. x R n

If one fixes an a priori bound on the degree of the s.o.s. polynomials {σ j }, checking f Q(g 1,..., g m ) reduces to solving a SDP!! Moreover, Assumption 1 holds true if e.g. : - all the g j s are linear (hence K is a polytope), or if - the set { x g j (x) 0} is compact for some j {1,..., m}. If x K x M for some (known) M, then it suffices to add the redundant quadratic constraint M 2 X 2 0, in the definition of K.

If one fixes an a priori bound on the degree of the s.o.s. polynomials {σ j }, checking f Q(g 1,..., g m ) reduces to solving a SDP!! Moreover, Assumption 1 holds true if e.g. : - all the g j s are linear (hence K is a polytope), or if - the set { x g j (x) 0} is compact for some j {1,..., m}. If x K x M for some (known) M, then it suffices to add the redundant quadratic constraint M 2 X 2 0, in the definition of K.

If one fixes an a priori bound on the degree of the s.o.s. polynomials {σ j }, checking f Q(g 1,..., g m ) reduces to solving a SDP!! Moreover, Assumption 1 holds true if e.g. : - all the g j s are linear (hence K is a polytope), or if - the set { x g j (x) 0} is compact for some j {1,..., m}. If x K x M for some (known) M, then it suffices to add the redundant quadratic constraint M 2 X 2 0, in the definition of K.

If one fixes an a priori bound on the degree of the s.o.s. polynomials {σ j }, checking f Q(g 1,..., g m ) reduces to solving a SDP!! Moreover, Assumption 1 holds true if e.g. : - all the g j s are linear (hence K is a polytope), or if - the set { x g j (x) 0} is compact for some j {1,..., m}. If x K x M for some (known) M, then it suffices to add the redundant quadratic constraint M 2 X 2 0, in the definition of K.

Krivine-Handelman-Vasilescu Positivstellensatz Assumption I: With g 0 = 1, the family {g 0,..., g m } generates the algebra R[x], that is, R[x 1,..., x n ] = R[g 0,..., g m ]. Assumption II: Recall that K is compact. Hence we also assume with no loss of generality (but possibly after scaling) that for every j = 1,..., m: 0 g j (x) 1 x K.

Krivine-Handelman-Vasilescu Positivstellensatz Assumption I: With g 0 = 1, the family {g 0,..., g m } generates the algebra R[x], that is, R[x 1,..., x n ] = R[g 0,..., g m ]. Assumption II: Recall that K is compact. Hence we also assume with no loss of generality (but possibly after scaling) that for every j = 1,..., m: 0 g j (x) 1 x K.

Notation: for α, β N m, let g(x) α = g 1 (x) α1 g m (x) αm (1 g(x)) β = (1 g 1 (x)) β1 (1 g m (x)) βm Theorem (Krivine,Vasilescu Positivstellensatz) Let Assumption I and Assumption II hold: If f R[x 1,..., x m ] is POSITIVE on K then f (x) = c αβ g(x) α (1 g(x)) β, x R n, α,β N m for finitely many positive coefficients (c αβ ).

Notation: for α, β N m, let g(x) α = g 1 (x) α1 g m (x) αm (1 g(x)) β = (1 g 1 (x)) β1 (1 g m (x)) βm Theorem (Krivine,Vasilescu Positivstellensatz) Let Assumption I and Assumption II hold: If f R[x 1,..., x m ] is POSITIVE on K then f (x) = c αβ g(x) α (1 g(x)) β, x R n, α,β N m for finitely many positive coefficients (c αβ ).

f (x) = Testing whether α,β N m c αβ g(x) α (1 g(x)) β, x R n, for finitely many positive coefficients (c αβ ) and with i α i + β i d... reduces to solving a LP!. Indeed, recall that f (x) = γ f γ x γ. So expand and state that α,β N m c αβ g(x) α (1 g(x)) β = γ N n θ γ (c) x γ f γ = θ γ (c), γ N n 2d ; c 0. a linear system!

f (x) = Testing whether α,β N m c αβ g(x) α (1 g(x)) β, x R n, for finitely many positive coefficients (c αβ ) and with i α i + β i d... reduces to solving a LP!. Indeed, recall that f (x) = γ f γ x γ. So expand and state that α,β N m c αβ g(x) α (1 g(x)) β = γ N n θ γ (c) x γ f γ = θ γ (c), γ N n 2d ; c 0. a linear system!

f (x) = Testing whether α,β N m c αβ g(x) α (1 g(x)) β, x R n, for finitely many positive coefficients (c αβ ) and with i α i + β i d... reduces to solving a LP!. Indeed, recall that f (x) = γ f γ x γ. So expand and state that α,β N m c αβ g(x) α (1 g(x)) β = γ N n θ γ (c) x γ f γ = θ γ (c), γ N n 2d ; c 0. a linear system!

DUAL side: The K-moment problem Let {X α } be a canonical basis for R[X], and let y := {y α } be a given sequence indexed in that basis. Recall the K-moment problem Given K R n, does there exist a measure µ on K, such that y α = X α dµ, α N n? (where X α = X α 1 1 X αn n ). K

Given y = {y α }, let L y : R[X] R, be the linear functional f (= α f α X α ) L y (f ) := α N n f α y α. Moment matrix M d (y) with rows and columns also indexed in the basis {X α }. M d (y)(α, β) := L y (X α+β ) = y α+β, α, β N n, α, β d.

For instance in R 2 : M 1 (y) = 1 {}}{ y 00 X 1 {}}{ {}}{ X 2 y 10 y 01 y 10 y 20 y 11 y 01 y 11 y 02 Importantly... M d (y) 0 L y (h 2 ) 0, h R[X] d

Localizing matrix The Localizing matrix" M d (θy) w.r.t. a polynomial θ R[X] with X θ(x) = γ θ γ X γ, has its rows and columns also indexed in the basis {X α } of R[X] d, and with entries: M d (θ y)(α, β) = L y (θ X α+β ) = θ γ y α+β+γ, γ N n { α, β N n α, β d.

For instance, in R 2, and with X θ(x) := 1 X 2 1 X 2 2, M 1 (θ y) = 1 {}}{ y 00 y 20 y 02, X 1 {}}{ y 10 y 30 y 12, X 2 {}}{ y 01 y 21 y 03 y 10 y 30 y 12, y 20 y 40 y 22, y 11 y 21 y 12 y 01 y 21 y 03, y 11 y 21 y 12, y 02 y 22 y 04. Importantly... M d (θ y) 0 L y (h 2 θ) 0, h R[X] d

Putinar s dual conditions Again K := { x R n g j (x) 0, j = 1,..., m}. Assumption 1: For some M > 0, the quadratic polynomial M X 2 is in the quadratic module Q(g 1,..., g m ) Theorem (Putinar: dual side) Let K be compact, and Assumption 1 hold. Then a sequence y = (y α ), α N n, has a representing measure µ on K if and only if ( ) L y (f 2 ) 0; L y (f 2 g j ) 0, j = 1,..., m; f R[X].

Putinar s dual conditions Again K := { x R n g j (x) 0, j = 1,..., m}. Assumption 1: For some M > 0, the quadratic polynomial M X 2 is in the quadratic module Q(g 1,..., g m ) Theorem (Putinar: dual side) Let K be compact, and Assumption 1 hold. Then a sequence y = (y α ), α N n, has a representing measure µ on K if and only if ( ) L y (f 2 ) 0; L y (f 2 g j ) 0, j = 1,..., m; f R[X].

Checking whether (**) holds for all f R[X] with degree d reduces to checking whether M d (y) 0 and M d (g j y) 0, for all j = 1,..., m! m + 1 LMI conditions to verify!

Krivine-Vasilescu: dual side Theorem Let K be compact, and Assumption I and II hold. Then the sequence y = (y α ), α N n, has a representing measure µ on K if and only if L y (g α (1 g) β ) 0, α, β N m.

LP-relaxations With f R[x], consider the hierarchy of LP-relaxations ρ d = max λ,c αβ λ f λ = α,β N m c αβ g α (1 g) β c αβ 0, α + β 2d and of course ρ d = ρ d for all d.

Theorem Assume that K is compact and Assumption I and II hold. Then the LP-relaxations CONVERGE, that is, ρ d f as d. The SHERALI-ADAMS RLT s hierarchy is exactly this type of LP-relaxations. Its convergence for 0/1 programs was proved with ah-hoc arguments. In fact, the rationale behind such convergence if Krivine-Vasilescu Positivstellensatz.

Theorem Assume that K is compact and Assumption I and II hold. Then the LP-relaxations CONVERGE, that is, ρ d f as d. The SHERALI-ADAMS RLT s hierarchy is exactly this type of LP-relaxations. Its convergence for 0/1 programs was proved with ah-hoc arguments. In fact, the rationale behind such convergence if Krivine-Vasilescu Positivstellensatz.

Theorem Assume that K is compact and Assumption I and II hold. Then the LP-relaxations CONVERGE, that is, ρ d f as d. The SHERALI-ADAMS RLT s hierarchy is exactly this type of LP-relaxations. Its convergence for 0/1 programs was proved with ah-hoc arguments. In fact, the rationale behind such convergence if Krivine-Vasilescu Positivstellensatz.

Theorem Assume that K is compact and Assumption I and II hold. Then the LP-relaxations CONVERGE, that is, ρ d f as d. The SHERALI-ADAMS RLT s hierarchy is exactly this type of LP-relaxations. Its convergence for 0/1 programs was proved with ah-hoc arguments. In fact, the rationale behind such convergence if Krivine-Vasilescu Positivstellensatz.

Some remarks on LP-relaxations 1. Notice the presence of binomial coefficients in both primal and dual LP-relaxations... which yields numericall ill-conditioning for relatively large d. 2. Let x K be a global minimizer, and for x K, let J(x) be the set of active constraints, i.e., g j (x) = 0 or...1 g k (x) = 0. Then FINITE convergence CANNOT occur if there exists nonoptimal x K with J(x) J(x )! And so... not possible for CONVEX problems in general! For instance, if K is a Polytope then FINITE convergence is possible only if every global minimizer is a vertex of K!

Some remarks on LP-relaxations 1. Notice the presence of binomial coefficients in both primal and dual LP-relaxations... which yields numericall ill-conditioning for relatively large d. 2. Let x K be a global minimizer, and for x K, let J(x) be the set of active constraints, i.e., g j (x) = 0 or...1 g k (x) = 0. Then FINITE convergence CANNOT occur if there exists nonoptimal x K with J(x) J(x )! And so... not possible for CONVEX problems in general! For instance, if K is a Polytope then FINITE convergence is possible only if every global minimizer is a vertex of K!

Some remarks on LP-relaxations 1. Notice the presence of binomial coefficients in both primal and dual LP-relaxations... which yields numericall ill-conditioning for relatively large d. 2. Let x K be a global minimizer, and for x K, let J(x) be the set of active constraints, i.e., g j (x) = 0 or...1 g k (x) = 0. Then FINITE convergence CANNOT occur if there exists nonoptimal x K with J(x) J(x )! And so... not possible for CONVEX problems in general! For instance, if K is a Polytope then FINITE convergence is possible only if every global minimizer is a vertex of K!

Some remarks on LP-relaxations 1. Notice the presence of binomial coefficients in both primal and dual LP-relaxations... which yields numericall ill-conditioning for relatively large d. 2. Let x K be a global minimizer, and for x K, let J(x) be the set of active constraints, i.e., g j (x) = 0 or...1 g k (x) = 0. Then FINITE convergence CANNOT occur if there exists nonoptimal x K with J(x) J(x )! And so... not possible for CONVEX problems in general! For instance, if K is a Polytope then FINITE convergence is possible only if every global minimizer is a vertex of K!

Some remarks on LP-relaxations 1. Notice the presence of binomial coefficients in both primal and dual LP-relaxations... which yields numericall ill-conditioning for relatively large d. 2. Let x K be a global minimizer, and for x K, let J(x) be the set of active constraints, i.e., g j (x) = 0 or...1 g k (x) = 0. Then FINITE convergence CANNOT occur if there exists nonoptimal x K with J(x) J(x )! And so... not possible for CONVEX problems in general! For instance, if K is a Polytope then FINITE convergence is possible only if every global minimizer is a vertex of K!

LP- and SDP-Relaxations with their dual Primal LP-relaxation Primal SDP-relaxation min y L y (f ) L y (g α (1 g) β ) 0 α, β N m, α + β 2d min y L y (f ) L y (h 2 g j ) 0, j = 1,..., m h, deg(h g j ) 2d, j m Dual LP-relaxation Dual SDP-relaxation max λ λ,{c αβ } f λ = c αβ g α (1 g) β α,β N m c αβ 0; α + β 2d max λ,{σ j } λ f λ = deg(σ j g j ) σ m σ j g j j=0 2d, j m s.o.s., j m