The moment-lp and moment-sos approaches LAAS-CNRS and Institute of Mathematics, Toulouse, France CIRM, November 2013
Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators
Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators
Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators
Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators
Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators
Semidefinite Programming Why polynomial optimization? LP- and SDP- CERTIFICATES of POSITIVITY The moment-lp and moment-sos approaches Two examples outside optimization: Approximating sets defined with quantifiers Convex polynomial underestimators
Semidefinite Programming The CONVEX optimization problem: P min x R n { c x n A i x i i=1 b}, where c R n and b, A i S m (m m symmetric matrices), is called a semidefinite program. The notation 0" means the real symmetric matrix " is positive semidefinite, i.e., all its (real) EIGENVALUES are nonnegative.
Example P : min {x 1 + x 2 : x [ 3 + 2x1 + x s.t. 2 x 1 5 x 1 5 x 1 2x 2 or, equivalently ] }, 0 P : min {x 1 + x 2 : x [ 3 5 s.t. 5 0 ] + x 1 [ 2 1 1 1 ] + x 2 [ 1 0 0 2 ] } 0
P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...
P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...
P and its dual P are convex problems that are solvable in polynomial time to arbitrary precision ɛ > 0. = generalization to the convex cone S + m (X 0) of Linear Programming on the convex polyhedral cone R m + (x 0). Indeed, with DIAGONAL matrices Semidefinite programming = Linear Programming! Several academic SDP software packages exist, (e.g. MATLAB LMI toolbox, SeduMi, SDPT3,...). However, so far, size limitation is more severe than for LP software packages. Pioneer contributions by A. Nemirovsky, Y. Nesterov, N.Z. Shor, B.D. Yudin,...
Why Polynomial Optimization? After all... the polynomial optimization problem: f = min{f (x) : g j (x) 0, j = 1,..., m} is just a particular case of Non Linear Programming (NLP)! True!... if one is interested with a LOCAL optimum only!!
Why Polynomial Optimization? After all... the polynomial optimization problem: f = min{f (x) : g j (x) 0, j = 1,..., m} is just a particular case of Non Linear Programming (NLP)! True!... if one is interested with a LOCAL optimum only!!
When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!
When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!
When searching for a local minimum... Optimality conditions and descent algorithms use basic tools from REAL and CONVEX analysis and linear algebra The focus is on how to improve f by looking at a NEIGHBORHOOD of a nominal point x K, i.e., LOCALLY AROUND x K, and in general, no GLOBAL property of x K can be inferred. The fact that f and g j are POLYNOMIALS does not help much!
BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}.... and so to compute f one needs TRACTABLE CERTIFICATES of POSITIVITY on K!
BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}.... and so to compute f one needs TRACTABLE CERTIFICATES of POSITIVITY on K!
BUT for GLOBAL Optimization... the picture is different! Remember that for the GLOBAL minimum f : f = sup { λ : f (x) λ 0 x K}.... and so to compute f one needs TRACTABLE CERTIFICATES of POSITIVITY on K!
REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)
REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)
REAL ALGEBRAIC GEOMETRY helps!!!! Indeed, POWERFUL CERTIFICATES OF POSITIVITY EXIST! Moreover... and importantly, Such certificates are amenable to PRACTICAL COMPUTATION! ( Stronger Positivstellensatzë exist for analytic functions but are useless from a computational viewpoint.)
SOS-based certificate K = {x : g j (x) 0, j = 1,..., m} Theorem (Putinar s Positivstellensatz) If K is compact (+ a technical Archimedean assumption) and f > 0 on K then: f (x) = σ 0 (x) + m σ j (x) g j (x), x R n, j=1 for some SOS polynomials (σ j ) R[x]. Testing whether holds for some SOS (σ j ) R[x] with a degree bound, is SOLVING an SDP!
SOS-based certificate K = {x : g j (x) 0, j = 1,..., m} Theorem (Putinar s Positivstellensatz) If K is compact (+ a technical Archimedean assumption) and f > 0 on K then: f (x) = σ 0 (x) + m σ j (x) g j (x), x R n, j=1 for some SOS polynomials (σ j ) R[x]. Testing whether holds for some SOS (σ j ) R[x] with a degree bound, is SOLVING an SDP!
LP-based certificate K = {x : g j (x) 0; (1 g j (x)) 0, j = 1,..., m} Theorem (Krivine-Vasilescu-Handelman s Positivstellensatz) Let K be compact and the family {g j, (1 g j )} generate R[x]. If f > 0 on K then: f (x) = α,β c αβ m j=1 g j (x) α j (1 g j (x)) β j,, x R n, for some NONNEGATIVE scalars (c αβ ). Testing whether holds for some NONNEGATIVE (c αβ ) with α + β M, is SOLVING an LP!
LP-based certificate K = {x : g j (x) 0; (1 g j (x)) 0, j = 1,..., m} Theorem (Krivine-Vasilescu-Handelman s Positivstellensatz) Let K be compact and the family {g j, (1 g j )} generate R[x]. If f > 0 on K then: f (x) = α,β c αβ m j=1 g j (x) α j (1 g j (x)) β j,, x R n, for some NONNEGATIVE scalars (c αβ ). Testing whether holds for some NONNEGATIVE (c αβ ) with α + β M, is SOLVING an LP!
SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY! Farkas Lemma Krivine-Stengle KKT-Optimality conditions Schmüdgen-Putinar
SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY! Farkas Lemma Krivine-Stengle KKT-Optimality conditions Schmüdgen-Putinar
SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY! Farkas Lemma Krivine-Stengle KKT-Optimality conditions Schmüdgen-Putinar
SUCH POSITIVITY CERTIFICATES allow to infer GLOBAL Properties of FEASIBILITY and OPTIMALITY,... the analogue of (well-known) previous ones valid in the CONVEX CASE ONLY! Farkas Lemma Krivine-Stengle KKT-Optimality conditions Schmüdgen-Putinar
In addition, polynomials NONNEGATIVE ON A SET K R n are ubiquitous. They appear in many important applications, and not only in global optimization! For instance, one may also want: To approximate sets defined with QUANTIFIERS, like.e.g., R f := {x B : f (x, y) 0 for all y such that (x, y) K} D f := {x B : f (x, y) 0 for some y such that (x, y) K} where f R[x, y], B is a simple set (box, ellipsoid). To compute convex polynomial underestimators p f of a polynomial f on a box B R n. (Very useful in MINLP.)
In addition, polynomials NONNEGATIVE ON A SET K R n are ubiquitous. They appear in many important applications, and not only in global optimization! For instance, one may also want: To approximate sets defined with QUANTIFIERS, like.e.g., R f := {x B : f (x, y) 0 for all y such that (x, y) K} D f := {x B : f (x, y) 0 for some y such that (x, y) K} where f R[x, y], B is a simple set (box, ellipsoid). To compute convex polynomial underestimators p f of a polynomial f on a box B R n. (Very useful in MINLP.)
The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Stengle s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In may situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.
The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Stengle s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In may situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.
The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Stengle s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In may situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.
The moment-lp and moment-sos approaches consist of using a certain type of positivity certificate (Krivine-Stengle s or Putinar s certificate) in potentially any application where such a characterization is needed. (Global optimization is only one example.) In may situations this amounts to solving a HIERARCHY of : LINEAR PROGRAMS, or SEMIDEFINITE PROGRAMS... of increasing size!.
LP- and SDP-hierarchies for optimization Replace f = sup λ,σj { λ : f (x) λ 0 x K} with: The SDP-hierarchy indexed by d N: m fd = sup { λ : f λ = σ 0 + σ }{{} j g j ; deg (σ j g j ) 2d } }{{} SOS j=1 SOS or, the LP-hierarchy indexed by d N: θ d = sup { λ : f λ = α,β c αβ }{{} 0 m α g j j (1 g j ) β j ; α + β 2d} j=1
LP- and SDP-hierarchies for optimization Replace f = sup λ,σj { λ : f (x) λ 0 x K} with: The SDP-hierarchy indexed by d N: m fd = sup { λ : f λ = σ 0 + σ }{{} j g j ; deg (σ j g j ) 2d } }{{} SOS j=1 SOS or, the LP-hierarchy indexed by d N: θ d = sup { λ : f λ = α,β c αβ }{{} 0 m α g j j (1 g j ) β j ; α + β 2d} j=1
Theorem Both sequence (f d ), and (θ d), d N, are MONOTONE NON DECREASING and when K is compact (and satisfies a technical Archimedean assumption) then: f = lim fd = lim θ d. d d
What makes this approach exciting is that it is at the crossroads of several disciplines/applications: Commutative, Non-commutative, and Non-linear ALGEBRA Real algebraic geometry, and Functional Analysis Optimization, Convex Analysis Computational Complexity in Computer Science, which BENEFIT from interactions! As mentioned... potential applications are ENDLESS!
What makes this approach exciting is that it is at the crossroads of several disciplines/applications: Commutative, Non-commutative, and Non-linear ALGEBRA Real algebraic geometry, and Functional Analysis Optimization, Convex Analysis Computational Complexity in Computer Science, which BENEFIT from interactions! As mentioned... potential applications are ENDLESS!
Has already been proved useful and successful in applications with modest problem size, notably in optimization, control, robust control, optimal control, estimation, computer vision, etc. HAS initiated and stimulated new research issues: in Convex Algebraic Geometry (e.g. semidefinite representation of convex sets, algebraic degree of semidefinite programming and polynomial optimization) in Computational algebra (e.g., for solving polynomial equations via SDP and Border bases) Computational Complexity where LP- and SDP-HIERARCHIES have become an important tool to analyze Hardness of Approximation for 0/1 combinatorial problems ( links with quantum computing)
A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.
A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.
A remarkable property of the SOS hierarchy: I When solving the optimization problem P : f = min {f (x) : g j (x) 0, j = 1,..., m} one does NOT distinguish between CONVEX, CONTINUOUS NON CONVEX, and 0/1 (and DISCRETE) problems! A boolean variable x i is modelled via the equality constraint x 2 i x i = 0". In Non Linear Programming (NLP), modeling a 0/1 variable with the polynomial equality constraint xi 2 x i = 0" and applying a standard descent algorithm would be considered stupid"! Each class of problems has its own ad hoc tailored algorithms.
Even though the moment-sos approach DOES NOT SPECIALIZES to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. (Finite convergence also occurs for general convex problems.) (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems!
Even though the moment-sos approach DOES NOT SPECIALIZES to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. (Finite convergence also occurs for general convex problems.) (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems!
Even though the moment-sos approach DOES NOT SPECIALIZES to each class of problems: It recognizes the class of (easy) SOS-convex problems as FINITE CONVERGENCE occurs at the FIRST relaxation in the hierarchy. (Finite convergence also occurs for general convex problems.) (NOT true for the LP-hierarchy.) The SOS-hierarchy dominates other lift-and-project hierarchies (i.e. provides the best lower bounds) for hard 0/1 combinatorial optimization problems!
A remarkable property: II FINITE CONVERGENCE of the SOS-hierarchy is GENERIC!... and provides a GLOBAL OPTIMALITY CERTIFICATE, the analogue for the NON CONVEX CASE of the KKT-OPTIMALITY conditions in the CONVEX CASE!
Theorem (Marshall, Nie) Let x K be a global minimizer of P : f = min {f (x) : g j (x) 0, j = 1,..., m}. and assume that: (i) The gradients { g j (x )} are linearly independent, (ii) Strict complementarity holds (λ j g j (x ) = 0 for all j.) (iii) Second-order sufficiency conditions hold at (x, λ ) K R m +. m Then f (x) f = σ0 (x) + σj (x)g j(x), SOS polynomials {σ j }. j=1 x R n, for some Moreover, the conditions (i)-(ii)-(iii) HOLD GENERICALLY!
Theorem (Marshall, Nie) Let x K be a global minimizer of P : f = min {f (x) : g j (x) 0, j = 1,..., m}. and assume that: (i) The gradients { g j (x )} are linearly independent, (ii) Strict complementarity holds (λ j g j (x ) = 0 for all j.) (iii) Second-order sufficiency conditions hold at (x, λ ) K R m +. m Then f (x) f = σ0 (x) + σj (x)g j(x), SOS polynomials {σ j }. j=1 x R n, for some Moreover, the conditions (i)-(ii)-(iii) HOLD GENERICALLY!
In summary: KKT-OPTIMALITY when f and g j are CONVEX m f (x ) λ j g j(x ) = 0 f (x) f j=1 m λ j g j(x) j=1 PUTINAR s CERTIFICATE in the non CONVEX CASE m f (x ) σ j (x ) g j (x ) = 0 f (x) f j=1 m σj (x)g j(x) j=1 0 for all x R n (= σ0 (x)) 0 for all x Rn. for some SOS {σj }, and σj (x ) = λ j.
In summary: KKT-OPTIMALITY when f and g j are CONVEX m f (x ) λ j g j(x ) = 0 f (x) f j=1 m λ j g j(x) j=1 PUTINAR s CERTIFICATE in the non CONVEX CASE m f (x ) σ j (x ) g j (x ) = 0 f (x) f j=1 m σj (x)g j(x) j=1 0 for all x R n (= σ0 (x)) 0 for all x Rn. for some SOS {σj }, and σj (x ) = λ j.
II. Approximation of sets with quantifiers Let f R[x, y] and let K R n R p be the semi-algebraic set: K := {(x, y) : g j (x, y) 0, j = 1,..., m}, and let B R n be the unit ball or the [ 1, 1] n. Suppose that one wants to approximate the set: R f := {x B : f (x, y) 0 for all y such that (x, y) K} as closely as desired by a sequence of sets of the form: Θ k := {x B : J k (x) 0 } for some polynomials J k.
With g 0 = 1 and with K R n R p and k N, let m Q k (g) := σ j (x, y) g j (x, y) : σ j Σ[x, y], deg σ j g j 2k j=0 Let x F(x) := max {f (x, y) : (x, y) K }, and for every integer k consider the optimization problem: ρ k = min J R[x] k { } (J F ) dx : J(x) f (x, y) Q k (g) B
1. The criterion B (J F) dx = F dx B }{{} + α unknown but constant J α x α dx B }{{} easy to compute is LINEAR in the coefficients J α of the unknown polynomial J R[x] k! 2. The constraint J(x) f (x, y) = m σ j (x, y) g j (x, y) j=0 is just LINEAR CONSTRAINTS + LMIs!
1. The criterion B (J F) dx = F dx B }{{} + α unknown but constant J α x α dx B }{{} easy to compute is LINEAR in the coefficients J α of the unknown polynomial J R[x] k! 2. The constraint J(x) f (x, y) = m σ j (x, y) g j (x, y) j=0 is just LINEAR CONSTRAINTS + LMIs!
ρ k = min J R[x] k Hence, the optimization problem { } (J F ) dx : J(x) f (x, y) Q k (g) B IS AN SDP! Moreover, it has an optimal solution J k R[x] k! Alternatively, if one uses LP-based positivity certificates for J(x) f (x, y), one ends up with solving an LP! From the definition of Jk, the sublevel sets Θ k := {x B : J k (x) 0} R f, k N, provide a nested sequence of INNNER approximations of R f.
ρ k = min J R[x] k Hence, the optimization problem { } (J F ) dx : J(x) f (x, y) Q k (g) B IS AN SDP! Moreover, it has an optimal solution J k R[x] k! Alternatively, if one uses LP-based positivity certificates for J(x) f (x, y), one ends up with solving an LP! From the definition of Jk, the sublevel sets Θ k := {x B : J k (x) 0} R f, k N, provide a nested sequence of INNNER approximations of R f.
ρ k = min J R[x] k Hence, the optimization problem { } (J F ) dx : J(x) f (x, y) Q k (g) B IS AN SDP! Moreover, it has an optimal solution J k R[x] k! Alternatively, if one uses LP-based positivity certificates for J(x) f (x, y), one ends up with solving an LP! From the definition of Jk, the sublevel sets Θ k := {x B : J k (x) 0} R f, k N, provide a nested sequence of INNNER approximations of R f.
Theorem (Lass) (Strong) convergence in L 1 (B)-norm takes place, that is: Jk F dx = 0 lim k B and, if in addition the set {x B : F(x) = 0} has Lebesgue measure zero, then lim VOL(R f \ Θ k ) = 0 k
Theorem (Lass) (Strong) convergence in L 1 (B)-norm takes place, that is: Jk F dx = 0 lim k B and, if in addition the set {x B : F(x) = 0} has Lebesgue measure zero, then lim VOL(R f \ Θ k ) = 0 k
Ex: Polynomial Matrix Inequalities: (with D. Henrion) Let x A(x) R p p where A(x) is the matrix-polynomial ) x A(x) = ( A α x α = A α x α 1 1 x n αn α N n α N n. for finitely many real symmetric matrices (A α ), α N n.... and suppose one wants to approximate the set R A := {x B : A(x) 0} = {x : λ min (A(x)) 0}. R A = Then: x B : y T A(x)y }{{} f (x,y) 0, y s.t. y 2 = 1
Ex: Polynomial Matrix Inequalities: (with D. Henrion) Let x A(x) R p p where A(x) is the matrix-polynomial ) x A(x) = ( A α x α = A α x α 1 1 x n αn α N n α N n. for finitely many real symmetric matrices (A α ), α N n.... and suppose one wants to approximate the set R A := {x B : A(x) 0} = {x : λ min (A(x)) 0}. R A = Then: x B : y T A(x)y }{{} f (x,y) 0, y s.t. y 2 = 1
Illustrative example (continued) Let B be the unit disk {x : x 1} and let: { ( [ 1 16x1 x R A := x B : A(x) = 2 x 1 x 1 1 x1 2 x 2 2 ]) } 0 Then by solving relatively simple semidefinite programs, one may approximate R A with sublevel sets of the form: Θ k := {x B : J k (x) 0 } for some polynomial J k of degree k = 2, 4,... and with VOL (R A \ Θ k ) 0 as k.
Illustrative example (continued) Let B be the unit disk {x : x 1} and let: { ( [ 1 16x1 x R A := x B : A(x) = 2 x 1 x 1 1 x1 2 x 2 2 ]) } 0 Then by solving relatively simple semidefinite programs, one may approximate R A with sublevel sets of the form: Θ k := {x B : J k (x) 0 } for some polynomial J k of degree k = 2, 4,... and with VOL (R A \ Θ k ) 0 as k.
1 1 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0.2 x 2 0 x 2 0 0.2 0.2 0.4 0.4 0.6 0.6 0.8 0.8 1 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 x 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 x 1 Θ 2 (left) and Θ 4 (right) inner approximations (light gray) of (dark gray) embedded in unit disk B (dashed).
1 1 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0.2 x 2 0 x 2 0 0.2 0.2 0.4 0.4 0.6 0.6 0.8 0.8 1 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 x 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 x 1 Θ 6 (left) and Θ 8 (right) inner approximations (light gray) of (dark gray) embedded in unit disk B (dashed).
Similarly, suppose that one wants to approximate the set: D f := {x B : f (x, y) 0 for some y such that (x, y) K} as closely as desired by a sequence of sets of the form: Θ k := {x B : J k (x) 0 } for some polynomials J k.
Let x F(x) := min {f (x, y) : (x, y) K }, and for every integer k the optimization problem: ρ k = max J R[x] k { } (F J) dx : J(x) f (x, y) Q k (g) B IS AN SDP with an optimal solution J k R[x] k. From the definition of Jk, the sublevel sets Θ k := {x B : J k (x) 0} D f, k N, provide a nested sequence of OUTER approximations of D f.
Theorem (Lass) (Strong) convergence in L 1 (B)-norm takes place, that is: F Jk dx = 0 lim k B and, if in addition the set {x B : F(x) = 0} has Lebesgue measure zero, then lim VOL(Θ k \ D f ) = 0 k
Theorem (Lass) (Strong) convergence in L 1 (B)-norm takes place, that is: F Jk dx = 0 lim k B and, if in addition the set {x B : F(x) = 0} has Lebesgue measure zero, then lim VOL(Θ k \ D f ) = 0 k
III. Convex underestimators of polynomials In large scale Mixed Integer Nonlinear Programming (MINLP), a popular method is to use B & B where LOWER BOUNDS at each node of the search tree must be computed EFFICIENTLY! In such a case... one needs CONVEX UNDERESTIMATORS of the objective function, say on a BOX B R n! Message: Good" CONVEX POLYNOMIAL UNDERESTIMATORS can be computed efficienty!
III. Convex underestimators of polynomials In large scale Mixed Integer Nonlinear Programming (MINLP), a popular method is to use B & B where LOWER BOUNDS at each node of the search tree must be computed EFFICIENTLY! In such a case... one needs CONVEX UNDERESTIMATORS of the objective function, say on a BOX B R n! Message: Good" CONVEX POLYNOMIAL UNDERESTIMATORS can be computed efficienty!
inf p R[x] d Solving { (f (x) p(x)) dx : B s.t. f p 0 on B and p convex on B} will provide a degree-d POLYNOMIAL CONVEX UNDERESTIMATOR p of f on B that minimizes the L 1 (B)-norm f p 1! Notice that: (f (x) p(x)) dx is LINEAR in the coefficients of p! B p convex on B y T 2 p(x) y }{{} R[xy] d 0 on B {y : y 2 = 1}!
Hence replace the positivity and convexity constraints f p 0 on B and p convex on B with the positivity certificates f (x) p(x) = y T 2 p(x) y = m σ j (x) g j (x) }{{} SOS m ψ(x, y) g }{{} j (x) + ψ m+1 (x, y) (1 y 2 ) SOS k=0 k=0
Hence replace the positivity and convexity constraints f p 0 on B and p convex on B with the positivity certificates f (x) p(x) = y T 2 p(x) y = m σ j (x) g j (x) }{{} SOS m ψ(x, y) g }{{} j (x) + ψ m+1 (x, y) (1 y 2 ) SOS k=0 k=0
and apply the moment-sos approach to obtain a sequence of polynomials p k R[x] d, k N, of degree d which converges to the BEST convex polynomial underestimator of degree d.
Conclusion The moment-sos hierarchy is a powerful general methodology. Works for problems of modest size (or larger size problem with sparsity and/or symmetries) f (x) = α,β Mixed LP-SOS positivity certificate c αβ }{{} 0 where k IS FIXED! g j (x) α j j An alternative for larger size problems? j (1 g j (x)) β j + σ 0 (x) }{{} SOS of degree k
Conclusion The moment-sos hierarchy is a powerful general methodology. Works for problems of modest size (or larger size problem with sparsity and/or symmetries) f (x) = α,β Mixed LP-SOS positivity certificate c αβ }{{} 0 where k IS FIXED! g j (x) α j j An alternative for larger size problems? j (1 g j (x)) β j + σ 0 (x) }{{} SOS of degree k
THANK YOU!!