Solving systems of polynomial equations and minimization of multivariate polynomials

Similar documents
Lecture 15: Algebraic Geometry II

POLYNOMIAL DIVISION AND GRÖBNER BASES. Samira Zeada

Algebraic Optimization with Applications to System Theory

PREMUR Seminar Week 2 Discussions - Polynomial Division, Gröbner Bases, First Applications

Problem Set 1 Solutions

A decoding algorithm for binary linear codes using Groebner bases

Groebner Bases, Toric Ideals and Integer Programming: An Application to Economics. Tan Tran Junior Major-Economics& Mathematics

Groebner Bases and Applications

A gentle introduction to Elimination Theory. March METU. Zafeirakis Zafeirakopoulos

Abstract Algebra for Polynomial Operations. Maya Mohsin Ahmed

Rewriting Polynomials

ABSTRACT. Department of Mathematics. interesting results. A graph on n vertices is represented by a polynomial in n

4. Convex optimization problems (part 1: general)

Lecture 1. (i,j) N 2 kx i y j, and this makes k[x, y]

Lesson 14 Properties of Groebner Bases

Optimization over Polynomials with Sums of Squares and Moment Matrices

Math 615: Lecture of January 10, 2007

Counting Zeros over Finite Fields with Gröbner Bases

On the minimal free resolution of a monomial ideal.

Gröbner Bases. eliminating the leading term Buchberger s criterion and algorithm. construct wavelet filters

4. Convex optimization problems

Lecture 2: Gröbner Basis and SAGBI Basis

Minimizing Cubic and Homogeneous Polynomials over Integers in the Plane

4-1 The Algebraic-Geometric Dictionary P. Parrilo and S. Lall, ECC

Polynomials, Ideals, and Gröbner Bases

M3P23, M4P23, M5P23: COMPUTATIONAL ALGEBRA & GEOMETRY REVISION SOLUTIONS

SUMS OF SQUARES, MOMENT MATRICES AND OPTIMIZATION OVER POLYNOMIALS. February 5, 2008

Onthe Role of the Critical Value Polynomial in Algebraic Optimization

Lecture: Convex Optimization Problems

MCS 563 Spring 2014 Analytic Symbolic Computation Monday 27 January. Gröbner bases

Reversely Well-Ordered Valuations on Polynomial Rings in Two Variables

4 Unit Math Homework for Year 12

1.3 Algebraic Expressions. Copyright Cengage Learning. All rights reserved.

GRÖBNER BASES AND POLYNOMIAL EQUATIONS. 1. Introduction and preliminaries on Gróbner bases

arxiv: v1 [hep-lat] 13 Oct 2017

Math 4370 Exam 1. Handed out March 9th 2010 Due March 18th 2010

SJÄLVSTÄNDIGA ARBETEN I MATEMATIK

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

MY PUTNAM PROBLEMS. log(1 + x) dx = π2

Lecture 5. 1 Goermans-Williamson Algorithm for the maxcut problem

Handout - Algebra Review

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

B.Sc. MATHEMATICS I YEAR

Global Optimization with Polynomials

Notes 6: Polynomials in One Variable

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include

Common Core State Standards for Mathematics - High School PARRC Model Content Frameworks Mathematics Algebra 2

S-Polynomials and Buchberger s Algorithm

QR Decomposition. When solving an overdetermined system by projection (or a least squares solution) often the following method is used:

COMPUTATIONAL COMMUTATIVE ALGEBRA NOTES

MATH 497A: INTRODUCTION TO APPLIED ALGEBRAIC GEOMETRY

Chapter III. Unconstrained Univariate Optimization

Polynomial Rings. i=0

RINGS: SUMMARY OF MATERIAL

Modular Algorithms for Computing Minimal Associated Primes and Radicals of Polynomial Ideals. Masayuki Noro. Toru Aoyama

Polynomial interpolation over finite fields and applications to list decoding of Reed-Solomon codes

Gröbner Bases, Ideal Computation and Computational Invariant Theory

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Convex optimization problems. Optimization problem in standard form

Semester Review Packet

4. Algebra and Duality

Pearson Mathematics Algebra 2 Common Core 2015

4 Hilbert s Basis Theorem and Gröbner basis

Improved Fast Gauss Transform. Fast Gauss Transform (FGT)

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 18

We say that a polynomial is in the standard form if it is written in the order of decreasing exponents of x. Operations on polynomials:

Another algorithm for nonnegative matrices

Analysis of California Mathematics standards to Common Core standards Algebra I

Algebra 2 Standards. Essential Standards:

4. Convex optimization problems

POSITIVE POLYNOMIALS LECTURE NOTES (09: 10/05/10) Contents

UNIT-I CURVE FITTING AND THEORY OF EQUATIONS

Chapter 8. Exploring Polynomial Functions. Jennifer Huss

An explicit construction of distinguished representations of polynomials nonnegative over finite sets

LMI Methods in Optimal and Robust Control

The simplex algorithm

Unit 13: Polynomials and Exponents

Lecture Note 5: Semidefinite Programming for Stability Analysis

Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra

Semidefinite Programming Basics and Applications

Local Parametrization and Puiseux Expansion

MATH III CCR MATH STANDARDS

Representations of Positive Polynomials: Theory, Practice, and

Non commutative Computations with SINGULAR

1. Algebra 1.5. Polynomial Rings

Introduction to Gröbner Bases for Geometric Modeling. Geometric & Solid Modeling 1989 Christoph M. Hoffmann

Mathematics Standards for High School Algebra II

Advanced SDPs Lecture 6: March 16, 2017

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Non-Convex Optimization via Real Algebraic Geometry

Semidefinite Programming

Convex sets, conic matrix factorizations and conic rank lower bounds

Rational Exponents. Polynomial function of degree n: with leading coefficient,, with maximum number of turning points is given by (n-1)

A field F is a set of numbers that includes the two numbers 0 and 1 and satisfies the properties:

MCS 563 Spring 2014 Analytic Symbolic Computation Friday 31 January. Quotient Rings

Algebraic Geometry for CAGD

Symbolic methods for solving algebraic systems of equations and applications for testing the structural stability

Roots and Coefficients Polynomials Preliminary Maths Extension 1

Groebner Bases in Boolean Rings. for Model Checking and. Applications in Bioinformatics

Solution of Algebric & Transcendental Equations

Transcription:

François Glineur, Polynomial solving & minimization - 1 - First Prev Next Last Full Screen Quit Solving systems of polynomial equations and minimization of multivariate polynomials François Glineur UCL/CORE glineur@core.ucl.ac.be ARC Graphs/CESAME May 12 th 2006

François Glineur, Polynomial solving & minimization - 2 - First Prev Next Last Full Screen Quit Plan Single variable case Newton-type methods, bisection techniques Sum of squares reformulation, companion matrix Algebraic geometry Division, ideals, varieties and Groebner bases Elimination method Stetter and Möller method (companion matrix) Hanzon and Jibetean method Other methods Resultant methods, sum of squares relaxations Concluding remarks

François Glineur, Polynomial solving & minimization - 3 - First Prev Next Last Full Screen Quit Introduction Definitions A monomial with n variables x α = x α 1 1 xα 2 2 xα n n with x R n and α N n A polynomial is a finite linear combination of monomials with coefficients in a field k f = α A a α x α with a α k and A N n finite The set of polynomials over k is denoted k[x 1, x 2,..., x n ] Our main concern is R[x 1, x 2,..., x n ]

Polynomial optimization Given a polynomial f 0 R[x 1, x 2,..., x n ], we will mainly consider unconstrained polynomial optimization: min f 0(x 1, x 2,..., x n ) x R n Obviously many applications Global optimization problem ( local minima) NP-hard problem (w.r.t. number of variables) (take f 0 = i (x2 i 1)2 binary constraints) May have a finite of infinite number of solutions May exhibit solutions at infinity: x 2 1x 4 2 + x 1 x 2 2 + x 2 1 François Glineur, Polynomial solving & minimization - 4 - First Prev Next Last Full Screen Quit

François Glineur, Polynomial solving & minimization - 5 - First Prev Next Last Full Screen Quit Polynomial optimization (cont.) Strong links with polynomial equation solving: find all x R n such that f i (x) = 0 for all 1 i m Stationarity conditions f 0 = f 0 = = f 0 = 0 x 1 x 2 x n define a finite number of connected components: check each to find global minimum Some methods will also be applicable to the constrained case without much change min f 0(x) such that f i (x) = 0 for all 1 i m x R n

Single variable case min x R a nx n + a n 1 x n 1 + + a 2 x 2 + a 1 x + a 0 Various available solution techniques: Newton methods and variants Bisection techniques Sum of squares convex reformulation Companion matrix Utility: Model single variable problems Use as a subroutine for multiple variable case François Glineur, Polynomial solving & minimization - 6 - First Prev Next Last Full Screen Quit

François Glineur, Polynomial solving & minimization - 7 - First Prev Next Last Full Screen Quit Newton methods and variants Iterative schemes Usually very fast convergence to local minimum Multiple zeros might cause trouble Convergence only guaranteed if starting point is close to a (local) minimum Local technique without globalization not suitable to find (all) global minimizers

Bisection techniques Principle: solve stationarity condition as polynomial equation, using Weyl s quadtree algorithm Principle of Weyl s quadtree algorithm to find all complex roots of polynomial equation p(x) = n i=0 a ix i = 0 a. Consider square {x Re(x), Im(x) M} where M is a constant such that M > 2 max i<n (a i /a n ): all roots of p(x) must lie therein b. Recursively partition square into four congruent subsquares François Glineur, Polynomial solving & minimization - 8 - First Prev Next Last Full Screen Quit

François Glineur, Polynomial solving & minimization - 9 - First Prev Next Last Full Screen Quit c. At the center of each subsquare, use computationally cheap procedure to estimate proximity to the closest root of p d. Eliminate subsquare if test guarantees that it contains no root e. Use Newton-type iterations to refine solutions once they are isolated f. Complexity: all zeros are computed with accuracy 2 b M at overall cost O ( n 2 log n log(bn) ) a. o. Optimal methods are even better [V. Pan] : Õ (n) arithmetic or Õ ( n 2 (b + n) ) boolean operations

François Glineur, Polynomial solving & minimization - 10 - First Prev Next Last Full Screen Quit Sum of squares reformulation A single-variable polynomial is nonnegative if and only if it can be written as a sum of squares A polynomial p(x) = 2n i=0 a ix i is a sum of squares if and only if there exists a positive semidefinite matrix M S n+1 such that a k = i+j=k M ij Minimizing p(x) is equivalent to maximizing an additional variable t with the condition that polynomial p(x) t is nonnegative (largest lower bound)

François Glineur, Polynomial solving & minimization - 11 - First Prev Next Last Full Screen Quit Minimizing a single-variable polynomial can thus be reformulated as a convex semidefinite optimization problem It is efficiently solvable, with a polynomial complexity (O ( n) iterations but total complexity worse than quadtree) Note that in principle, this procedure can be used to solve a single-variable polynomial equation

François Glineur, Polynomial solving & minimization - 12 - First Prev Next Last Full Screen Quit Companion matrix Again, try to solve the stationarity condition as a monic polynomial equation p(x) = x n +a n 1 x n 1 +a n 2 x n 2 + +a 2 x 2 +a 1 x+a 0 Fact: p(x) is the characteristic polynomial of 0 1 0 1 A =...... 0 1 a 0 a 1 a 2 a n 1 The eigenvalues of A are thus the r roots of p(x), and complexity of computing these n eigenvalues is O ( n 3) (again worse than quadtree)

François Glineur, Polynomial solving & minimization - 13 - First Prev Next Last Full Screen Quit Algebraic geometry Ideals and varieties Let f 1, f 2,..., f s polynomials in k[x 1, x 2,..., x n ], define the affine variety V (f 1, f 2,..., f s ) = { (x 1, x 2,..., x n ) f i (x) = 0 i } An ideal is a subset of k[x 1, x 2,..., x n ] closed under addition and multiplication by an element of k[x 1, x 2,..., x n ] Let f 1, f 2,..., f s polynomials in k[x 1, x 2,..., x n ], they generate an ideal defined by f 1, f 2,..., f s = { s h i f i with h i k[x 1, x 2,..., x n ] i } i=1

François Glineur, Polynomial solving & minimization - 14 - First Prev Next Last Full Screen Quit Once again, the application of algebraic geometric techniques to polynomial optimization will rely on the resolution of the stationarity conditions as a system of polynomials equations f 1 (x) = f 2 (x) = = f s (x) = 0 identify elements of the affine variety V (f 1, f 2,..., f s ) Key property: f 1, f 2,..., f s = g 1, g 2,..., g t is easily seen to imply V (f 1, f 2,..., f s ) = V (g 1, g 2,..., g t ) look for a generating set that allows for easy identification of the underlying affine variety Ideals/generating sets resemble vector spaces/bases but lack the independence/unicity property The use of Groebner bases is an answer to this problem

Polynomials and division Single variable case: divide f by g unique q and r such that f = qg + r (degr < degg) easy to compute (Euclidean algorithm) Consequence : all ideals in k[x] have the form f Multiple variables: first issue is to define degree (for remainder condition) Define an ordering on the monomials such that a. it is a total order b. x α > x β x α+γ > x β+γ c. well ordered (nonempty subsets have smallest element) François Glineur, Polynomial solving & minimization - 15 - First Prev Next Last Full Screen Quit

The multidegree of a polynomial f = α A a αx α is the maximal α A such that a α 0 Define also the leading term of a polynomial: LM(f) = a α x α with α = multidegf Examples lexicographic order (lex) decide which monomial is greater by looking first at the exponent of x 1, then in case of a tie at the exponent of x 2,, then in case of another tie at the exponent of x 3, etc. François Glineur, Polynomial solving & minimization - 16 - First Prev Next Last Full Screen Quit

graded lexicographic order (grlex) decide which monomial is greater by looking at the sum of their exponents, and use lexicographic order in case of a tie graded reverse lexicographic order (grevlex) decide which monomial is greater by looking at the sum of their exponents, and use reverse lexicographic order in case of a tie (which means look first at x n, then x n 1, etc. but choose the opposite of the first difference observed) François Glineur, Polynomial solving & minimization - 17 - First Prev Next Last Full Screen Quit

François Glineur, Polynomial solving & minimization - 18 - First Prev Next Last Full Screen Quit Multivariable division Divide f by (g 1, g 2,..., g s ) (ordered s-uple) to obtain f = q 1 g 1 + q 2 g 2 + + q s g s + r with multideg f multideg q i g i i and no monomial of r is divisible by any LM(g i ) This can be achieved by a naive iterative scheme (try repeatedly to divide by g 1, then by g 2, etc. ) but uniqueness is not guaranteed Example divide xy 2 x by (xy + 1, y 2 1) to find xy 2 x = y(xy + 1) + 0(y 2 1) + ( x y) while one also has xy 2 x = 0(xy + 1) + x(y 2 1) + 0

François Glineur, Polynomial solving & minimization - 19 - First Prev Next Last Full Screen Quit Groebner bases Definition: the generating set {g 1, g 2,..., g s } of an ideal I = g 1, g 2,..., g s is a Groebner basis if and only if LM(g 1 ),..., LM(g s ) = {x α x α = LM(f) for some f I} a. In a division by the elements of a Groebner basis, the remainder is independent of the order of the g i s (but not the quotients q i ) b. This means that Groebner bases can be used to determine whether a polynomial belongs or not to a given ideal c. For a given monomial ordering, Groebner bases are essentially unique

d. Groebner bases can be computed with Buchberger s algorithm but its complexity is exponential: polynomials of degree bounded by d can exhibit basis elements of degree O (2 2d) e. Computational experience tends to suggest that Groebner bases using the grevlex have much less elements than those computed using the lex or the grlex order f. However, bad behavior can still occur: x n+1 yz n 1 w, xy n 1 z n, x n z y n w with x > y > z > w grevlex leads to z n2 +1 y n2 w in the basis François Glineur, Polynomial solving & minimization - 20 - First Prev Next Last Full Screen Quit

François Glineur, Polynomial solving & minimization - 21 - First Prev Next Last Full Screen Quit The elimination method Let I = f 1, f 2,..., f s k[x 1, x 2,..., x n ] an ideal Define the l th elimination ideal I l = I k[x l+1, x l+2,..., x n ] i.e. the set of all consequences of f 1, f 2,..., f s that do not involve the first l variables x 1, x 2,..., x l If G = {g i } is a Groebner basis for the lex-order, G l = G k[x l+1, x l+2,..., x n ] is a Groebner basis for I l This means in particular that G n 1 involves only x n, from which we can compute potential values for x n. Using then G n 2, one computes values for x n 1, then x n 2 etc. Finally check the validity of all potential solutions found

François Glineur, Polynomial solving & minimization - 22 - First Prev Next Last Full Screen Quit The elimination method (cont.) Easily applicable to the constrained case (use slacks for inequality constraints) Requires finite number of solutions to work Number of potential solutions can grow exponentially with the number of variables Computation of the lex-order Groebner basis can be very slow in practice The following method will tackle this last drawback by allowing any Groebner basis to be used

François Glineur, Polynomial solving & minimization - 23 - First Prev Next Last Full Screen Quit The Stetter-Möller method (1993) Let I = f 1, f 2,..., f s k[x 1, x 2,..., x n ] Consider the space k[x 1, x 2,..., x n ]/I, i.e. the quotient space of the whole set of polynomials by the ideal: this means we do not distinguish between polynomials if they differ by a member of the ideal One can show that this is a vector space whose dimension is finite if and only if the affine variety V (I) is finite Choosing now a Groebner basis for the ideal I, one can check that each class of this quotient space can be represented by its remainder when divided by the Groebner basis

François Glineur, Polynomial solving & minimization - 24 - First Prev Next Last Full Screen Quit The Stetter-Möller method (cont.) Using a basis of k[x 1, x 2,..., x n ]/I (which can be easily done by inspection of the Groebner basis), one can write down A i, the matrix corresponding to the linear operator describing multiplication by x i in the quotient space k[x 1, x 2,..., x n ]/I One has then the following key result: The eigenvalues of A i correspond to the values that variable x i takes on the affine variety V (I) This allows to readily compute all solutions to the polynomial system of equations

François Glineur, Polynomial solving & minimization - 25 - First Prev Next Last Full Screen Quit The Stetter-Möller method (cont.) Easily applicable to the constrained case (use slacks for inequality constraints) Requires finite number of solutions to work Number of potential solutions can grow exponentially with the number of variables Key step is identification of a basis of k[x 1, x 2,..., x n ]/I, which requires the computation of any Groebner basis (this can still be difficult) The next method will try to alleviate the need for the explicit computation of any type of Groebner basis

François Glineur, Polynomial solving & minimization - 26 - First Prev Next Last Full Screen Quit The Hanzon-Jibetean method (2001) Instead of minimizing p(x 1, x 2,..., x n ), consider q λ (x 1, x 2,..., x n ) = p(x 1, x 2,..., x n )+λ(x 2m 1 +x 2m 2 +...+x 2m n ) where λ is a parameter and m an integer such that 2m is greater than the total degree of p We have inf p(x) = lim min q λ(x) x Rn λ 0 x R n minimize the (parameterized) polynomial q(x)

François Glineur, Polynomial solving & minimization - 27 - First Prev Next Last Full Screen Quit The Hanzon-Jibetean method (cont.) Key result: the polynomials arising from the stationarity conditions of q λ always constitute a Groebner basis for the ideal they generate implicit Groebner basis Stetter-Möller method Limit of the eigenvalues λ 0 can be evaluated using a power series technique Not applicable to the constrained case! (breaks implicit Groebner basis) Allows for infinite number of solutions (finds minimumnorm solution one in each connected component)!

François Glineur, Polynomial solving & minimization - 28 - First Prev Next Last Full Screen Quit Resultant methods Resultant methods are based on the fact that a sufficient condition for that fact that two polynomials in k[x] share a common root is the vanishing of the determinant of a certain matrix called the Sylvester matrix (easy to build from the polynomial coefficients) This result can be used to solve two polynomial equations involving two variables by considering one of the variables to be a parameter and solving the polynomial equation obtained by equating the resulting Sylvester determinant to zero This can be applied recursively to solve any multivariable equation

François Glineur, Polynomial solving & minimization - 29 - First Prev Next Last Full Screen Quit Sum of square methods In the multivariable case, sum of squares methods only provide relaxations since there exists nonnegative polynomials than cannot be expressed as a sum of squares: x 4 y 2 + x 2 y 4 + z 6 3x 2 y 2 z 2 (Motzkin) Use some more theory, it is however possible to build a sequence of semidefinite relaxations of increasing size whose optima converge to the global optimum These methods also provide a proof for the lower bound they compute Practical computational efficiency (using the first few relaxations) seems promising (GloptiPoly, D. Henrion)

Concluding remarks Many different methods for the single variable case. Optimal methods not always practical. Convex reformulation is not competitive. Why is it more efficient to concentrate on the stationarity conditions rather than optimizing directly? Groebner bases techniques are powerful but their worstcase complexity is awful and practical performance sometimes too slow Wanted: serious numerical experiments comparing algebraic geometric and linear algebraic approaches François Glineur, Polynomial solving & minimization - 30 - First Prev Next Last Full Screen Quit