Optimization of a Convex Program with a Polynomial Perturbation

Similar documents
The Lovasz-Vempala algorithm for computing the volume of a convex body in O*(n^4) - Theory and Implementation

Lecture 16: October 29

On the complexity of approximate multivariate integration

Solving Convex Programs by Random Walks

Baum s Algorithm Learns Intersections of Halfspaces with respect to Log-Concave Distributions

The centroid body: algorithms and statistical estimation for heavy-tailed distributions

Sampling Contingency Tables

Constrained Optimization and Lagrangian Duality

Lecture 2: Ellipsoid Method and Reductions Between Convex Oracles

Linear Algebra Massoud Malek

FRACTIONAL PACKING OF T-JOINS. 1. Introduction

Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007

A Bernstein-Chernoff deviation inequality, and geometric properties of random families of operators

Simulated annealing in convex bodies and an O (n 4 ) volume algorithm

LATTICE POINT COVERINGS

The Split Closure of a Strictly Convex Body

c 2000 Society for Industrial and Applied Mathematics

On the monotonicity of the expected volume of a random. Luis Rademacher Computer Science and Engineering The Ohio State University

Unions of perfect matchings in cubic graphs

Some Background Math Notes on Limsups, Sets, and Convexity

Szemerédi s Lemma for the Analyst

The Kannan-Lovász-Simonovits Conjecture

Convexity and constructive infima

Analysis III. Exam 1

1 Strict local optimality in unconstrained optimization

Geometric Random Walks: A Survey

Unions of perfect matchings in cubic graphs

Convex hull of two quadratic or a conic quadratic and a quadratic inequality

Optimal global rates of convergence for interpolation problems with random design

Compact orbit spaces in Hilbert spaces and limits of edge-colouring models

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

SELECTIVELY BALANCING UNIT VECTORS AART BLOKHUIS AND HAO CHEN

Polynomiality of Linear Programming

Geometric Random Walks: A Survey

1 Introduction Semidenite programming (SDP) has been an active research area following the seminal work of Nesterov and Nemirovski [9] see also Alizad

GAUSSIAN MEASURE OF SECTIONS OF DILATES AND TRANSLATIONS OF CONVEX BODIES. 2π) n

Bulletin of the. Iranian Mathematical Society

A Robust APTAS for the Classical Bin Packing Problem

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

PROJECTIVE PRE-CONDITIONERS FOR IMPROVING THE BEHAVIOR OF A HOMOGENEOUS CONIC LINEAR SYSTEM

On the cells in a stationary Poisson hyperplane mosaic

Dual methods for the minimization of the total variation

47-831: Advanced Integer Programming Lecturer: Amitabh Basu Lecture 2 Date: 03/18/2010

EFFICIENT HIGH-DIMENSIONAL SAMPLING AND INTEGRATION. A Dissertation Presented to The Academic Faculty. Ben Cousins

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

APPROXIMATE ISOMETRIES ON FINITE-DIMENSIONAL NORMED SPACES

Lecture and notes by: Daniel Dadush and Alejandro Toriello November 20, Volume Computation Using Geometric Random Walks

EXISTENCE OF SOLUTIONS TO ASYMPTOTICALLY PERIODIC SCHRÖDINGER EQUATIONS

PYTHAGOREAN PARAMETERS AND NORMAL STRUCTURE IN BANACH SPACES

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Comparison of Lasserre s measure based bounds for polynomial optimization to bounds obtained by simulated annealing

Dimensional behaviour of entropy and information

Topics in Stochastic Geometry. Lecture 4 The Boolean model

Convergence rates of moment-sum-of-squares hierarchies for volume approximation of semialgebraic sets

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

CS 435, 2018 Lecture 3, Date: 8 March 2018 Instructor: Nisheeth Vishnoi. Gradient Descent

Optimization Tutorial 1. Basic Gradient Descent

EE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1

Dispersion of Mass and the Complexity of Geometric Problems. Luis Alexis Rademacher

8 Numerical methods for unconstrained problems

Model Counting for Logical Theories

LINEAR PROGRAMMING III

5 Handling Constraints

Szemerédi s Lemma for the Analyst

Approximating a Convex Body by An Ellipsoid

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

26.1 Metropolis method

Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures

Jeong-Hyun Kang Department of Mathematics, University of West Georgia, Carrollton, GA

The Knaster problem and the geometry of high-dimensional cubes

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Star bodies with completely symmetric sections

Rate of convergence of geometric symmetrizations

A Proximal Method for Identifying Active Manifolds

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers.

Richard F. Bass Krzysztof Burdzy University of Washington

PROJECTIVE PRE-CONDITIONERS FOR IMPROVING THE BEHAVIOR OF A HOMOGENEOUS CONIC LINEAR SYSTEM

Helly's Theorem and its Equivalences via Convex Analysis

CONVERGENCE OF APPROXIMATING FIXED POINTS FOR MULTIVALUED NONSELF-MAPPINGS IN BANACH SPACES. Jong Soo Jung. 1. Introduction

c 2006 Society for Industrial and Applied Mathematics

Hit-and-Run from a Corner

arxiv: v1 [math.oc] 21 Mar 2015

CHAPTER 1. Metric Spaces. 1. Definition and examples

15-850: Advanced Algorithms CMU, Spring 2017 Lecture #17: The Ellipsoid Algorithm March 3, 2017

On the convergence of normal forms for analytic control systems

On a functional equation connected with bi-linear mappings and its Hyers-Ulam stability

Convex Functions and Optimization

Approximating sparse binary matrices in the cut-norm

A note on : A Superior Representation Method for Piecewise Linear Functions

Topology. Xiaolong Han. Department of Mathematics, California State University, Northridge, CA 91330, USA address:

The Split Closure of a Strictly Convex Body

MAXIMAL PERIODS OF (EHRHART) QUASI-POLYNOMIALS

Whitney s Extension Problem for C m

arxiv: v2 [math-ph] 24 Feb 2016

New Conjectures in the Geometry of Numbers

Interior Point Methods for Mathematical Programming

Some Sieving Algorithms for Lattice Problems

Notes on the decomposition result of Karlin et al. [2] for the hierarchy of Lasserre by M. Laurent, December 13, 2012

Numerical Optimization

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Transcription:

Optimization of a Convex Program with a Polynomial Perturbation Ravi Kannan Microsoft Research Bangalore Luis Rademacher College of Computing Georgia Institute of Technology Abstract We consider the problem of minimizing a convex function plus a polynomial p over a convex body K. We give an algorithm that outputs a solution x whose value is within ε range K (p) of the optimum value, where range K (p) = sup x K p(x) inf x K p(x). When p depends only on a constant number of variables, the algorithm runs in time polynomial in 1/ε, the degree of p, the time to round K and the time to solve the convex program that results by setting p = 0. Keywords: nonlinear programming, approximation algorithms, polynomial optimization 1 Introduction We give an algorithm to minimize approximately f(x)+p(x) over a convex body K in R n where f is any convex function and p is any polynomial in a constant number of variables. Our solution x satisfies f(x) + p(x) f(x ) + p(x ) + ε range K (p) where ε is a given error parameter, x is an optimum solution and The algorithm runs in time range K (p) = sup p(x) inf p(x). x K x K ( ( )) kd 2 k O T (f, K) + T (K) ε Corresponding author. Email: lrademac@cc.gatech.edu. Address: 266 Ferst Drive, Atlanta GA 30332, USA. 1

where k is the number of variables that appear in p, d is the degree of p, T (f, K) is the time to solve the convex program min x K f(x) and T (K) is the time to put K in near-isotropic position (discussed in Section 2). In situations where p is a small perturbation, the range of p is small and hence the error the algorithms makes. Also clearly the algorithm generalizes traditional convex optimization. This paper is inspired by a result of Vavasis [7] for the case when p is quadratic with slightly different error bounds. 2 Preliminaries For a bounded set C R n, let w C denote the width of C, that is w C = inf (sup h S n 1 x C h T x inf x C ht x). Let B n denote the n-dimensional Euclidean unit ball. The volume of B n is π n/2 Γ( n 2 + 1). A convex body K is said to be in isotropic position if its center of gravity is the origin and for any unit vector v, one has 1 (v x) 2 = 1. vol(k) K It is known that for any convex body there is an affine transformation which puts the body in isotropic position. We will now quote definitions and results from [2] about the existence of an efficient randomized algorithm that puts a convex body in near isotropic position. For a convex body K, let b(k) denote its center of gravity. Definition 1 (2.4 in [2]). We say that a convex body K is in θ-nearly isotropic position (0 < θ 1), if b(k) θ, and for every vector v R n, (1 θ) v 2 1 vol(k) K b(k) (v T x) 2 dx (1 + θ) v 2 Theorem 2 (Corollary 5.2 in [2]). Let θ < 1/2. If K is in θ-near isotropic position, then (1 2θ)B n K (1 + 2θ)(n + 1)B n. The following result is a specialization of Theorem 2.5 from [2]. There are some mistakes in the statement of that theorem, but a correct statement can be obtained by looking at the results on which it depends (Lemma 5.18 and Theorem 5.20 from [2]) 2

Theorem 3. There exists a randomized algorithm that, when given numbers 0 < r R and access to a membership oracle of a convex body K R n with rb n K RB n, finds an affine transformation A for which AK is in 1/4- nearly isotropic position with probability at least 1 δ. The number of oracle calls is ( ( O n 5 log R ) ) 3. r Here O means that logarithmic factors in n and 1/δ are ignored, where δ is the probability of failure. The improved random walks in [4] imply that one can approximately round a convex body in time O (n 4 log c (R/r)) for some constant c, but we use the results from [2] because they are in a form that is slightly closer to what we need. 3 Optimizing a Convex Program with a Polynomial Perturbation Formally, for a convex body K R n, a convex function f : R n R and a polynomial p that depends on only k variables, let problem (P ) be: min g(x) = f(x) + p(x) subject to x K (P ) The algorithm that we propose to solve (P ) is essentially an enumeration over a suitable grid in the space of the variables that appear in the polynomial. It is possible to guarantee a certain quality of approximation because of a known bound on the gradient of a polynomial in a convex set as a function of the range of the polynomial (Theorem 5). Let proj : R n R k be the orthogonal projection onto the subspace of the variables that appear in p. Consider a covering C of 3 2 (k + 1)B k with cubes having side s = π ε 8d 2 k. (1) The algorithm will put proj(k) in 1/4-nearly isotropic position by an affine transformation A (see Section 2), which implies that A proj(k) 3 2 (k + 1)B k, and C is also a covering of A proj(k). We will take C to be the set of all cubes with centers in sz k that intersect 3 2 (k + 1)B k. These cubes are contained in ( 3 2 (k + 1) + s k)b k, and thus C ( O( ) k k) (2) s (using that the volume of each cube is s k and the volume of the ball given in Section 2). 3

The algorithm uses the following piecewise linear approximation of p: For every cube C with center y C, define Algorithm 1 φ C (x) = p(a 1 y C ) + p(a 1 y C ) T (x A 1 y C ). 1. Compute an affine transformation A : R k R k such that A proj(k) is in 1/4-nearly isotropic position. 2. For every cube C C, compute an optimal solution x C to min f(x) + φ C (x). x proj 1 A 1 (C) K If this problem is infeasible, remove C from C. 3. Let C = argmin C C f(x C ) + p(x C ). Output x C. Let T (f, K) be the time to solve the convex program obtained from P when p = 0. This can be done in polynomial time with mild assumptions on K and f and their representation [1, 3]. Let T (K) be the time to put the projection of K in 1/4-nearly isotropic position. If K is given by a membership oracle and numbers 0 < r R so that rb n K RB n, then an efficient algorithm is guaranteed by Theorem 3. Theorem 4. Let x be an optimum solution of Problem P. Algorithm 1 outputs x K satisfying f(x) + p(x) f(x ) + p(x ) + ε range K (p) in time ( ( )) kd 2 k O T (f, K) + T (K). ε Proof. By means of Theorem 2 we have that B k /2 A proj(k), which implies that the width of A proj(k) is at least 1/2. With this, we will now analyze the error that the algorithm makes when replacing p by φ C. We apply Corollary 6 and Taylor s theorem on the multivariate polynomial q : R k R, q(x) = p(a 1 x) satisfying range A proj(k) (q) = range K (p) to get for any cube C and for all x proj 1 A 1 (C) K: φ C (x) p(x) 1 2 sup h T 2 q(az)h A proj(x) y C 2 z K,h S n 1 16 d 4 π 2 wa 2 range A proj(k) (q)s 2 k proj(k) 64 π 2 d4 range K (p)s 2 k ε range K (p). 4

This implies g(x C ) OP T + ε range K (p). The bound on the running time follows from Equations (1) and (2). The following result is a slight reformulation of a result by Skalyga [6] (who proved it with constant 4/π, it was proved before with constant 2 by Wilhelmsen [8], [5, Section 6.1.5]). This is a multivariate generalization of a univariate inequality by A. A. Markov [5, Chapter 6, Theorem 1.2.1] Theorem 5. Let K R n be a convex body. Let p : K R be a multivariate polynomial of degree d. Then for any x K: p(x) 4 π d 2 range K (p) w K. If we just used Theorem 5, we could only use a constant approximation in the role of our linear approximation φ C to p and the dependence on ε of our algorithm would be 1/ε. But one can easily use Theorem 5 inductively to get a second order version of it, and be able to use a linear approximation φ C : Corollary 6. Let K R n be a convex body. Let p : K R be a multivariate polynomial of degree d. Then for any x K and h S n 1 : h T 2 p(x)h 32 d 4 range K (p) π 2. Proof. Use Theorem 5 twice on the polynomials x p(x) T h and p to get: w 2 K h T 2 p(x)h 4 d 2 range K ( p( ) T h) π w K 8 d 2 sup x K p(x) T h π w K 32 d 4 range K (p) π 2 wk 2 4 Discussion Vavasis s paper is crucially based on a result similar to Theorem 5 for quadratic polynomials. Note that the above theorem essentially asserts that a degree d polynomial cannot be wild at any place. The only factor that can make its gradient too large is the width which we bound using isotropic position. An interesting open problem is to extend the results to a larger class of perturbations of convex minimization problems. In effect any function satisfying Theorem 5 type of conclusion would be amenable to this. In a sense with Theorem 5 and efficient methods to find near-isotropic position on hand, the current paper can be viewed as just using them together. But given that polynomials are very general and not too many clean generalizations of convex optimization are known, this records one such while raising the question of possibly others. 5

Acknowledgments We would like to thank an anonymous referee for helpful comments and suggestions. References [1] M. Grötschel, L. Lovász, and A. Schrijver. Geometric algorithms and combinatorial optimization, volume 2 of Algorithms and Combinatorics: Study and Research Texts. Springer-Verlag, Berlin, 1988. [2] R. Kannan, L. Lovász, and M. Simonovits. Random walks and an O (n 5 ) volume algorithm for convex bodies. Random Structures Algorithms, 11(1):1 50, 1997. [3] L. Lovász and S. Vempala. Fast algorithms for logconcave functions: Sampling, rounding, integration and optimization. Proceedings of the 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS 06), pages 57 68, 2006. [4] L. Lovász and S. Vempala. Simulated annealing in convex bodies and an O (n 4 ) volume algorithm. Journal of Computer and System Sciences, 72(2):392 417, 2006. [5] G. V. Milovanović, D. S. Mitrinović, and T. M. Rassias. Topics in polynomials: extremal problems, inequalities, zeros. World Scientific Publishing Co. Inc., River Edge, NJ, 1994. [6] V. I. Skalyga. Estimates of the derivatives of polynomials on convex bodies. Proc. Steklov Inst. Math, 218:372 383, 1997. [7] S. A. Vavasis. Approximation algorithms for indefinite quadratic programming. Math. Program., 57:279 311, 1992. [8] D. R. Wilhelmsen. A Markov inequality in several dimensions. J. Approximation Theory, 11:216 220, 1974. 6