The master equality polyhedron with multiple rows

Similar documents
The master equality polyhedron with multiple rows

MEP123: Master Equality Polyhedron with one, two or three rows

On mixed-integer sets with two integer variables

Cutting planes from extended LP formulations

The continuous knapsack set

The Master Equality Polyhedron: Two-Slope Facets and Separation Algorithm

On the Relative Strength of Split, Triangle and Quadrilateral Cuts

The continuous knapsack set

A Lower Bound on the Split Rank of Intersection Cuts

The Strength of Multi-row Aggregation Cuts for Sign-pattern Integer Programs

Split Rank of Triangle and Quadrilateral Inequalities

3. Linear Programming and Polyhedral Combinatorics

3. Linear Programming and Polyhedral Combinatorics

A geometric perspective on lifting

Polyhedral Approach to Integer Linear Programming. Tepper School of Business Carnegie Mellon University, Pittsburgh

Mixing Inequalities and Maximal Lattice-Free Triangles

Lattice closures of polyhedra

Chapter 1. Preliminaries

A geometric perspective on lifting

On the Relative Strength of Split, Triangle and Quadrilateral Cuts

Computational Experiments with Cross and Crooked Cross Cuts

Linear Programming Redux

On Some Polytopes Contained in the 0,1 Hypercube that Have a Small Chvátal Rank

n-step mingling inequalities: new facets for the mixed-integer knapsack set

Cutting planes from two rows of a simplex tableau

Some Relationships between Disjunctive Cuts and Cuts based on S-free Convex Sets

Integer Programming ISE 418. Lecture 12. Dr. Ted Ralphs

On the Relative Strength of Split, Triangle and Quadrilateral Cuts

March 2002, December Introduction. We investigate the facial structure of the convex hull of the mixed integer knapsack set

Computational Experiments with Cross and Crooked Cross Cuts

Multi-Row Cuts in Integer Programming. Tepper School of Business Carnegie Mellon University, Pittsburgh

Lift-and-Project Inequalities

Corner Polyhedron and Intersection Cuts

Minimal inequalities for an infinite relaxation of integer programs

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

Lattice closures of polyhedra

On the knapsack closure of 0-1 Integer Linear Programs

Copositive Plus Matrices

A Review of Linear Programming

LP Relaxations of Mixed Integer Programs

The Split Closure of a Strictly Convex Body

Minimal Valid Inequalities for Integer Constraints

THE EXISTENCE AND USEFULNESS OF EQUALITY CUTS IN THE MULTI-DEMAND MULTIDIMENSIONAL KNAPSACK PROBLEM LEVI DELISSA. B.S., Kansas State University, 2014

MAT-INF4110/MAT-INF9110 Mathematical optimization

CS 6820 Fall 2014 Lectures, October 3-20, 2014

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

THE MIXING SET WITH FLOWS

Lifting Integer Variables in Minimal Inequalities Corresponding To Lattice-Free Triangles

Lifted Inequalities for 0 1 Mixed-Integer Bilinear Covering Sets

Spring 2017 CO 250 Course Notes TABLE OF CONTENTS. richardwu.ca. CO 250 Course Notes. Introduction to Optimization

Approximation of Minimal Functions by Extreme Functions

3.7 Cutting plane methods

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Minimal inequalities for an infinite relaxation of integer programs

Asteroide Santana, Santanu S. Dey. December 4, School of Industrial and Systems Engineering, Georgia Institute of Technology

On Sublinear Inequalities for Mixed Integer Conic Programs

Key words. Integer nonlinear programming, Cutting planes, Maximal lattice-free convex sets

3.8 Strong valid inequalities

3.7 Strong valid inequalities for structured ILP problems

CO 250 Final Exam Guide

3.10 Lagrangian relaxation

Graph Coloring Inequalities from All-different Systems

Multicommodity Flows and Column Generation

Lagrangian Relaxation in MIP

Computational Experiments with Cross and Crooked Cross Cuts

Gaussian integers. 1 = a 2 + b 2 = c 2 + d 2.

The Triangle Closure is a Polyhedron

Change in the State of the Art of (Mixed) Integer Programming

IBM Research Report. Lattice-Free Sets, Branching Disjunctions, and Mixed-Integer Programming

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

The Simplex Algorithm

Critical Reading of Optimization Methods for Logical Inference [1]

Cut-Generating Functions for Integer Variables

1 Maximal Lattice-free Convex Sets

Lecture #21. c T x Ax b. maximize subject to

Combinatorial Optimization

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method

Notes taken by Graham Taylor. January 22, 2005

Linear Algebra Review: Linear Independence. IE418 Integer Programming. Linear Algebra Review: Subspaces. Linear Algebra Review: Affine Independence

Duality of LPs and Applications

Sparse Optimization Lecture: Dual Certificate in l 1 Minimization

The Split Closure of a Strictly Convex Body

On the relative strength of families of intersection cuts arising from pairs of tableau constraints in mixed integer programs

MIP reformulations of some chance-constrained mathematical programs

TRISTRAM BOGART AND REKHA R. THOMAS

A packing integer program arising in two-layer network design

Cut-Generating Functions for Integer Linear Programming

Carathéodory Bounds for Integer Cones

Integer Programming, Part 1

arxiv:math/ v1 [math.ac] 11 Nov 2005

7. Lecture notes on the ellipsoid algorithm

A New Subadditive Approach to Integer Programming: Theory and Algorithms

MIR Closures of Polyhedral Sets

Chapter 3. Vector spaces

A Parametric Simplex Algorithm for Linear Vector Optimization Problems

Lecture 6 - Convex Sets

Lecture 9: Dantzig-Wolfe Decomposition

Optimization methods NOPT048

Some Properties of Convex Hulls of Integer Points Contained in General Convex Sets

Valid inequalities based on simple mixed-integer sets

Transcription:

The master equality polyhedron with multiple rows Sanjeeb Dash IBM Research sanjeebd@us.ibm.com Ricardo Fukasawa University of Waterloo rfukasaw@math.uwaterloo.ca September 16, 2010 Oktay Günlük IBM Research gunluk@us.ibm.com Abstract The master equality polyhedron (MEP) is a canonical set that generalizes the Master Cyclic Group Polyhedron (MCGP) of Gomory. We recently characterized a nontrivial polar for the MEP, i.e., a polyhedron T such that an inequality defines a nontrivial facet of the MEP if and only if its coefficient vector forms a vertex of T. In this paper we study the MEP when it is defined by m > 1 rows. We define the notion of a polaroid, a set containing all nontrivial facet defining inequalities. We show how to use linear programming (LP) to efficiently solve the separation problem for the MEP when the polaroid has a compact polyhedral description. We obtain such descriptions via subadditivity conditions when m = 2 or m = 3 and, using LP duality, show how to efficiently optimize over the MEP. These results yield a pseudo-polynomial time LP-based algorithm to solve the problem min{cx : Ax = b, x 0, x Z n } when A has at most 3 constraints. For the MCGP and the MEP defined by a single constraint, the notions of two-term subadditivity and valid inequalities for MEP are essentially equivalent. We show this is not true in the case of the MEP when m 3; In fact, we prove that subadditivity conditions with a sub-exponential number of terms do not imply validity. In particular, when m = 3, we show that four-term subadditivity conditions are necessary and sufficient for validity. 1 Introduction In [12], Gomory defined master polyhedra, and characterized the convex hull of their nontrivial facets, i.e., facets which do not have the form x i 0 for some variable x i. He used these facets to define valid inequalities for general integer programs. Gomory and Johnson [13] later studied a mixed-integer extension of the master cyclic group polyhedron (MCGP) a special case of the master polyhedron when it is defined by a single constraint and also studied an infinite group variant of the MCGP. Over the last decade, the MCGP has been studied in great detail; see, for example, [3, 8, 17, 20] for some facet classes, and [14, 11] for properties of the infinite group variant. Recently, many authors have studied polyhedral properties of sets defined by two or more rows. Andersen, Louveaux, Weismantel, and Wolsey [1] derive valid inequalities for {x R t : Ax b(mod 1), x 0} where A has two rows, 1 is a vector with all components 1 and a vector 1

equals another modulo a third vector if and only if the component-wise modular equations hold. Cornuéjols and Margot [6] study the infinite group version of the two-constraint problem above, while Borozan and Cornuéjols [4] study the m-constraint version, where m 2. Dey and Richard [9, 10] study the infinite group variant of the master polyhedron defined by two constraints. In [7], we defined the master equality polyhedron (MEP), a generalization of the MCGP, and also of the master knapsack polyhedron of Araóz [2], and described a nontrivial polar for the MEP, i.e., a polyhedron T with the property that an inequality defines a nontrivial facet of MEP if and only if its coefficient vector defines a vertex of T. In this paper we study the m-row extension (for m 2) of the MEP, which can be viewed as a generalization of the m-row master polyhedron. Let n be a strictly positive vector in Z m for some m 1. Let I + be the set of all non-zero integer vectors with kth component contained in [0, n k ) for k = 1,..., m. Then for r I +, the m-row MCGP, in its most general form, equals conv {x Z I+ : } ix i r(mod n), x 0. i I + Let I be the set of vectors in Z m with kth component contained in [ n k, n k ]. If the values of m and n are not clear from the context, we refer to I as I m (n). For a nonzero integer vector r such that n r 0, we define the (m-row) MEP as { K m (n, r) = conv x Z I : } ix i = r, x 0. (1) i I where m is a positive integer. Note that we require r 0 without loss of generality because if r 0, one can simply multiply some rows by 1 and rename the variables afterwards to obtain a nonnegative right-hand-side. Let e j stand for the unit vector in R m with a 1 in the jth component and 0 in the remaining components, and let I N polyhedron is = { n d e d } m d=1 ; then an equivalent definition of the m-row master P m (n, r) = conv {x Z I+ I N : } ix i + ix i = r, x 0. (2) i I + i I N It is clear that P m (n, r) defines a face of K m (n, r) (fix x i to 0 for i I \ (I + I N )). In the rest of the paper, we assume the vector n in K m (n, r) has equal-valued components; our results proved under this restriction hold under the more general case. Therefore, the parameter n in K m (n, r) can be considered to be a number. Gomory characterized the convex hull of nontrivial facets of P m (n, r) as follows. Theorem 1.1 [12] The inequality πx 1 defines a nontrivial facet of P m (n, r) if and only if π R I+ is an extreme point of π i + π j π (i+j) (mod n) i, j I +, (SA) Q m π i + π j = π r i, j I + such that r (i + j) (mod n), (CO) (n, r) = π j 0 j I +, π r = 1. 2

We call the property (SA) two-term subadditivity and the property (CO) complementarity. Twoterm subadditivity is a central concept in the study of the master polyhedron. It is essentially equivalent to validity: if π satisfies (SA), then πx π r is a valid inequality for the master polyhedron, and any minimal valid inequality for the master polyhedron satisfies (SA). In [7], we described a nontrivial polar for K 1 (n, r) and proved a result similar to Theorem 1.1. Theorem 1.2 [7] The inequality πx 1 defines a nontrivial facet of K 1 (n, r) for n r > 0 if and only if π R 2n+1 is an extreme point of π i + π j π i+j, i, j I, 1 i + j n (RSA1) π i + π j + π k π i+j+k, i { n,..., 1}, j, k, i + j + k {1,..., n} (RSA2) T 1 (n, r) = π i + π j = π r, i, j I, i + j = r (RCO) π r = 1, (NC1) π n = π 0 = 0. (NC2) Theorem 1.2 implies that given a point x K 1 (n, r), a violated facet can be found in polynomial time by finding an extreme point solution of min{π x : π T 1 (n, r)}. Notice that any set Q = conv{x Z t : t i=1 a ix i = b, x 0} defined by a single constraint with integral coefficients can be viewed as a face of K 1 (n, r) where r = b and n is taken larger than all a i and b. This result yields a pseudo-polynomial time algorithm to solve the separation problem for the set Q. In [7] we also show that two-term subadditivity conditions imply validity for K 1 (n, r) as they do for P m (n, r). In this paper we show that two-term subadditivity conditions imply validity for K 2 (n, r) as well. However, two and three-term subadditivity conditions do not imply validity for K 3 (n, r), but four-term subadditivity conditions are sufficient for validity. Further, we show that subadditivity conditions with exponentially many terms are necessary for validity for K m (n, r) for m 3. Therefore, in a fundamental sense, K m (n, r) is a more complicated set than P m (n, r). Our second main result is a polynomial-time separation algorithm for K 2 (n, r) and K 3 (n, r) based on the notion of a polaroid. This is a polyhedral set, similar to a nontrivial polar, containing the coefficient vectors of all nontrivial facets. We present compact descriptions of polaroids for K 2 (n, r) and K 3 (n, r) and show that given a point not contained in K m (n, r) when m = 2 or 3, a violated facet can be obtained by optimizing two linear functions over the corresponding polaroid. Additionally, using the dual of the constraints defining the polaroid, we show how to optimize a linear function over K 2 (n, r) and K 3 (n, r). Furthermore, this implies that one can solve the problem min{cx : Ax = b, x 0, x Z n } when A has at most 3 constraints in pseudo-polynomial time by solving a single LP. The paper is organized as follows. In Section 2, we discuss the difference between K m (n, r) and P m (n, r) as a source of cutting planes for integer programs and argue why it is hard to generalize Theorem 1.2 to K m (n, r) when m > 1. In Section 3, we describe polaroids, and show how to use them to solve the associated separation problem. In Section 4, we give polaroids for m = 2 and m = 3, and in Section 5, we study the relationship between validity and subadditivity for K m (n, r). In Section 6 we show that the size of certain natural polaroids grow exponentially with m. 3

2 Preliminaries 2.1 Group relaxation of MEP As discussed earlier, P m (n, r) is a lower dimensional face of K m (n, r). In addition, it is possible to map points in K m (n, r) to points in P m (n, r) using a simple mapping which also gives a relaxation of K m (n, r) which we call the group relaxation. For simplicity, we next describe the group relaxation for m = 1 and present an example that shows that this relaxation can be arbitrarily bad. It is easy to extend the description and example to m > 1. Let n, r Z + \ {0} be given and let P = P 1 (n, r) and K = K 1 (n, r). Furthermore, let K C denote the continuous relaxation of K and let { n 1 } P = conv y Z n + : ny n + iy i = r i=1 = { } y R n + : Ay b where Ay b gives a linear description of P. Note that for x K, we have n i= n ix i = n 1 i=1 i(x i + x i n ) n( n i=1 x i x n ) = r and therefore x K y P where { n j=1 y i = x j x n i = n x i + x i n i = 1,..., n 1. Based on this observation, it is possible to obtain the group relaxation K G of K as the collection of points in K C whose image satisfies Ay b. More precisely, K G = { x R 2n+1 + : n i= n y n = ix i = r, y R n +, Ay b, n i=1 } x i x n, y i = x i + x i n i = 1,..., n 1. Clearly, K C K G K. We next give an example that shows that when optimizing a linear function, the group relaxation can be very weak. Example 2.1 Consider minimizing cx over K 1 (15, 3) where c i is 1 if i is odd, and 0 otherwise. Let z denote the optimal value of this problem. Clearly any integer feasible solution has to have x i 1 for some i odd, and therefore z 1. Furthermore, as x K K 1 (15, 3) where x K i i = 3 and 0 otherwise, we have z = 1. is 1 if Let K C and K G denote the continuous and group relaxation of K 1 (15, 3) and let z C and z G be the optimal value of the corresponding relaxed problem. Clearly, z C, z G 0 as both c and x are nonnegative. Now consider x where x i is 1/2 if i = 8, 10, 12 and 0 otherwise. Clearly, c x = 0. As n i= n i x i = r, we have x K C and z C = 0. Furthermore, defining ȳ 15 = 15 i=1 x i x 15 and ȳ i = x i + x i 15 for i = 1,..., 14, it is not hard to see that ȳ = 1/2(y 1 + y 2 ) where { { yi 1 1 i = 3 = y 2 1 i = 8, 10, 15 i = 0 otherwise, 0 otherwise, 4

Furthermore, notice that y 1, y 2 P 1 (15, 3) and therefore ȳ P 1 (15, 3) implying Aȳ b. Therefore, x K G and z G = 0. 2.2 Basic properties of K m (n, r) and its facets We now define some basic notation that will be used throughout the paper. As defined previously, e d R m + will denote the dth unit vector. Given a point i I, let f i R I + denote the unit vector with a 1 in the i-th component and 0 everywhere else. Finally, let I r = {i I : r i I} and let I N = { ne d } m d=1 I. These sets will be used later on to represent, respectively, vectors involved in complementarity equations and vectors involved in normalization equations. We next make some basic observations regarding the polyhedral structure of K m (n, r). These observations are extensions of the results in [7]. We start with the dimension of K m (n, r) R I. Lemma 2.2 The dimension of K m (n, r) is I m. In other words, there are no extra implied equations. Proof. All points in K m (n, r) satisfy the m linearly independent equations that define K m (n, r). Therefore, the dimension is at most I m. Now suppose αx = β holds for all x K m (n, r). Then, we can assume that α ned = 0 for all d = 1,..., m. Now, since f r K m (n, r), we have that α r = β. Moreover, since nf ed + f r + f ned K m (n, r), then nα ed + α r + α ned = β α ed = 0. Similarly, f ed + f ed + f r K m (n, r) so α ed = 0. Finally, for any point i I, f i + d:i d <0 i d f ed + d:i d 0 i df ed + f r K m (n, r), therefore α i + d:i d <0 i d α ed + d:i d 0 i dα ed + α r = β α i = 0. Therefore, α = 0 and β = 0. By using a very similar proof (which we omit for the sake of brevity), we can also show the following result: Lemma 2.3 If n 2, then x i 0 defines a facet of K m (n, r) for all i I. We call any facet-defining inequality of K m (n, r) which is not a nonnegativity constraint nontrivial. We next show some basic properties of these inequalities. Lemma 2.4 The following conditions are satisfied by any nontrivial facet-defining inequality πx η of K m (n, r): π i + π r i = π r, i I r, π r = η, In addition, for all q Z I + such that q := i I q i i I q i π i π q. i I (RCO) (NC1) (KSA) 5

Proof. Let q Z I + and t I such that i I q i i = t. Since πx η is a nontrivial facet-defining inequality of K m (n, r), there is a point x K m (n, r) such that πx = η and x t > 0. But then define x = x f t + i I q if i. We have that η πx = πx π t + i I q iπ i and (KSA) follows. As f r K m (n, r), we have π r η. If π r η, let x K m (n, r) be such that such that π x = η and x r > 0. Let p = x + x f r and note that p 0 and therefore p K m (n, r). As π r > η, we have πp = 2η π r < η even though p K m (n, r), a contradiction. Therefore, π r = η and (NC1) holds. To prove (RCO), let x and x be points in K m (n, r) such that πx = πx = η, x i > 0 and x r i > 0. Then x = x +x f i f r i K m (n, r) η πx = πx +πx π i π r i π i +π r i η. But then, (KSA) implies (RCO). Note that for i = 0, inequality (RCO) implies that π 0 = 0 for all nontrivial facet-defining inequalities. We will call inequalities (KSA) as k-term subadditive inequalities whenever q 1 k. As K m (n, r) is not a full-dimensional polyhedron, any valid inequality πx η for K m (n, r) has an equivalent representation with π i = 0 for all i I N. If a valid inequality does not satisfy π i = 0 for some i I N, one can add an appropriate multiple of the equation i I i dx i = r d to it. We next show that after this normalization, all facet-defining nontrivial inequalities have a strictly positive right-hand-side. Lemma 2.5 Let πx η be a nontrivial facet-defining inequality of K m (n, r) that satisfies π i = 0 for all i I N. Then η > 0. Proof. For any i I such that i 0 and i 0, we have that n i + m d=1 i d( ne d ) = 0 so (KSA) implies that nπ i + m d=1 i dπ ned π 0 π i 0. Therefore, η 0 as η = π r. Now assume, for the sake of contradiction that η = 0. By (NC1) we have that π r = 0. We also assume, without loss of generality, that there is a number m 1 such that r d > 0 for all d D + = {1,..., m } and r d = 0 for all d D o = {m + 1,..., m}. For all 0 i r, (RCO) implies that 0 π i +π r i = π r = 0 and therefore π i = 0. In particular, π ed = 0 for all d D +. Furthermore, by inequality (KSA) nπ ed π ned = 0, and thus π ed 0 for all d = 1,..., m. In addition, for d D +,we have π ned + (n 1)π ed π ed implying π ed = 0. So far we have shown that π ed = π ed = 0 for all d D +. For d D o, on the other hand, π ned + π r+ned = π r and therefore π r+ned = 0. Now pick any point i I and let S i = {d : i d < 0} and define i = i + n d S i e d = i + d S i (r + ne d ) + S i d D + r d ( e d ). Note that the last term above is indeed S i r as r d = 0 for d D +. Clearly, 0 i I and therefore π i 0. Therefore, (KSA) implies that π i + (π r + nπ ed ) + π r+ned + S i r d π ed d D + S i d D o S i d D + π i 0. Therefore π i 0 for all i I. 6

But then πx η is implied by all nonnegativity inequalities x 0, which contradicts the fact that it is a facet-defining inequality. Thus η > 0. Therefore, without loss of generality, we can assume that if πx η is a nontrivial facet-defining inequality then it satisfies the following normalization conditions: π i = 0, i I N. (NC2) η = 1. (NC3) Throughout the rest of the paper we assume that all nontrivial facet-defining inequalities satisfy equations (NC3) and (NC2). 2.3 Extending the result from K 1 (n, r) An important property of T 1 (n, r) that we use in the proof of Theorem 1.2 is that it contains all unit vectors f i for i / I r I N as rays. Using this property, we can argue that extreme points of T 1 (n, r) are in one-to-one correspondence with nontrivial facet-defining inequalities of K 1 (n, r). Notice that the conditions that guarantee that all π T 1 (n, r) yield valid inequalities are the relaxed subadditivity conditions (RSA1) and (RSA2) which have the form q i π i π q, q Q (3) i I where q denotes i I q i i and Q is a given subset of Z I +. In our description all q Q satisfy q 1 3 and therefore Q and consequently the number of inequalities defining T 1 (n, r) is small. Moreover, as we wanted to have f i as rays for i / I r I N, we excluded from Q points q that have i I q i i I r I N. A natural generalization of this idea for K m (n, r) would require a set Q m Z Im (n) + with the following properties: (i) if π satisfies subadditivity condition (3) for all q Q m, then πx 1 is valid for K m (n, r), and (ii) all q Q m have small norm and i I m (n) q i i I r I N. We next present an example that shows that such a set does not exist for m 2 because for the first condition to hold, Q m must contain an element q that has q 1 = Ω(n). Example 2.6 (Swirszcz [18]) Consider the following point x K 2 (n, r) where r = [n, n] T 2 if i = [ n, n] T x i = n, if i = [3, 1] T, 0 otherwise. Clearly x 0 and i I x i i = r implying x K 2 (n, r). Now consider the following π R I 0 if i {[ n, n] T, [3, 1] T, [ n, 0] T, [0, n] T, [0, 0] T }, π i = 1 if i = [n, n] T, 1/2 otherwise 7

and note that π satisfies the normalization and complementarity conditions (NC1), (NC2) and (RCO). Clearly πx 1 is not a valid inequality as π x = 0. However it is possible to show that π satisfies i I q i π i π q for all q ˆQ where ˆQ = q (n) Z I2 + : q i i I r I N and 1 < q 1 < n/3 + 1. i I 2 (n) Notice that ˆQ is exponential in size and contains all q with small norm. This, in turn, implies that to guarantee the validity, one needs a q Q with large norm in (3). We can show that π satisfies i I q i π i π q for all q ˆQ by contradiction. Assume that there actually exists a vector q ˆQ such that i I q i π i < π q. Let W = {[ n, n] Ț [3, 1] Ț [ n, 0] Ț [0, n] Ț [0, 0] T } and note that I r = I Z 2 +. As q, π 0, we must have π q > 0, implying q W and π q {1/2, 1}. Furthermore, π q > 0 implies that q I N, and therefore, q I r. If π q = 1/2, then we must have i I q i π i = 0 implying π i = 0 for all i I such that q i > 0. In other words, if q i > 0 then i W. In this case, if q [ n,n] T = 0, then the remaining elements in W clearly can not produce a nonnegative vector other than 0 as a conic combination. On the other hand, if we have q [ n,n] T 1, then the first component of q is negative unless q [3, 1] T n/3 implying q 1 n/3 + 1 if the conic combination gives a nonnegative vector. Finally, if π q = 1 then q = r = [n, n] T. In this case, for all but one i I we must have q i π i = 0. Furthermore, if q i π i > 0 for some i I, then π i = 1/2 and q i = 1. As the elements in W can not produce the nonnegative vector [n, n] as a conic combination unless q 1 n/3 + 1, we assume that there indeed exists a j I \ {[n, n]} such that π j = 1/2 and q j = 1. Therefore j plus a conic combination of vectors in W must give r = [n, n] T. This implies that the elements in W must produce the nonnegative vector [n, n] j 0 as a conic combination. This again is not possible unless q 1 n/3 + 1. Until this example was presented to us by our colleague Grzegorz Swirszcz, we made multiple attempts to characterize a set Q m with properties (i) and (ii) described above. Based on this example we decided that, even for m = 2, it would be hard to find a compact characterization of a nontrivial polar for K m (n, r). 3 Separation and optimization using polaroids In Section 2.3 we argued that it appears hard to obtain a compact description of a nontrivial polar for K m (n, r). In spite of this, we discuss in this section how to solve the separation problem efficiently. Given a point x R I, the separation problem is to first decide if x K m (n, r), and if x K m (n, r), then to find a valid inequality violated by x. The separation problem can be easily solved if an explicit and compact description of the polar or nontrivial polar is known. In this section, we present a polar-like set which also leads to an efficient separation algorithm. We next formally define the properties of this set. 8

Definition 3.1 A polaroid T for K m (n, r) is a polyhedron in R I such that: 1. If π T, then π satisfies the normalization conditions (NC1) and (NC2). 2. If π T then πx 1 is valid for all x K m (n, r); 3. If πx 1 is a nontrivial facet-defining inequality of K m (n, r) satisfying (NC1) and (NC2), then π T. Let LP (K m (n, r)) denote the continuous relaxation of K m (n, r). Theorem 3.2 Given a point x LP (K m (n, r)), and a polaroid T for K m (n, r). Then 1. x can be separated from K m (n, r) by solving a linear program (LP) over T, and, 2. if x K m (n, r) then a violated facet-defining inequality can be obtained by solving a second LP over T provided that the solution to the second LP is an extreme point solution. Proof. Since x LP (K m (n, r)), it satisfies all trivial inequalities x 0. To solve the separation problem, solve: min πx (LP1) s.t. π T We start by establishing some properties of a solution ˆπ T. As ˆπx 1 is valid for K m (n, r) it can be obtained by relaxing the right hand side of a conic combination of facet-defining inequalities plus a linear combination of the equations i I i dx i = r d for d = 1,..., m. In other words, there exists multipliers λ R M + and α R d such that ˆπ i 1 M λ k πi k + iα, i I (4) k=1 M λ k + rα (5) k=1 hold. Here π k x 1 for k = 1,..., M is the collection of all nontrivial facet-defining inequalities normalized according to (NC1) and (NC2). The inequalities in (4) follow from the fact that ˆπ i = M k=1 λ kπi k + m d=1 i dα d + µ i, for i I where the last term µ i 0, corresponds to adding µ i times the trivial inequality x i 0. Notice that, for each d = 1,..., m, inequality (4) for i = ne d becomes: ˆπ i M k=1 λ kπi k nα d. But since ˆπ i = πi k = 0, we have that α d 0. From the fact that M k=1 λ k + m d=1 r dα d. Combining this with ˆπ r = πr k = 1 we have that (4) for i = r becomes 1 inequality (5) we obtain M k=1 λ k + rα = 1. We have already shown that α 0, so this also implies that 1 M k=1 λ k 0. Now let h = min{π k x } M k=1 and remember that both rα and M k=1 λ k are nonnegative. For any ˆπ T, we have that ˆπx M k=1 λ kπ k x + i I (x i i)α M k=1 λ kh + rα where the last term gives a convex combination of h and 1 and therefore is at least min{1, h}. Therefore, (LP1) is bounded and has an optimal solution π. If π is such that π x < 1, then π x 1 is a valid 9

inequality separating x from K m (n, r). If π x 1, then all nontrivial facet-defining inequalities are satisfied by x and all trivial ones are also satisfied by it. Therefore x K m (n, r). For the facet separation problem, note that the violated inequality π x 1 obtained using (LP1) is not necessarily facet-defining. To handle this, let z = π x and solve: min s.t. i I π i π T πx = z (LP2) Clearly π is feasible for (LP2) and any (extreme) solution π to (LP2) gives a violated inequality π x 1 that separates x from K m (n, r). We now need to argue that π x 1 defines a facet of K m (n, r). Since π T we can still say that there exists multipliers λ R M + and α R d such that (4) and (5) hold. As π gives a violated inequality, we have that M M 0 < 1 π x λ k + rα k=1 k=1 λ k π k x i I M M iαx i = λ k λ k π k x. (6) But notice that if M k=1 λ k = 0, then λ k = 0 for all k which contradicts (6). Hence M k=1 λ k > 0. Thus we can set γ k = λ k / M k=1 λ k and rewrite (6) as ( M ) ) M 0 < 1 π x λ k (1 γ k π k x (7) which implies that 1 M γ k π k x > 0. k=1 Using M k=1 λ k 1, we can now write k=1 k=1 M M 0 < 1 π x 1 γ k π k x = γ k (1 π k x ) 1 π x. (8) The last inequality follows from the fact that π is also an optimal solution to (LP1). Since equality has to hold throughout, we have that M k=1 λ k = 1. But now, notice that k=1 k=1 k=1 k=1 π i i I i I M λ k πi k + iα k=1 i I = i I = i I M λ k πi k + m i d α d k=1 i I d=1 M m M λ k πi k + α d i d = λ k πi k i I k=1 d=1 i I k=1 i I π i where the first inequality follows from (4) and the last equation follows from i I i d = 0 for all d = 1,..., m. Since equality has to hold throughout, all inequalities in (4) hold as equality. In particular, the inequalities for i = ne d imply that α = 0. Thus, π = M k=1 λ kπi k, where M k=1 λ k = 1 and λ k 0. As π is assumed to be an extreme point solution to (LP2), it can not 10

be a convex combination of other optimal solutions and therefore π is a nontrivial facet-defining inequality. Notice that in the proof above we also showed the following. Corollary 3.3 If T is a polaroid and πx 1 is a nontrivial facet-defining inequality of K m (n, r), then π is an extreme point of T. We now show that the polaroid can be used to give an extended formulation of K m (n, r). Let T = {π : Aπ b} be a polaroid for K m (n, r) for some matrix A and vector b of appropriate dimensions. By Theorem 3.2, a point x in LP (K m (n, r)) belongs to conv(k m (n, r)) if and only if πx 1 for all π T, that is, x π 1 defines a valid inequality for T. Therefore by LP duality, x conv(k m (n, r)) if and only if there exists a vector λ 0 such that x = λa and λb 1. In other words, x is a nonnegative linear combination of the inequalities defining the polaroid and conv(k m (n, r)) = {x R I : x = λa, λb 1, λ 0, x LP (K m (n, r))}. This yields the following corollary. Corollary 3.4 Given a polaroid T for K m (n, r), one can solve the problem min{cx : x K m (n, r)} by solving a single LP. One can similarly frame the problem of optimizing over the MCGP as a linear program, via Gomory s characterization of its non-trivial polar, and LP duality. 4 A polaroid for K m (n, r) for m = 2, 3 In Section 3, we showed that a compact description of a polaroid is useful for separation over K m (n, r). In this section, we obtain compact descriptions of polaroids when m = 2 and m = 3. Lemmas 2.4 and 2.5 imply that a polyhedral set in R Im (n) defined by (NC1), (NC2) and any subset of the k-term subadditivity constraints (KSA) will automatically satisfy properties 1 and 3 in the definition of a polaroid. If for some fixed l > 0, the k-term subadditivity constraints with k l imply (along with (NC1)) validity for K m (n, r), then we can define a polaroid T via (NC1), (NC2) and a subset of the constraints (KSA). Clearly there are at most I m (n) k k- term subadditivity inequalities. To obtain a polaroid defined by few inequalities, we would like to compute the smallest number k (m) such that k (m)-term subadditivity and (NC1) imply validity for K m (n, r). We emphasize that Gomory [12] shows that the 2-term subadditivity constraints π i + π k π (i+k) (mod n) are sufficient for validity for P m (n, r) for all m Z + (for P m (n, r), these constraints are defined in a modular fashion). Therefore, if we show that k (m) > 2 for any m 2, we will have shown that the set K m (n, r) is fundamentally a more complicated object than P m (n, r). In the remainder of this Section, we obtain a constant upper bound on k (m) for m = 2, 3 via some geometrical properties of vectors in R m. We start with the following notation. 11

A vector b R 2 has sign pattern (+ ) (for example) if b 1 0 and b 2 < 0. We also denote this fact by b (+ ). We use similar notation for vectors in R m for m 2. Given a collection of vectors (b 1,..., b t ), we say that a sign pattern p is absent from the collection if no vector in the collection has sign pattern p, otherwise we say that p is present. If two sign patterns have opposite signs in each co-ordinate, we say they are opposing sign patterns. The following observation regarding vectors with opposing sign patterns is true for any m. Observation 4.1 If b, c [ 1, 1] m have opposing sign patterns, then b + c [ 1, 1] m. Proof. A vector v R m is in [ 1, 1] m if and only if v 1. But since b and c have opposing sign patterns, for j = 1,..., m the number (b + c) j has absolute value less than or equal to max j { b j, c j },. Thus b + c max{ b, c } 1. 4.1 The case m = 2 The following result regarding collections of vectors in [ 1, 1] 2 will imply that k (2) 2. Lemma 4.2 In any collection of vectors S = (b 1,..., b t ) such that t 2, b i [ 1, 1] 2 for i = 1,..., t and t i=1 b i [ 1, 1] 2, there is a pair of vectors b k, b l with b k + b l [ 1, 1] 2. Proof. Without loss of generality we can assume that t i=1 b i = r = (r 1, r 2 ) with (0, 0) (r 1, r 2 ) (1, 1). If S contains vectors with opposing sign patterns, then by Observation 4.1 the result holds for S. Assume that opposing sign patterns are not present in S; without loss of generality we can assume the sign pattern ( +) is not present in S. Then the pattern (+ +) is present in S as r 1 0, and therefore ( ) is not present in S. Let b k be any vector with sign pattern (+ +). If S has a vector b l with sign pattern (+ ), then b k + b l [ 1, 1] 2 as the first component of b k + b l is contained in [0, r 1 ] [0, 1]. Otherwise, all vectors in S have sign pattern (+ +) and any vector b l b k satisfies b k + b l [0, 1] 2. Note that, in Lemma 4.2, the choice of the interval [ 1, 1] 2 is arbitrary. The following remark formalizes this comment. Remark 4.3 Let m be a positive integer and let α R m with α > 0 and b i R m for all i S. Then α i S b i α, if and only if i S b i [ 1, 1]m, where b i Rm is defined by dividing the kth component of b i by α k for k = 1,..., m. Let I = I 2 (n). Let I ++ be the set of points in I with sign pattern (+ +), and let I, I + and I + be defined in a similar manner. Given a vector i I, remember that f i denotes the unit vector in R I with a 1 in component i and 0 in all other components. Theorem 4.4 If π R I satisfies 2-term subadditivity and π r 1, then πx 1 is valid for K 2 (n, r). Moreover, the theorem is still true if π only satisfies π r 1 and the following 2-term subadditivity conditions: π i + π j π i+j, i I +, j I +, (9) π i + π j π i+j, i I ++, j I, i + j I, (10) 12

Proof. Suppose πx 1 is not valid for K 2 (n, r). Then let x be a point in K 2 (n, r) such that πx < 1 and x 1 is as small as possible. If x 1 = 1, then i I ix i = r implies that x r = 1, and x i = 0 for i r; which contradicts the fact that π r 1. Therefore x 1 2. Define t = x 1 and a collection S = (b 1,..., b t ) of vectors having x i copies of vector i for all i I. Therefore b j [ n, n] 2 for j = 1,..., t. Using Lemma 4.2 and Remark 4.3, we know that there are vectors b and c in S such that b + c [ n, n] 2. Moreover, as b, c I = [ n, n] 2 Z 2, we know that b + c I. Define x := x f c f b + f b+c. Then x K 2 (n, r) and yet x 1 < x 1 and πx = πx π b π c + π b+c πx. The previous inequality holds because π satisfies the 2-term subadditivity condition π b + π c π b+c. But this contradicts the choice of x. The second part of the theorem follows from the proof of Lemma 4.2, which guarantees that either b and c can be chosen to have opposing sign patterns or b can be chosen to be in I ++. (notice that the case where b I ++ and c I satisfies both conditions). Theorem 4.4 implies that k (2) = 2. Moreover, we can state the following corollary. Corollary 4.5 The following is a description of a polaroid for K 2 (n, r): π i + π j π i+j, i I +, j I +, (11) π i + π j π i+j, i I ++, j I, i + j I, (12) π r = 1, (13) π i = 0, i I N. (14) 4.2 The case m = 3 We now prove an analogue of Lemma 4.2 for vectors in R 3. As noted in Remark 4.3, such a result is equivalent to its variant obtained by replacing 1 by any α > 0. Lemma 4.6 Let S = (b 1,..., b t ) be a collection of vectors with t 2, b i [ 1, 1] 3 for i = 1,..., t and t i=1 b i [ 1, 1] 3. Then there is a set T {1,..., t} with 2 T 4 such that i T b i [ 1, 1] 3. Proof. First, note that the order of elements in the collection S does not matter. In addition we allow S to have multiple copies of the same element. In this proof we abuse notation and use set operations on collections as follows: (i) S S means that for every distinct element in S, the collection S has at least as many copies of that element as S does, (ii) S = S means that S S and S S, (iii) S S denotes the collection obtained by concatenating S and S, (iv) For S S, we define S \ S to denote their difference, that is (S \ S ) S = S. Without loss of generality we can assume that t i=1 b i = r has sign pattern (+ + +). If the lemma is not true, then consider the collection of vectors S which forms the smallest counterexample to the claim, i.e., t i=1 b i = r, where 0 r 1, and no subcollection of four or fewer vectors in S has its sum contained in [ 1, 1] 3. Clearly, t 5. If any vector in S (say b 1 ) has the sign pattern (+ + +), then t i=2 b i = r b 1 [ 1, 1] 3, and there is a subcollection of four or fewer 13

vectors from b 2,..., b t with sum contained in [ 1, 1] 3, a contradiction to the fact that S has no such subcollection. We can therefore assume that the sign pattern (+ + +) is not present in S. We now divide the remaining sign patterns into three groups. A = {(+ + ), (+ +), ( + +)}, B = {( +), ( + ), (+ )}, C = {( )}. Case 1: Assume none of the sign patterns from A are present in S. Let p be the sign pattern from B with fewest vectors in S (we may assume by symmetry that p is ( +)). Let S be a subcollection of vectors consisting of all vectors with sign pattern p, and an equal number of vectors with the other two sign patterns from B. If S, let c 1, c 2, c 3 be vectors from S where each has a distinct sign pattern from B (say c 1 has the first sign pattern and so on). Clearly c 1 + c 2 [ 1, 1] 3, and as a consequence the first component of c 1 + c 2 is less than -1 (c 1 and c 2 have opposite signs in the last two co-ordinates, and therefore the sums in these co-ordinates are contained in [ 1, 1]). As c 2 + c 3 and c 1 + c 3 are also not contained in [ 1, 1] 3, we conclude that c 1 + c 2 + c 3 < 0. Since this holds for all triples of vectors in S with distinct sign patterns, the sum of vectors in S is negative if S, or 0 if S =. But in either case the vectors in S \ S have a negative sign in the last co-ordinate, and thus t i=1 b i cannot equal r 0. Case 2: Assume exactly one of the sign patterns from A is present in S; without loss of generality, we can assume it is (+ + ). The first sign patterns of A and B are opposing patterns, and therefore the first sign pattern from B cannot be present in S. Then all sign patterns in S have a negative sign in the third co-ordinate, and t i=1 b i cannot equal r 0. Case 3: Assume exactly two sign patterns from A are present in S. Assume, without loss of generality, that the third sign pattern from A is absent from S. Note that this rules out the presence in S of the first two sign patterns from B. Case 3a: Assume ( ) is not present in S. In this case all vectors in S have a positive sign in the first co-ordinate. By Lemma 4.2, there are two vectors in S, say b i and b j, such that the last two components of b i + b j are contained in [ 1, 1] 2. But the first component of b i + b j is contained in [0, r 1 ], and therefore b i + b j [ 1, 1] 3, a contradiction. Case 3b: Assume the pattern ( ) is present in S. As the sum of no pair of vectors from S is contained in [ 1, 1] 3, it follows that if c 1, c 2 and c 3 are any three vectors with the sign patterns (+ + ), (+ +), ( ) respectively, then c 1 +c 2 +c 3 has the sign pattern (+ ). Now let S be a maximal subcollection of S with equal number of vectors from the above three sign patterns. Again i S b i has sign pattern (+ ). If S \ S does not contain either of the first two patterns, then t i=1 b i is negative in some co-ordinate. Therefore S \ S contains a vector b i (+ + ) and b j (+ +), and no vector with sign pattern ( ). In this case, the first component of b i + b j [0, r 1 ], and therefore b i + b j [ 1, 1] 3, a contradiction. Case 4: Assume all three sign patterns from A are present in S. Then all sign patterns from B are absent from S. Let c 1, c 2, c 3 be any three vectors having the distinct sign patterns from A (in order). As c 1 + c 2 [ 1, 1] 3, the first component of c 1 + c 2 lies in (1, 2], and therefore the first component of c 1 + c 2 + c 3 lies in (0, 2). Arguing in a similar fashion for the other components, we 14

conclude that c 1 + c 2 + c 3 (0, 2) 3, and at least one of its components is greater than 1. Case 4a: The pattern ( ) is present in S. Let c 4 ( ). Then 3 i=1 c i > 0 4 i=1 c i 1. Furthermore, for i = 1, 2, 3, the ith component of c 4 + c 4 i is contained in [ 2, 1) and hence 4 i=1 c i 1. Therefore 4 i=1 c i [ 1, 1] 3. Case 4b: The pattern ( ) is absent from S. Let S (c 1, c 2, c 3 ) be a maximal subcollection of S with equal number of vectors from the three sign patterns in A. Clearly, b i S b i has sign pattern (+ + +), as the sum of every subset of three vectors with distinct sign patterns is strictly positive in each component. If S = S, c 1 + c 2 + c 3 [0, 1] 3, a contradiction. We can therefore assume that S S. Without loss of generality we can assume that the pattern ( + +) is absent from S \ S ; then all vectors in S \ S have a positive sign in the first co-ordinate. This implies that both b i S b i and b i S\S b i have their first component in the range [0, r 1 ]. If there exist two vectors d 1 and d 2 in S \ S with distinct sign patterns, then the first component of d 1 + d 2 is contained in [0, r 1 ], and d 1 and d 2 have opposite signs in the last two co-ordinates. Therefore, d 1 + d 2 [ 1, 1] 3, a contradiction. Therefore we can assume that all vectors in S \ S have the same pattern from A, say (+ + ). This implies that b i S b i and b i S\S b i have their first two components contained in [0, r 1 ] [0, r 2 ]. As t 5, it follows that either S \ S = 1 and S 6, or S \ S 2. In the first case, at least one subset of three vectors (say c 1, c 2, c 3 ) from S with distinct sign patterns has the third component of its sum bounded above by 1 (otherwise the sum of vectors in S would have its last component greater than 2). Then c 1 + c 2 + c 3 [0, 1] 3. In the second case, let d 1 and d 2 be any two vectors from S \ S. Then the first two components of d 1 + d 2 are contained in [0, 1] and therefore its last component is contained in [ 2, 1). This implies that u = c 2 + c 3 + d 1 + d 2 is contained in [ 1, 1] 3. For the first two components, the lower bound of -1 follows from the fact that only c 3 (respectively, c 2 ) has a negative number in the first (respectively, second) component, and the upper bound follows because these components of c 1 + u are bounded above by 1. Also the last component of c 2 + c 3 is contained in (1, 2], and therefore the last component of c 2 + c 3 + d 1 + d 2 is contained in [ 1, 1]. Let I = I 3 (n). Given a vector i I, let f i be the unit vector in R I with a 1 in component i and 0 in all other components. Theorem 4.7 If π R I satisfies 4-term subadditivity and π r K 3 (n, r). 1 then πx 1 is valid for Proof. Suppose πx 1 is not valid for K 2 (n, r). Let x be a point in K 3 (n, r) such that πx < 1 and x 1 is as small as possible. As in the proof of Theorem 4.4, we can assume that x 1 2, and that there is a collection S = (b 1,..., b t ) of integer vectors from [ n, n] 3 with t = x 1 such that t j=1 b j = r, and each b j satisfies x b j > 0. By the previous theorem, there is a subset T of {1,..., t} with 2 T 4 such that j T b j = λ I. If x = x j T f b j + f λ, then x 1 < x 1. Further, x K 3 (n, r) and πx = πx j T π b j + π λ πx, a contradiction. Therefore, k (3) 4 and we have the following corollary. 15

Corollary 4.8 The following is a description of a polaroid for K 3 (n, r): π i + π j + π k + π l π i+j+k+l i, j, k, l I, with (i + j + k + l) I, (15) π r = 1, (16) π i = 0, i I N. (17) 5 Validity via subadditivity In the previous Section, we focused on obtaining an upper bound on k (m), that is, the smallest number such that k (m)-term subadditivity and (NC1) guarantee validity for K m (n, r). In this Section, we focus on deriving lower bounds for k (m). In particular, the results in this Section show that the upper bound of k (3) 4 obtained in Section 4 is tight. We start with an example for K 3 (n, r) that shows that 3-term subadditivity of the coefficients is not sufficient to guarantee validity of an inequality. Example 5.1 Let a = 10 10 10 b = 10 10 10 c = 1 1 9 and consider the inequality πx 1 and the point x K 3 (10, 2) where 2 R 3 is a vector of all twos and π i = { 0 if i {a, b, c} 1, otherwise, 1 if i {a, b} and x i = 2 if i = c 0, otherwise. Notice that even though π x = 0 1, π satisfies all 2 and 3-term subadditivity conditions. To see this, first note that as π {0, 1} I, if π i + π j π i+j is violated by some i, j I with i + j I, then π i = π j = 0 and π i+j = 1. In other words i, j {a, b, c} and i + j I. This, however, is not possible as i + j I if i, j {a, b, c}. Similarly, if π i + π j + π k π i+j+k is violated by some i, j, k I with i + j + k I, then π i = π j = π k = 0 and π i+j+k = 1. This, again, is not possible as i + j + k I if i, j, k {a, b, c}. Given a point x in Z I, we define its support, denoted as supp(x), to be the collection of indices i I for which x i > 0. For the point x in Example 5.1, supp( x) = {a, b, c}. We next define a function F which maps a matrix consisting of columns from I and a vector of column weights to a point in Z I. More formally, for some l > 0 let Q be a matrix in Z m l with Q j (the jth column of Q) contained in I for j = 1,..., l. Furthermore, assume that all Q j are distinct and let w Z l + be a vector with non-negative entries. Then p = F(Q, w) is a point in Z I + defined as follows: { w j if i = Q j, p i = 0 otherwise. 16

Therefore p 1 = w 1 and i I p i i = Qw and supp(p) = {Q j } {j=1,...,l : wj >0}. From this point on, whenever we say that p = F(Q, w), we are implicitly assuming that Q is a matrix in Z m l with distinct columns contained in I and w Z l +. Conversely, if y Z I + is a vector where supp(y) {Q 1,..., Q k }, then we define v = F 1 (Q, y) as a vector in Z k + such that for all j = 1,..., k, v j = y Qj. Note that Qv = i I y i i and v 1 = y 1. Using the above definitions, x in Example 5.1 equals F([a, b, c], [1, 1, 2] T ) and [1, 1, 2] T = F 1 ([a, b, c], x). We next formalize this example and produce increasing lower bounds on k (m) for m 3. We start with the following definition. Definition 5.2 Let p = F(Q, w) Z I +. The point p is called a bad point for Km (n, r) if it satisfies the following conditions: (i) r = Qw [1, n] m, (ii) w 1 > 2, and, (iii) Qt I for all t Z l + with 1 < t 1 < w 1. Further, we say p is an extremely bad point component of n for all j = 1,..., l. if in addition to being a bad point, Q j has a Using this definition, the point x of Example 5.1 is a bad point. We next formally state that bad points lead to lower bounds on k (m). Lemma 5.3 If p Z I + is a a bad point for Km (n, r p ), then k (m) p 1. Proof. The proof essentially follows from Example 5.1. Consider the inequality πx 1 where π i takes the value 1 if p i = 0 and 0 otherwise. Clearly πp = 0 1 while π satisfies k-term subadditivity for all k < i I p i. Clearly, the point presented in Example 5.1 is not an extremely bad point. However, it is easy to see that the following point q 3 = F(Q 3, w 3 ) K 3 (10, 1) is an extremely bad point where w 3 = 1 1 1 and Q 3 = 10 10 1 1 10 10 10 1 10 Define r 3 = 1 and n 3 = 10. We next describe an iterative procedure that, for m > 3, constructs an extremely bad point q m K m (n m, r m ) using q m 1 K m 1 (n m 1, r m 1 ). The points {q m } constructed by this procedure will have the property that q m 1 would grow exponentially with m. Given q m 1 = F(Q m 1, w m 1 ), we define the scaling factor α = 2 w m 1 1 and let n m = αn m 1. We define Q m Z m m and w m Z m + as follows: w m = [ w m 1 w m 1 1 1 ] and Q m = αq m 1 n m... n m. 1. 1 n m. (18) 17

Finally, we define the new point q m as F(Q m, w m ). Let r m = Q m w m ; then [ ] [ r m αq m 1 w m 1 ( w m 1 1 1)1 m 1 α r m 1 ( w m 1 1 1) 1 m 1 = n m w m 1 1 n m ( w m 1 = 1 1) n m ]. As r m 1 [1, n m 1 ] m 1 and α = 2 w m 1 1, one can verify that r m [1, n m ] m. Note that w m 1 = 2 w m 1 1 1, and therefore q m 1 = w m 1 = 2 m 2 + 1 for all m 3. Also notice that if all diagonal entries of Q m 1 are equal to n m 1, then all diagonal entries of αq m 1 are equal to αn m 1 and therefore all diagonal entries of Q m are equal to n m. As this condition holds for m = 3, it holds for all Q m with m 3. Lemma 5.4 The points {q m } generated by the procedure above are extremely bad for their corresponding K m (n m, r m ), for m 3. Proof. The case m = 3 can be easily checked to be true. When m > 3, due to the construction, we have (i) r m [1, n m ] m and (ii) q m 1 > 2. Furthermore, all columns of Q m have a component of n m. Let I = I m (n m ) and α = n m /n m 1 = 2 q m 1 1. To see that q m satisfies the remaining condition for being an extremely bad point, let t Z m + satisfy 1 < t 1 < q m 1. Let s = Qt, and t = [t 1,..., t m 1 ] T. We next show that s = Qt I by considering the following cases. [Case 1] t 1 = 0. As t m = t 1 > 1, the last component of s is strictly less than n m and s / I. [Case 2] t 1 = 1. In this case let t j = 1 for some m 1 j 1 and note that t m 1. As the j th diagonal entry of Q m is n m and as the last column only has negative entries, we have s I as its j th row is strictly less than n m. [Case 3] q m 1 1 > t 1 > 1. Let I = I m 1 (n m 1 ). As q m 1 is an extremely bad point for K m 1 (n m 1, r m 1 ) and q m 1 1 > t 1 > 1, we have Q m 1 t I. In other words, Q m 1 t has a component that is not contained in [ n m 1, n m 1 ]. Therefore, αq m 1 t has a component with absolute value greater than or equal to α(n m 1 + 1) = n m + α. As the first m 1 rows of s are given by αq m 1 t t m 1 m 1 and t m < t 1 < q m 1 < α, we have s I. [Case 4] t 1 q m 1 1. In this case, t m t 1 q m 1 1 < q m 1 q m 1 1 = q m 1 1 1, implying that t m q m 1 1 2. Therefore, the m th component of s is at least q m 1 1 n m ( q m 1 1 2)n m = 2n m and s I. As q m 1 = 2 m 2 + 1, Lemma 5.4 together with Lemma 5.3 imply that k (m) 2 m 2 + 1. We next present another family of points and show that this lower bound on k (m) can be improved. Let p 3 = F(P 3, v 3 ) be the bad point for K 3 ( n 3, r 3 ) presented in Example 5.1, where P 3 = [a, b, c], v 3 = [1, 1, 2] T, n 3 = 10 and r 3 = 2. For m 4, let q m 1 = F(Q m 1, w m 1 ) be the extremely bad point constructed above and define the scaling factor β = 4 q m 1 1 and let n m = βn m 1. We next 18

define the point p m = F(P m, v m ) where [ ] v m w m 1 = 2 w m 1 1 2 and P m = βq m 1 1. 1 n m... n m n m /2 1. and note that n m is even and n m /2 1 is indeed an integer. Also note that all entries of P m are in [ n m, n m ]. Let r m = P m v m, then, [ ] r m βq m 1 w m 1 (2 w m 1 1 2)1 m 1 = n m w m 1 1 n m ( w m 1 1 1) (2 w m 1 1 2) [ ] β r m 1 2( w m 1 1 1) 1 = n m 2 w m 1. 1 + 2 Note that as w m 1 1 > 1 we have r m < n m 1. In addition, r m 1 as (i) r m 1 1 and therefore βr m 1 β = 4 q m 1 1 = 4 w m 1 1 > 2 w m 1 1 + 2. Furthermore, (ii) n m > 2 q m 1. Lemma 5.5 The points {p m } generated by the procedure above are bad points for their corresponding K m ( n m, r m ), for all m > 3. Proof. The proof is very similar to that of Lemma 5.4. We again have (i) r m [1, n m ] m and (ii) p m 1 > 2. Let I = I m ( n m ) and β = n m /n m 1 = 4 q m 1 1. Assume t Z m + satisfy 1 < t 1 < p m 1 and let s = Qt, and t = [t 1,..., t m 1 ] T. If q m 1 1 > t 1, then the proof is the same as Lemma 5.4, substituting α by β, n m by n m, Q m by P m and q m by p m. Therefore we need to consider t 1 q m 1 1. In this case, t m t 1 q m 1 1 < p m 1 q m 1 1 = 2 q m 1 1 2, implying that t m 2 q m 1 1 3. Therefore, the m th component of s is at least q m 1 1 n m ( n m /2+1)(2 q m 1 1 3) = ( n m +3)+( n m 4 q m 1 1 )/2. As n m = βn m 1 = 4 q m 1 1 n m 1, the m th component of s is strictly greater than n m and we have s I. Remember that q m 1 = 2 m 2 +1 for all m 3 and therefore p m 1 = 3 2 m 3 +1 for all m > 3. Combining this observation and the bad point x K 3 (10, 2) of Example 5.1 with Lemma 5.3, we have the following result: Corollary 5.6 For all m 3, there exists a vector π of appropriate dimension that satisfies up to (3/8)2 m -term subadditivity conditions and yet the corresponding inequality π m x 1 is not valid for K m (n, r). In other words, k (m) (3/8)2 m + 1. In particular, this shows that k (3) 4 and hence from the results in Section 4, k (3) = 4. 19

6 Lower bounds on the size of the description of a polaroid In Section 5 we derived lower bounds on the smallest number k (m) that would guarantee that any π that satisfies k (m)-term subadditivity and (NC1) would yield a valid inequality πx 1 for K m (n, r). Even though this lower bound grows exponentially in m, this does not necessarily imply that one can not obtain a compact description of a polaroid. In particular, recall from Lemma 2.4 that all nontrivial facet-defining inequalities of K m (n, r) satisfy the complementarity conditions (RCO) and the normalization conditions (NC1) and (NC2). It is therefore, possible that, for some k (m) < k (m), k (m)-term subadditivity conditions plus (NC1), (NC2) and (RCO) would give a description of a polaroid. In this section, we will show that, even in the presence of (NC1), (NC2) and (RCO), k (m)-term subadditivity does not guarantee validity if k (m) is small. In fact, the same lower bounds for k (m) derived from the bad points constructed in the previous section still hold for k (m). We next show that bad points that satisfy some additional conditions lead to lower bounds on k (m). We will later show that the bad points constructed in the previous section satisfy such conditions. Lemma 6.1 Let m 3. Let p = F(Q, w) be a bad point for K m (n, r) such that 0 < w Z l. If Q(w t) / I for all t Z l + such that 1 t 1 < ( p 1 + 1)/2, then there exists π R I that satisfies ( p 1-1)-term subadditivity, (RCO), (NC1) and (NC2) and such that πx 1 is not valid for K m (n, r). Proof. Let γ = p 1 + 1. Construct π R I as follows: 1/γ if i supp(p), 1 if i = r, π i = 0 if i = 0, 1/2 otherwise. The inequality πx 1 is not valid for K m (n, r) as p K m (n, r), yet πp = p 1 /( p 1 + 1) < 1. Claim: π satisfies (RCO): Recall that I r = {i I : r i I} and note that I r supp(p) =. Indeed, if this were not true and p j > 0 for some j I r, then let c(j) be the index of the column of Q that is equal to j and define w = w e c(j), where e c(j) Z l is the unit vector with a 1 in the c(j)-th component. But then w Z l + and r = Qw = Qw + Q c(j) = Qw + j and thus Qw = r j I, which contradicts the fact that p is a a bad point. Therefore, for all i I r, we have π i + π r i = 1. Claim: π satisfies ( p 1-1)-term subadditivity: If for some k with 2 k p 1 1, π violates a k-term subadditivity condition, then q i π i < π q 1, (19) i I 20