Sampling Contingency Tables

Similar documents
Sampling Contingency Tables

On the complexity of approximate multivariate integration

Polynomial-time counting and sampling of two-rowed contingency tables

Minimal basis for connected Markov chain over 3 3 K contingency tables with fixed two-dimensional marginals. Satoshi AOKI and Akimichi TAKEMURA

Model Counting for Logical Theories

RESEARCH ARTICLE. An extension of the polytope of doubly stochastic matrices

arxiv:math/ v1 [math.co] 3 Sep 2000

A Note on Tiling under Tomographic Constraints

Approximate Counting and Markov Chain Monte Carlo

TRISTRAM BOGART AND REKHA R. THOMAS

Algebraic Methods in Combinatorics

Combinatorics for algebraic geometers

Detailed Proof of The PerronFrobenius Theorem

Bricklaying and the Hermite Normal Form

First Order Convergence and Roots

The Triangle Closure is a Polyhedron

Some Sieving Algorithms for Lattice Problems

A combinatorial algorithm minimizing submodular functions in strongly polynomial time

Lecture 16: October 29

The Shortest Vector Problem (Lattice Reduction Algorithms)

Chapter 1. Preliminaries

The Complexity of the Permanent and Related Problems

The cocycle lattice of binary matroids

A Characterization of (3+1)-Free Posets

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

A Simple Algorithm for Clustering Mixtures of Discrete Distributions

3. Linear Programming and Polyhedral Combinatorics

THE LECTURE HALL PARALLELEPIPED

The decomposability of simple orthogonal arrays on 3 symbols having t + 1 rows and strength t

constraints Ax+Gy» b (thus any valid inequalityforp is of the form u(ax+gy)» ub for u 2 R m + ). In [13], Gomory introduced a family of valid inequali

The Skorokhod reflection problem for functions with discontinuities (contractive case)

Alexander Barvinok August 2008

Bounds on the generalised acyclic chromatic numbers of bounded degree graphs

Chordal structure in computer algebra: Permanents

A thesis presented to the faculty of San Francisco State University In partial fulfilment of The Requirements for The Degree

1.1.1 Algebraic Operations

Sampling Good Motifs with Markov Chains

INEQUALITIES OF SYMMETRIC FUNCTIONS. 1. Introduction to Symmetric Functions [?] Definition 1.1. A symmetric function in n variables is a function, f,

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery

Spectra of Semidirect Products of Cyclic Groups

A Geometric Interpretation of the Metropolis Hastings Algorithm

Lecture 6: Random Walks versus Independent Sampling

Some notes on Coxeter groups

A Simple Computational Approach to the Fundamental Theorem of Asset Pricing

Conductance and Rapidly Mixing Markov Chains

Radon Transforms and the Finite General Linear Groups

Negative Examples for Sequential Importance Sampling of Binary Contingency Tables

A Geometric Approach to Graph Isomorphism

Excluded permutation matrices and the Stanley Wilf conjecture

Lecture 22: Counting

REPRESENTATIONS OF S n AND GL(n, C)

LATTICE POINT COVERINGS

QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY

Tactical Decompositions of Steiner Systems and Orbits of Projective Groups

Lecture 6: September 22

INITIAL COMPLEX ASSOCIATED TO A JET SCHEME OF A DETERMINANTAL VARIETY. the affine space of dimension k over F. By a variety in A k F

Efficient Algorithms for Universal Portfolios

Randomized Algorithms

A Faster Strongly Polynomial Time Algorithm for Submodular Function Minimization

Another algorithm for nonnegative matrices

Corrigendum: The complexity of counting graph homomorphisms

Determinants of Partition Matrices

COMMUTING PAIRS AND TRIPLES OF MATRICES AND RELATED VARIETIES

Optimization of a Convex Program with a Polynomial Perturbation

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

An Algorithmic Proof of the Lopsided Lovász Local Lemma (simplified and condensed into lecture notes)

The partial-fractions method for counting solutions to integral linear systems

An algebraic proof of the Erdős-Ko-Rado theorem for intersecting families of perfect matchings

A Polytope for a Product of Real Linear Functions in 0/1 Variables

Welsh s problem on the number of bases of matroids

Min-Wise independent linear permutations

MAXIMAL PERIODS OF (EHRHART) QUASI-POLYNOMIALS

Auerbach bases and minimal volume sufficient enlargements

1 Directional Derivatives and Differentiability

Lower Bounds for Testing Bipartiteness in Dense Graphs

Introduction to Real Analysis Alternative Chapter 1

Note on shortest and nearest lattice vectors

YOUNG TABLEAUX AND THE REPRESENTATIONS OF THE SYMMETRIC GROUP

Additive Combinatorics Lecture 12

MIT Algebraic techniques and semidefinite optimization May 9, Lecture 21. Lecturer: Pablo A. Parrilo Scribe:???

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

Office of Naval Research contract no. N K-0377 (9) FINAL TECHNICAL REPORT

Algebraic Methods in Combinatorics

Lower Bounds on the Complexity of Simplex Range Reporting on a Pointer Machine

ON THE EQUIVALENCE OF CONGLOMERABILITY AND DISINTEGRABILITY FOR UNBOUNDED RANDOM VARIABLES

MAT-INF4110/MAT-INF9110 Mathematical optimization

Proclaiming Dictators and Juntas or Testing Boolean Formulae

chapter 12 MORE MATRIX ALGEBRA 12.1 Systems of Linear Equations GOALS

On the mean connected induced subgraph order of cographs

On Linear Subspace Codes Closed under Intersection

Structural Grobner Basis. Bernd Sturmfels and Markus Wiegelmann TR May Department of Mathematics, UC Berkeley.

Dimension Quasi-polynomials of Inversive Difference Field Extensions with Weighted Translations

The Matrix-Tree Theorem

Matching Polynomials of Graphs

A robust APTAS for the classical bin packing problem

Vertex colorings of graphs without short odd cycles

The Advantage Testing Foundation Olympiad Solutions

Log-Convexity Properties of Schur Functions and Generalized Hypergeometric Functions of Matrix Argument. Donald St. P. Richards.

3. Linear Programming and Polyhedral Combinatorics

Transcription:

Sampling Contingency Tables Martin Dyer Ravi Kannan John Mount February 3, 995 Introduction Given positive integers and, let be the set of arrays with nonnegative integer entries and row sums respectively and column sums respectively. Elements of are called contingency tables with these row and column sums. We consider two related problems on contingency tables. Given, and ) Determine. ) Generate randomly an element of, each with probability. The counting problem is of combinatorial interest in many contexts. See, for School of Computer Studies, Leeds University, Leeds, England School of Computer Science, Carnegie-Mellon University, Pittsburgh, PA 53, USA. Partially supported by NSF Grant CCR 908597. School of Computer Science, Carnegie-Mellon University, Pittsburgh, PA 53, USA. Partially supported by NSF Grant CCR 908597.

example, the survey by Diaconis and Gangolli [5]. We show that even in the case when or is, this problem is #P hard. The random sampling problem is of interest in Statistics. In particular, it comes up centrally in an important new method formulated by Diaconis and Efron [4] for testing independence in two-way tables. (See also Mehta and Patel [] for a more classical test of independence). We show that this problem can be solved in approximately in polynomial time provided the row and column sums are sufficiently large, in particular, provided Ω and Ω. This then implies under the same conditions, a polynomial time algorithm to solve the counting problem () approximately. The algorithm we give relies on the random walk approach, and is closely related to those for volume computation and sampling from log-concave distributions [0,, 9, 0, 5, ]. Preliminaries and Notation The number of rows will always be denoted by and the number of columns by. We will let denote. For convenience, we number the coordinates of any vector in by a pair of integers or just where runs from through and from through. Suppose is an vector of reals (the row sums) and is an vector of reals (the column sums). We define : for ; for

can be thought of as the set of real matrices with row and column sums specified by and respectively. Let : 0 for We call such a polytope a contingency polytope. Then, is the set of vectors in with all integer coordinates. Note that the above sets are all trivially empty if throughout that.. So we will assume We will need another quantity denoted by defined for and : max We will abbreviate to satisfying when are clear from the context. An easy calculation shows that if is Ω and is Ω, then is. On input (we assume that and ), our algorithm runs for time bounded by times a polynomial in max log log, and "! where! will be an error parameter specifying the desired degree of accuracy. (See section 5, first paragraph for a description of the input/output specifications of the algorithm.) 3 Hardness of counting We will show that exactly counting contingency tables is hard, even in a very restricted case. We will use the familiar notion of # P-completeness introduced by Valiant [5]. 3

Theorem The problem of determining the exact number of contingency tables with prescribed row and column sums is is #P-complete, even in the case. Proof It is easy to see that this problem is in #P. We simply guess in the range 0 (where integers is the sum of the row sums) and check whether they satisfy the constraints. The number of accepting computations is the number of tables. For proving hardness, we proceed as follows : Given positive integers, it is shown in [8] that it is #P-hard to compute the the polytope 0 It follows that it is #P-hard to compute the polytope 0 -dimensional volume of -dimensional volume of the where. Hence, by substituting,, that it follows that it is #P-hard to compute the (with rows and columns) where -dimensional volume of the polytope Now the number of integer points in contingency tables with the required row and column sums. is clearly the number of But, for integral, is a polytope with integer vertices. (It is the polytope of a transportation problem [].) Consider the family of polytopes for 4

. It is well known [, Chapter ] that the number of integer points in will be a polynomial in, the Ehrhart polynomial, of degree, the dimension of. Moreover the coefficient of will be the -dimensional volume of. It is also straightforward to show that the coefficients of this polynomial are of size polynomial in the length of the description of. Suppose therefore we could count the number of contingency tables for arbitrary. Then we could compute for arbitrary integral. Hence we could compute for and thus determine the coefficients of the Ehrhart polynomial. But this would allow us to compute the volume of, which we have seen is a #P-hard quantity. Remark If both are fixed, we can apply a recent result of Barvinok s [3] (that the number of lattice points in a fixed dimensional polytope can be counted exactly in polynomial time) to compute. Diaconis and Efron [4] give an explicit formula for in case. Their formula has exponentially many terms (as expected from our hardness result). Mann has a similar formula for 3. Sturmfels demonstrates a structural theorem which allows for an effective counting scheme that runs in polynomial time for fixed.[4] Sturmfels has used the technique to count 4 4 contingency tables in the literature. We will, in later papers, discuss our work counting 4 4 and 5 4 contingency tables. 5

4 Sampling Contingency tables : Reduction to continuous sampling This section reduces the problem of sampling from the discrete set of contingency tables to the problem of sampling with near-uniform density from a contingency polytope. To this end, we first take a natural basis for the lattice of all integer points in and associate with each lattice point a parallelepiped with respect to this basis. Then we produce a convex set which has two properties : a) each parallelepiped associated with a point in is fully contained in and b) the volume of divided by the total volume of the parallelepipeds associated with points in is at most. Now the algorithm is simple : pick a random point from with near uniform density (this can be done in polynomial time), find the lattice point in whose parallelepiped lies and accept if it is in [acceptance occurs with probability at least ]; otherwise, reject and rerun. The volume of each parallelepiped is the same, so the probability distribution of the in is near uniform. This general approach has been used also for the simpler context of sampling from the 0- solutions to a knapsack problem [3]. [While is a convex set and by the general methods, we can sample from it in polynomial time with near uniform density, it will turn out that is also isomorphic to a contingency polytope. In the next section, we show how to exploit this to get a better polynomial time algorithm than the general one.] We will have to build up some geometric facts before producing the. 6

Let be the lattice: : 0 for ; 0 for ; For and, let be the vector in given by and 0 for other than the 4 above. Any vector in 0 0 can be expressed as a linear combination of the s as follows (the reader may check this by direct calculation) () It is also easy to check that the are all linearly independent. This implies that the subspace 0 0 has dimension and so does the affine set which is just a translate of 0 0. Also, we see that if is an integer vector, then the above linear combination has integer coeffients; so, the form a basis of the lattice. It is easy to see that if are positive vectors, then the dimension of is also. To see this, it suffices to come up with an with row and column sums given by and with each entry strictly positive because then we can add any small real multiples of the to and still remain in an. Such is easy to obtain : for example, we can choose to satisfy 0 and summing to. Then we subtract this amount from the column sums and repeat the process on the second row etc. We will denote by Vol the ( dimensional) volume of. Obviously, if either or has a non-integer coordinate, then is empty. 7

Lemma : If are vectors of positive reals and are vectors of positive reals with and (componentwise), then Vol Vol Proof By induction on the number of coordinates of that are strictly greater than the corresponding coordinates of. After changing row and column numbers if necessary, assume without loss of generality that is the least among all POSITIVE components of. After permuting columns if necessary, we may also assume that we have 0. Let. Let be defined by : ; 0 for Let be an vector defined by ; for 3. Let be defined by ; for 3. Denote by 3 the set. Then and 3 have the same volume as seen by the isomorphism. Clearly, contains. Also the row sums and column sums in 3 are no greater than the corresponding row and column sums in. Applying the inductive assumption to 3 and, we have the lemma. Let be fixed and consider. Let We may partition into fundamental parallelepipeds, one corresponding to each point of. Namely, : 0 8 for ;

We call : 0 for ;, the parallelepiped associated with and denote it by. Claim : For any, and for any belonging to we have. Proof : This follows from the fact that at most four of the s have a nonzero entry in each position and two of them have a +, the other two a -. Lemma : For nonnegative vectors R and R and any R, let Then, R : (i) for any in, we have where is the -vector of all s. (ii). [ will play the role of in the brief description of the algorithm given earlier.] Proof : (i) Since is in, we have 0 and so for any, we have proving (i). For (ii), observe that for any, there is a unique such that. If, then we must have 0, so the corresponding belongs to as claimed. Lemma 3 : For, we have R and R with for all and for all Vol Vol 9

Proof : By (ii) of the lemma above, Vol is at least Vol. But is isomorphic to where s are vectors of all s of suitable dimensions as seen from the substitution. Similarly, we have that is isomorphic to. Let. Consider the set :. This is precisely the set. By the definition of, we have and. So by lemma, the volume of is at least the volume of. But is a dilation of the dimensional object by a factor of and thus has volume precisely equal to times the volume of completing the proof. Since is isomorphic to, the essential problem we have is one of sampling from a contingency polytope with uniform density. Proposition For every, we have vol. Proof : Note that if, then we have (where if one of the subscripts is 0, we take the to be zero too. Thus for each, and each,, the height of the parallelepiped in the direction of perpendicular to the previous ( if and and ( or )) coordinates is proving our proposition. 0

5 The Algorithm We now describe the algorithm which given row sums and column sums, and an error parameter! 0, draws a multiset of samples from (the set of integer contingency tables with the given row and column sums) with the following property : if : is any function, then the true mean of and the sample mean (defined below) satisfy where! 0 0 max As stated in the last section, we will first draw samples from. We let and. We assume that and throughout this section. The function often takes on values 0 or. This is so in the motivating application described by Diaconis and Efron [4] where one wants to estimate the proportion of contingency tables that have (the so-called) measure of deviation from the independent greater than the observed table. So, in this case, 0 will be for a table if its zero. The evaluation of case. measure is greater than that of the observed, otherwise, it will be for a particular table is in general simple as it is in this At the end of the section, we describe how to use the sampling algorithm to estimate.

For reasons that will become clear later, we rearrange the rows and columns so that () We assume the above holds from now on. An table in is obviously specified completely by its entries in all but the last row and column. From this, it is easy to see that is isomorphic to the polytope in (recall the notation ) which is defined as the set of satisfying the following inequalities : 0 In the above, as well as in the rest of the section, will run through and will run through unless otherwise specified. The algorithm will actually pick points in do so, the algorithm first scales each coordinate so that becomes wellrounded. Let Min The scaling is given by with near uniform density. To for and (3) for and (4) This transformation transforms the polytope to a polytope in space. Note that : is the set of that satisfy the following constraints (5)

0 Instead of working in, the algorithm will work with a larger convex set. To describe the set, first consider the convex set obtained from by discarding the constraint 0. The corresponding set in space is given by upper bounds on the row and column sums and nonnegativity constraints. The scaling above transforms to the following set : : 0 (6) We will impose the following log-concave function : : Min where (7) which is on and falls off exponentially outside. The corresponding function on is given by : Min (8) We call a cube of the form : 0 4 0 4 in space where are all integers a lattice cube (it is a cube of side 0.4); its center where 0 4 0 will be called a lattice cube center (lcc). Let be the set of lattice cubes that intersect. We will interchangeably use to denote the set of lcc s of such cubes. (Note that the lcc itself may not be in.) If is an lcc, we denote by the lattice cube of which it is the center. Note that for a particular lcc, it is easy to check whether it is in - we just round down all coordinates (to an integer multiple of 0.4) and check if the point so obtained is in. In our algorithm, each step will only modify one coordinate of ; in 3

this case, by keeping running row and column sums, we can check in time whether the new is in. A simple calculation shows : Proposition : For any lcc and any, we have 0 8. 0 8 Now we are ready to present the algorithm. There are two steps of the algorithm of which the first is the more important. Some details of execution of the algorithm are given after the algorithm. In the second step, we may reject in one of three places. If we do so, then, this trial does not produce a final sample. 4

The Algorithm The first paragraph of this section describes the input / output behavior of the algorithm. See Remark following the Theorem for the choice of 0. I. Let 0. Their values are to determined by the accuracy needed. Run the random walk for 0 steps starting at any state of. II. for 0 to 0 do the following : a. Pick 8 b. Reject if c. Let 0 with as state space described below with uniform density from. Reject with probability. [Reject means end the execution for this.] is not in. be the point in corresponding to some, then add otherwise reject.. If is in a for to the multiset of samples to be output; Random walk for step I Step I is a random walk on. Two lcc s in will be called adjacent if they differ in one coordinate by 0.4 and are equal on all the other coordinates. The transition probabilities for the random walk for step I are given by : Pr Min for adjacent lcc s in Pr 0 for not adjacent and 5

Pr Pr This is an irreducible time-reversible Metropolis Markov Chain and from the identity Pr Pr it follows that the steady state probabilities,, exist and are proportional to. An execution of the random walk produces where chosen from according to the transition probababilities. is In the next section, we will use the results of Frieze, Kannan, Polson [[5]] to show that the second largest absolute value of an eigenvalue of the Markov Chain (we denote this quantity by ) satisfies (see Theorem 3): 4 35 where the term goes to zero as. As pointed out there, for 0 and, we have 36 4. The quantity is called the spectral gap. The following theorem is proved using a result of Aldous [[]]. It involves a quantity which is the limiting probability of acceptance in the above algorithm. A sequence of lemmas (4-7) is then used to bound from below. This lower bound may be just plugged into the Theorem. See Remark below for a more practical way to estimate. Theorem The multiset of samples produced by the algorithm satisfies : if : R is any function, with 0 max, and the 6

4 mean, we have 5 806 log log 0 5 4 0 Further, 4 Remark To use the theorem, we must choose 0 to be large enough. This can be done in a fairly standard manner : suppose we want! Then, note that using Theorem 3, we can choose We may estimate 0 5 log 840 4 35 4 0 4! from the run of the algorithm which potentially fares much better than the lower bound we have given. Proof of the Theorem Let. Let : 0 be such that is the probability that given in step I of the algorithm, we pick. Letting denote the linear transformation such that for each, gives us the corresponding point in, (the exact description of here), we have 5 8 is not needed 7

8 Let. [This is the expected value of given where we count a rejected as giving us the value 0.] Let 0 0 [Note that the expectation of given the expectation of with respect to : 8 is precisely.] Consider the vol 8 0 because is on all of which contains ; also the volume of is the same for all. If we had started the chain in the steady state, the expected value of would be 0. Thus as tends to infinity, the limit of is 0. Let be the probability of accepting. Let. We also need the variance of (wrt ) denoted. 0 0 It follows from Aldous s Proposition 4. (using the fact that his than here which implies that his is at most 0 log 0 will be less ) where, is the minimum of all. [We note that Aldous lets 0 actually be a random variable to avoid the dependence on negative eigenvalues of the chain. 8

4 In our case, we will be able to get an upper bound on, the second largest absolute value of an eigenvalue, so this is not necessary. A simple modification of Aldous s argument which we do not give here implies the above inequality even though our 0 is a deterministic quantity.] It also follows by similar arguments that 0 log (9) Consider now. Recall that 0 Given 0, the process of producing each is independent. For some 0. So, we have Var 0 Var So we have, (using the inequality ) 0 300 log 0 log (0) where we have used a lower bound on which we derive now. First note that vol 8 We have, where is the union of cubes in. In a sequence of lemmas below, we show that vol. So, using Lemma 3, 4 vol 4 vol 9 4 4 vol 4 vol 4 4

We will now use equations 9 and 0 to argue the conclusion of the Theorem. To this end, let and let Pr. Then equation 9 gives us 0 log which implies 0 log 0 Now 0 gives 300 log 0 log 0 Using the inequality, we get 5 806 log 0 log 0 which gives us the theorem using proposition 3. Lemma 4 For any real number (positive, negative or zero), let : and Vol For Min, we have 0

Proof : For a real 4 vector (we assume our vector is indexed as above), define as the set of tables satisfying for ; ; for ; is the set of tables in with their last row and column dictated by. Let us denote by and by. Define Λ to be the set of nonnegative vectors satisfying : ; ; for ; ; for ; Then, we have that is nonempty iff belongs to Λ. In general, is an dimensional set and by the volume of dimensional volume. We have Vol Vol for ; Consider the linear transformation ; given by where,, we mean its for ; for ; for

It is easy to see that is a - map of Λ into Λ. Vol Λ Vol Λ Vol det Vol It is easy to check that and. This implies that the integrand in the last integral is bounded above by Vol. This function of course integrates to Vol completing the proof. Lemma 5 For any where Min 0, we have (defined in lemma 4) satisfies. Vol Proof For in the range 0 Vol, we have vol Integrating this over this range of, we get the lemma. Vol Lemma 6 : vol

4 6 Proof : From the last lemma, we have for any 0, Now the lemma follows by integration. Vol Lemma 7 Let Proof : We have : : be the union of over all lcc s. Then we have This implies that 4 lemma using the last lemma. 0 4 4 3 4 0 (). Note also that for any in, we have 6. Thus we have the min Proposition 3 5 4 Proof The number of states is at most. It is easy to check that the ratio of 5 the maximum value of to the minimum over lcc s is at most 4 completing the proof. 3

Algorithm to estimate the number of contingency tables Using the sampling algorithm, we will be able to estimate the number of contingency tables in time polynomial in the data and. We will only sketch the method here. We first estimate by sampling from the following ratio : vol vol Next, we define a sequence of contingency polytopes, each obtained from the previous one by increasing and ( ) by min until we have for. Then and the volume of is read off as the coefficient of in the easily calculated polynomial (! ). 6 Bound on the spectral gap We refer the reader to Diaconis and Strook [6] or [5] for a discussion of the eigenvalues of a reversible Markov Chain. The second largest absolute value of such a chain gives us a bound on the time to converge to the steady state as described in these references. Here, we use Theorem of [5] to bound the second largest absolute value of our chain. Note that though their theorem does not explicitly state so, the in that Theorem is the second largest absolute value of any eigenvalue of the chain. This section is a technical evaluation of the various quantities needed to plug into the expression for in [5]; we do not redefine these quantities here. 4

5 We first calculate the diameter of points in ). To do so, we let : (the largest Euclidean distance between two : Then, for any, (using the fact that 4,) So, So the diameter : : of satisfies 4 4 and 4 4 For every unit (length) vector, with 0, let be the ray 0 from the origin along. Note that the ray intersects in a line segment (since iff 0 4 5 where floor denotes componentwise floor.). Let be the length of the segment and be the length of the segment. Then there exists an such that or a such that. Assume without loss of generality the first option. Since belongs to, the vector 0 4 belongs to. So, we also have, 0 4 5 :

6 Using the fact that, we get that 4 The above implies that as required by (7) of [[5]]. Also, belongs to implies which implies by [] Similarly 0 4 is also at most 0 4. Thus is at most. This implies that we may take of (8) of [[5]] to be 5 3. Also, 4. So of [[5]] is given by (since, the step size is 0.4 here) Also, 50 3 0 We now want to bound 50 3 () 8 3 (3) of [5]. To this end, we need to prove upper and lower bounds on where is an lcc. Since is a convex function, it is clear that is a lower bound on this ratio. To get an upper bound, we use Heoffding s bounds [8] on the probability that the sum deviates from its mean of 0 and (after some complicated integration) arrive at an upper bound of 05 8 5. Plugging all this into the formula for in Theorem of [[5]], we get : 6

Theorem 3 3 4 5 4 8 5 8 3 5 8 3 35 4 3 For 0, a calculation shows that the 35 may be replaced by 36. References [] D. ALDOUS, Some inequalities for reversible Markov chains, Journal of the London Mathematical Society, ()5, pp. 564-576, 98. [] D. APPLEGATE, AND R. KANNAN, Sampling and integration of near logconcave functions, 3rd ACM Symposium on the Theory of Computing, 99. [3] A. I. BARVINOK, A polynomial time algorithm for counting integral points in polyhedra when the dimension is fixed, Proceedings of the Symposium on the Foundations of Computer Science, 993. [4] P. DIACONIS, AND B. EFRON, Testing for independence in a two-way table, Annals of Statistics, 3, pp 845-93, 985. [5] P. DIACONIS, AND A. GANGOLLI, Rectangular arrays with fixed margins, To appear in the Proceedings of the workshop on Markov Chains, Institute of Mathematics and its Applications, University of Minnesota, (994). [6] P. DIACONIS, AND D. STROOCK, Geometric bound for eigenvalues of Markov chains, Annals of Applied Probability, pp. 36-6, 99. 7

[7] P. DIACONIS, AND B. STURMFELS, Algebraic algorithms for sampling from conditional distributions, Technical Report, Department of Statistics, Stanford University (993) [8] M. E. DYER, AND A. M. FRIEZE, On the complexity of computing the volume of a polyhedron, SIAM J. Comput., 7 (988), 7 37. [9] M. E. DYER, AND A. M. FRIEZE, Computing the volume of convex bodies: a case where randomness provably helps, Proceedings of Symposia in Applied Mathematics, Vol 44, 99. [0] M. E. DYER, A. M. FRIEZE, AND R. KANNAN, A Random Polynomial-Time Algorithm for Approximating the Volume of Convex Bodies, Journal of the Association for Computing Machinery, Vol 38 No. (January 99), pp.-7. [] A. SCHRIJVER, Theory of linear and integer programming, John Wiley, Chichester, 986. [] R. P. STANLEY, Enumerative combinatorics, Vol. I, Wadsworth and Brooks/Cole, Monterey, CA, 986. [3] M. E. DYER, A. FRIEZE, R. KANNAN, A. KAPOOR, L. PERKOVIC, AND U. VAZIRANI, A mildly exponential time algorithm for approximating the number of solutions to a multidimensional knapsack problem, Combinatorics, Probability and Computing, (to appear). [4] J. A. FILL, Eigenvalue bounds on the convergence to stationarity for nonreversible Markov Chains, with an application to the exclusion process, The Annals of Applied Probability, Vol., No., pp 6-87, 99. 8

[5] A. FRIEZE, R. KANNAN AND N. POLSON, Sampling from log-concave distributions, The Annals of Applied Probability, Vol. 4, No. 3 (994) pp 8-837. [6] A. R. GANGOLLI, Convergence Bounds for Markov Chains and Applications to Sampling, Phd. Thesis, Stanford STAN-CS-9-36, 99. [7] D. GILLMAN, A Chernoff bound for random walks on expander graphs, Proceedings of the 34th Annual IEEE Conference on Foundations of Computer Science, 993. [8] W. HOEFFDING Probability Inequalities For Sums of Bounded Random Variables, American Statistical Association Journal, March 963, pp. 3-30. [9] N. KAHALE, Large deviation bounds on Markov Chains, DIMACS Technical Report, 94-39, 994. [0] L. LOVÁSZ, AND M. SIMONOVITS, Random Walks in a Convex Body and an Improved Volume Algorithm, Revised version, March 993. [] C. R. MEHTA, N. R. PATEL, A Network Algorithm for Performing Fisher s Exact Test in Contingency Tables, Journal of the American Statistical Association 78, pp. 47-434, 983. [] R.KANNAN, L.LOVÁSZ AND M. SIMONOVITS, Random Walks and an volume algorithm for convex bodies, in preparation. 5 [3] E. SENETA, Non-negative matrices and Markov Chains, Springer-Verlag, New York, 973. [4] B. STURMFELS, How to Count Integer Matrices, manuscript, December 3, 993. 9

[5] L. G. VALIANT, The complexity of computing the permanent, Theoretical Computer Science 8, pp. 89-0, 979. 30