Small ball probability and Dvoretzky Theorem

Similar documents
arxiv:math/ v1 [math.fa] 30 Sep 2004

Isoperimetry of waists and local versus global asymptotic convex geometries

An example of a convex body without symmetric projections.

A Bernstein-Chernoff deviation inequality, and geometric properties of random families of operators

Empirical Processes and random projections

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011

GAUSSIAN MEASURE OF SECTIONS OF DILATES AND TRANSLATIONS OF CONVEX BODIES. 2π) n

ALEXANDER KOLDOBSKY AND ALAIN PAJOR. Abstract. We prove that there exists an absolute constant C so that µ(k) C p max. ξ S n 1 µ(k ξ ) K 1/n

Sections of Convex Bodies via the Combinatorial Dimension

Convex inequalities, isoperimetry and spectral gap III

Weak and strong moments of l r -norms of log-concave vectors

arxiv: v1 [math.pr] 22 May 2008

Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007

High-dimensional distributions with convexity properties

Anti-concentration Inequalities

On isotropicity with respect to a measure

Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures

Rate of convergence of geometric symmetrizations

Rapid Steiner symmetrization of most of a convex body and the slicing problem

Random sets of isomorphism of linear operators on Hilbert space

A Banach space with a symmetric basis which is of weak cotype 2 but not of cotype 2

Dimensional behaviour of entropy and information

Inverse Brascamp-Lieb inequalities along the Heat equation

Another Low-Technology Estimate in Convex Geometry

Approximately Gaussian marginals and the hyperplane conjecture

KLS-TYPE ISOPERIMETRIC BOUNDS FOR LOG-CONCAVE PROBABILITY MEASURES. December, 2014

A note on the convex infimum convolution inequality

Boundedly complete weak-cauchy basic sequences in Banach spaces with the PCP

Pointwise estimates for marginals of convex bodies

Probabilistic Methods in Asymptotic Geometric Analysis.

Topics in Asymptotic Convex Geometry

Estimates for the affine and dual affine quermassintegrals of convex bodies

Integral Jensen inequality

Super-Gaussian directions of random vectors

On the constant in the reverse Brunn-Minkowski inequality for p-convex balls.

On John type ellipsoids

The Knaster problem and the geometry of high-dimensional cubes

THE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX

Super-Gaussian directions of random vectors

On Z p -norms of random vectors

Approximately gaussian marginals and the hyperplane conjecture

A CLT FOR MULTI-DIMENSIONAL MARTINGALE DIFFERENCES IN A LEXICOGRAPHIC ORDER GUY COHEN. Dedicated to the memory of Mikhail Gordin

ON BOUNDEDNESS OF MAXIMAL FUNCTIONS IN SOBOLEV SPACES

Entropy extension. A. E. Litvak V. D. Milman A. Pajor N. Tomczak-Jaegermann

Geometry of Banach spaces and sharp versions of Jackson and Marchaud inequalities

On a Generalization of the Busemann Petty Problem

The covering numbers and low M -estimate for quasi-convex bodies.

A. Iosevich and I. Laba January 9, Introduction

INEQUALITIES FOR SUMS OF INDEPENDENT RANDOM VARIABLES IN LORENTZ SPACES

Fréchet algebras of finite type

The extension of the finite-dimensional version of Krivine s theorem to quasi-normed spaces.

Houston Journal of Mathematics. c 2002 University of Houston Volume 28, No. 3, 2002

QUOTIENTS OF ESSENTIALLY EUCLIDEAN SPACES

Reconstruction from Anisotropic Random Measurements

THE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX

Subspaces and orthogonal decompositions generated by bounded orthogonal systems

Threshold Intervals under Group Symmetries

Independence of some multiple Poisson stochastic integrals with variable-sign kernels

On the isotropic constant of marginals

Optimization of a Convex Program with a Polynomial Perturbation

Optimal compression of approximate Euclidean distances

A PROPERTY OF STRICTLY SINGULAR 1-1 OPERATORS

The Lusin Theorem and Horizontal Graphs in the Heisenberg Group

Diameters of Sections and Coverings of Convex Bodies

COARSELY EMBEDDABLE METRIC SPACES WITHOUT PROPERTY A

On polynomially integrable convex bodies

On the singular values of random matrices

ON THE CONVEX INFIMUM CONVOLUTION INEQUALITY WITH OPTIMAL COST FUNCTION

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

The cocycle lattice of binary matroids

On metric characterizations of some classes of Banach spaces

Extremal Solutions of Differential Inclusions via Baire Category: a Dual Approach

APPROXIMATE ISOMETRIES ON FINITE-DIMENSIONAL NORMED SPACES

A LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE. 1.

Mass transportation methods in functional inequalities and a new family of sharp constrained Sobolev inequalities

ON THE REGULARITY OF SAMPLE PATHS OF SUB-ELLIPTIC DIFFUSIONS ON MANIFOLDS

RESTRICTED UNIFORM BOUNDEDNESS IN BANACH SPACES

Sampling and high-dimensional convex geometry

Introduction to Bases in Banach Spaces

A DECOMPOSITION THEOREM FOR FRAMES AND THE FEICHTINGER CONJECTURE

A NOTE ON POSITIVE DEFINITE NORM DEPENDENT FUNCTIONS

Comparison of Sums of Independent Identically Distributed Random Variables

Neighborly families of boxes and bipartite coverings

On the Equivalence Between Geometric and Arithmetic Means for Log-Concave Measures

ESTIMATES FOR MAXIMAL SINGULAR INTEGRALS

Asymptotic shape of the convex hull of isotropic log-concave random vectors

POSITIVE DEFINITE FUNCTIONS AND MULTIDIMENSIONAL VERSIONS OF RANDOM VARIABLES

(convex combination!). Use convexity of f and multiply by the common denominator to get. Interchanging the role of x and y, we obtain that f is ( 2M ε

Into a Hilbert Space

arxiv:math/ v1 [math.fa] 21 Mar 2000

A COUNTEREXAMPLE TO AN ENDPOINT BILINEAR STRICHARTZ INEQUALITY TERENCE TAO. t L x (R R2 ) f L 2 x (R2 )

Szemerédi-Trotter theorem and applications

On the mean connected induced subgraph order of cographs

Citation Osaka Journal of Mathematics. 41(4)

A Note on Jackknife Based Estimates of Sampling Distributions. Abstract

A RELATIONSHIP BETWEEN THE DIRICHLET AND REGULARITY PROBLEMS FOR ELLIPTIC EQUATIONS. Zhongwei Shen

The Rademacher Cotype of Operators from l N

Concentration inequalities: basics and some new challenges

TOPICS. P. Lax, Functional Analysis, Wiley-Interscience, New York, Basic Function Theory in multiply connected domains.

CONTINUITY OF CP-SEMIGROUPS IN THE POINT-STRONG OPERATOR TOPOLOGY

Transcription:

Small ball probability and Dvoretzy Theorem B. Klartag, R. Vershynin Abstract Large deviation estimates are by now a standard tool in Asymptotic Convex Geometry, contrary to small deviation results. In this note we present a novel application for a small deviations inequality to a problem that is related to the diameters of random sections of high dimensional convex bodies. Our results imply an unexpected distinction between the lower and upper inclusions in Dvoretzy Theorem. Introduction In probability theory, the large deviation theory (or the tail probabilities) and the small deviation theory (or the small ball probabilities) are in a sense two complementary directions. The large deviation theory, which is a more classical direction, sees to control the probability of deviation of a random variable X from its mean M, i.e. one loos for upper bounds on Prob( X M > t). The small deviation theory sees to control the probability of X being very small, i.e. it loos for upper bounds on Prob( X < t). There are a number of excellent texts on large deviations, see e.g. recent boos [DZ] and [dh]. A recent exposition of the state of the art in small deviation theory can be found in [LS]. A modern powerful approach to large deviations is via the celebrated concentration of measure phenomenon. One of the early manifestations of this idea was V. Milman s proof of Dvoretzy Theorem in the 970s. Recall that Dvoretzy Theorem entails that any n-dimensional convex body has a section of dimension c log n which is approximately a Euclidean ball. Since Milman s proof, the concentration of measure philosophy plays a major role in geometric functional analysis and in many other areas. A recent boo by M. Ledoux [L] gives an account of many ramifications of this method. A standard instance of the concentration of measure phenomenon is the case of a Lipschitz function School of Mathematics, Institute for Advanced Study, Princeton, NJ 08540, USA. Supported by NSF grant DMS-0298 and the Bell Companies Fellowship. Department of Mathematics, University of California, One Shields Ave, Davis, CA 9566, USA. Supported by NSF grant 040032 and the Sloan Research Fellowship.

on the unit Euclidean sphere S n. In view of the geometric applications, we shall state it for a norm on R n, or equivalently for its unit ball, which is a centrally-symmetric convex body K R n. We equip the sphere S n with the unique rotation invariant probability measure σ. Two parameters are responsible for many geometric properties of the convex body K; the maximal and the average values of the norm on the sphere S n : b = b(k) = sup x, M = M(K) = x dσ(x). () x S n S n The concentration of measure inequality, which appears e.g. in the first pages of [MS] states that the norm is close to its mean M on most of the sphere. For any t >, σ { x S n } ( : x M > tm < exp ct 2 ) (2) where = (K) = n ( ) 2 M(K). b(k) Here and thereafter the letters c, C, c, c, c, c 2 etc. denote some positive universal constants, whose values may be different in various appearances. The symbol denotes equivalence of two quantities up to an absolute constant factor, i.e. a b if ca b Ca for some absolute constants c, C > 0. The concentration of measure inequality can of course be interpreted as a large deviation inequality for the random variable x, and the connection to probability theory becomes even more sound when one recalls an analogous inequality for Gaussian measures, see [L]. The quantity (K) plays a crucial role in high dimensional convex geometry, as it is the critical dimension in Dvoretzy Theorem. We will call this dimension (K) the Dvoretzy dimension. Milman s proof of Dvoretzy Theorem [M] (see also [MS, 5.8]) provides accurate information regarding the dimension of the almost spherical sections of K. Milman s argument shows that if l < c(k), then with probability larger than e c l, a random l-dimensional subspace E G n,l satisfies c M (Dn E) K E C M (Dn E), (3) where M = M(K), D n denotes the unit Euclidean ball in R n, and the randomness is induced by the unique rotation invariant probability measure on the grassmanian G n,l of l-dimensional subspaces in R n. The Dvoretzy dimension (K) was proved in [MS2] to be the exact critical dimension for a random section to satisfy (3), in the following strong sense. If a random l-dimensional subspace E G n,l satisfies (3) with probability larger than, say, n, then necessarily l < C(K). Thus a random section of dimension l < c(k) is close to Euclidean with high probability, and a random section of dimension l > C(K) is typically far from Euclidean. These arguments completely clarify the question of the dimensions in which random sections of a given convex body are close to Euclidean. Once b(k) and M(K) are 2

calculated, the behavior of a random section is nown. For instance, Dvoretzy dimension of the cube is log n, while the cross polytope K = {x R n : xi } has Dvoretzy dimension as large as (K) n. In this note we investigate Dvoretzy Theorem from a different direction, which does not involve the standard large deviations inequality (2). The second named author conjectured that a phenomenon similar to the concentration of measure should also occur for the small ball probability, and he proved a weaer statement. The conjecture has been recently proved by R. Lata la and K. Olesziewitz [LO], using the solution to the B-conjecture by Cordero, Fradelizi and Maurey [CFM]: Theorem. (Small ball probability). For every 0 < ε < 2, where c > 0 is a universal constant. σ { x S n : x < εm } < ε c(k) This theorem is related to the small ball probability (as a direction of the probability theory) in exactly the same way as the concentration of measure is related to large deviations. Here we apply Theorem. to study questions arising from Dvoretzy Theorem. We show that for some purposes, it is possible to relax the Dvoretzy dimension (K), replacing it by a quantity independent of the Lipschitzness of the norm (which is quantified by the Lipschitz constant b(k)). Precisely, we wish to replace (K) by d(k) = min{ log σ { x S n : x 2 M}, n}, where log stands for the natural logarithm. Selecting t = 2 in the concentration of measure inequality (2), we conclude that d(k) must be at least of the same order of magnitude as Dvoretzy dimension (K): d(k) C(K). The small ball Theorem. indeed holds with d(k) (this is a part of the argument of Lata la and Olesziewicz, reproduced below). The resulting inequality can be viewed as Kahane-Khinchine type inequality for negative exponents: Proposition.2 (Negative moments of a norm). Assume that 0 < l < cd(k). Then ) cm < x l l dσ(x) < CM S n where c, C > 0 are universal constants. For positive exponents, this inequality was proved in [LMS]: for 0 < < c(k), ( cm < x dσ(x) < CM. (4) S n 3

For negative exponents < l < 0, inequality (4) follows from results of Guédon [G] that generalize Lovász-Simonovits inequality [LoSi]. Proposition.2 extends (4) to the range [ cd(k), c(k)] (which of course includes the range [ c(k), c(k)]). In Proposition.2, x can be regarded as the radius of the one-dimensional section of the body K. Combining this with the recent inequality for diameters of sections due to the first named author [K], we are able to lift the dimension of the section and thus compute the average diameter of l-dimensional sections of any centrally-symmetric convex body K. Theorem.3 (Diameters of random sections). Assume that 0 < l < cd(k). Select a random l-dimensional subspace E G n,l. Then with probability larger than e c l, K E C M (Dn E). (5) Furthermore, l c M < diam(k E) l dµ(e) < C G n,l M (6) where µ is the unique rotation invariant probability measure on G n,l, and c, c, c, C, C > 0 are universal constants. The relation between Theorem.3 and Dvoretzy Theorem is clear. We show that for dimensions which may be much larger than (K), the upper inclusion in Dvoretzy Theorem (3) holds with high probability. This reveals an intriguing point in Dvoretzy Theorem. Milman s proof of Dvoretzy Theorem focuses on the left-most inclusion in (3). Once it is proved that the left-most inclusion in (3) holds with high probability, the right-most inclusion follows almost automatically in his proof. Furthermore, Milman-Schechtman s argument [MS2] implies in fact that the left-most inclusion does not hold (with large probability) for dimensions larger than the Dvoretzy dimension. The reason that a random l-dimensional section is far from Euclidean when l > c(k) is that a typical section does not contain a sufficiently large Euclidean ball. In comparison, we observe that the upper inclusion in (3) holds for a much wider range of dimensions. There are cases, such as the case of the cube, where the Dvoretzy dimension satisfies (K) log n, while d(k) is a polynomial in n. Hence, while sections of the cube of dimension n c are already contained in the appropriate Euclidean ball (for any fixed c <, independent of n), only when the dimension is log n, the sections start to fill from inside, and an isomorphic Euclidean ball is observed. The case of the cube is contained, using different terminology, in [LMT]. The fact that d(k) is typically larger than (K) is a little unexpected. It implies that the correct upper bound for random sections of a convex body appears 4

sometimes in much larger dimensions than those for which we have the lower bound. In the past decade, diameters of random lower-dimensional sections of convex bodies attracted a considerable amount of attention, see in particular [GM, GM2, GMT]. Theorem.3 is a significant addition to this line of results. It implies that diameters of random sections are equivalent for a wide range of dimensions starting from dimension one, when the random diameter simply equals M(K), and up to the critical dimension d(k). See also Remar right after the proof of Theorem.3. Remar. The proof shows that d(k) can be further relaxed in all of our results. For any fixed u >, the parameter d(k) can be replaced by d u (K) = min{ log σ { x S n : x u M}, n}. The rest of the paper is organized as follows. In Section 2 we discuss the negative moments of the norm, proving Proposition.2 and Theorem. by the Lata la-olesziewicz argument. In Section 3 we perform the dimension lift and compute the average diameters of random sections, proving Theorem.3. 2 Concentration of measure and the small ball probability We begin by proving Proposition.2. This proposition is a reformulation of the small ball probability conjecture due to the second named author. It was recently deduced by R. Lata la and K. Olesziewicz [LO] from the B-conjecture proved by Cordero, Fradelizi and Maurey [CFM]. We will reproduce the Lata la- Olesziewicz argument here. We start with a standard and well-nown lemma, on the close relation between the uniform measure σ on the sphere S n and the standard gaussian measure γ on R n. For the reader s convenience, we include its proof. Lemma 2.. For every centrally-symmetric convex body K, 2 σ(sn 2 K) γ( nk) σ(s n 2K) + e cn where c > 0 is a universal constant. Proof. We will use the following two estimates on the Gaussian measure of the Euclidean ball, γ(2 nd n ) > 2, γ( 2 nd n ) < e cn. 5

The first estimate is simply Chebychev s inequality, and the second follows from standard large deviation inequalities, e.g. Cramer s Theorem [V]. Since K is star-shaped, γ( nk) γ(2 nd n nk) γ(2 nd n ) σ (2 ns n nk) where σ denotes the probability rotation invariant measure on the sphere 2 ns n, 2 σ(sn 2 K). This proves the lower estimate in the lemma. For the upper estimate, note that no points of nk can lie outside both the ball 2 nd n and the positive cone generated by 2 Sn nk. Adding the two measures together, we obtain γ( nk) γ( 2 nd n ) + σ 2 ( 2 ns n nk) where σ 2 denotes the probability rotation invariant measure on the sphere 2 ns n, This completes the proof. e cn + σ(s n 2K). Proof of Proposition.2. As usual, K will denote the unit ball of the norm. The B-conjecture, proved in [CFM], asserts that the function t γ(e t K) is log-concave. This means that for any a, b > 0 and 0 < λ <, γ ( a λ b λ K ) γ (ak) λ γ (bk λ. (7) Let Med = Med(K) be the median of the norm on the unit sphere S n. By Chebychev s inequality, Med 2M(K). Set L = Med nk. According to Lemma 2., γ(2l) 2 σ(sn Med K) 4 by the definition of the median. On the other hand, again by Lemma 2., γ( 8 L) σ(sn 4Med K) + e cn = σ(x S n : x 4 Med) + e cn (9) σ(x S n : x 2 M(K)) + e cn e d(k) + e c n < 2e Cd(K) because d(k) n. We may assume that ε < e 3, and apply (7) for a = ε, b = 2, λ = 3. This yields log ε 3 γ(εl) log(/ε) γ(2l) 3 log(/ε) γ (ε 3 log(/ε) 2 3 log(/ε) L ) γ( 8 L). (8) 6

Combining this with (8) and (9), we obtain that γ (εl) 8e C d(k) log ε 8ε cd(k) < (c ε) cd(k) and according to Lemma 2. we can transfer this to the spherical measure, obtaining σ(x S n : x < εm) < (Cε) cd(k). By integration by parts, this yields that for any 0 < l < cd(k) 0, ( S n x/m l dσ(x) l C, which implies the left hand side of the inequality in Proposition.2. The right hand side follows easily by Hölder s inequality. By Chebychev s inequality, Proposition.2 yields the desired tail inequality for the small ball probability: Corollary 2.2 (The small ball probability). For every 0 < ε < 2, σ { x S n : x < εm } < ε cd(k) < ε c (K), where c, c > 0 are universal constants. Theorem. is contained in Corollary 2.2. Let us give some interpretation of the expression in Proposition.2. For a subspace E R n, let S(E) = S n E and σ E be the unique rotation invariant probability measure on the sphere S(E). We will use the fact that Vol(K) = Vol(D n ) S x n dσ(x). The n volume radius of a -dimensional set T is defined as ( Vol(T ) v.rad.(t ) =. Vol(D ) Thus ( /n v.rad.(k) = x n dσ(x). S n By the rotation invariance of all the measures (as in [K]), we conclude that x dσ(x) = x dσ E (x) dµ(e) S n G n, S(E) = v.rad.(k E) dµ(e), (0) G n, where, as before, µ is the unique rotation invariant probability measure on G n,. Thus Proposition.2 asymptotically computes the average volume radius of random sections. This perfectly fits the estimates for diameters of sections in [K], to be applied next. 7

3 Diameters of random sections In this section we prove the main result of the paper, Theorem.3. We regard x as the radius of the one-dimensional section spanned by x; thus Proposition.2 is an asymptotically sharp bound on the diameters of random one-dimensional sections. Theorem.3 extends this bound to -dimensional sections, for all up to the critical dimension d(k). We start with a dimension lift, which is based on the low M estimate, Proposition 3.9 in [K] (the case λ = 2 there). Here we estimate the L norm, rather than only the tail probability as in Proposition 3.9 in [K]. Proposition 3. (Dimension lift for diameters of sections). Let 0 < n. Then for any integer < 0 /4, G n, diam(k E) dµ(e) ( ) 2 0 CM(K) x 0 dσ(x). S n Remar. Note the special normalization in Proposition 3. (compare with Proposition 3.9 in [K]). In fact, as follows from the proof, for any λ > 0, the right hand side of Proposition 3. may be replaced by +λ C(λ)M(K) (S λ 0 x 0 dσ(x) n at the cost of increasing 0 (simply replace Cauchy-Schwartz inequality with the appropriate Hölder inequality in the proof). However, we cannot have λ = 0, as is demonstrated by the example where K = R n. The average diameter of onedimensional sections of K is zero, while the diameter of any section of dimension larger than one is infinity. Thus there cannot be any formal dimension lift unless one has an extra factor which must be infinity for flat bodies. Such a factor is M(K) = S n x dσ(x). In order to prove Theorem.2 we need a standard lemma on the stability of the average norm M. We are unaware of a reference for the exact statement we need (a similar result appears e.g. in Lemma 6.6 of [M2]), so a proof is provided. The average norm on a subspace E G n, is denoted by M E = S(E) x dσ E(x). Lemma 3.2. For every norm on R n and every integer 0 < < n, cm < G n, (M E ) 2 dµ(e) where c, C > 0 are universal constants. 2 < CM () 8

Proof. The left hand size inequality in () follows easily from Hölder s inequality. In the proof of the right hand side inequality, we will use a variant of Raz s argument (see [R, MW]). We normalize so that M =. Let X,.., X be independent random vectors, distributed uniformly on S n. It is well-nown that a norm of a random vector on the sphere has a subgaussian tail (e.g. [LMS, 3.]. It actually follows from (2) above): E exp (s X i ) < exp ( cs 2) for all i and all s > which by independence implies ( ) E exp s ( ) Cs 2 X i < exp i= for s >. Using Chebychev s inequality and optimizing over s (e.g. [MS, 7.4]), we obtain { } Prob X i > Ct < exp ( t 2 ) for t >. (2) i= Let E be the linear span of X,.., X. Then E is distributed uniformly in G n, (up to an event of measure zero). Since for any two events one has Prob(A) Prob(B) Prob(B A), we conclude that { } i= X i > ct Prob Prob {M E > 2ct} Prob { i= X i > ct ME > 2ct }. (3) The numerator in (3) is bounded by (2). To bound the denominator from below, note that X i < C M E pointwise for all i; This is a consequence of a simple comparison inequality for the Gaussian analogs of M and M E (see e.g. [MS, 5.9]). Let us fix a subspace E G n,. Note that, conditioning on E = span{x,..., X }, each of the vectors X i is distributed uniformly in S(E). Next, we estimate the probability P E = Prob via Chebychev s inequality as ( M E = E X i Hence P E i= { span{x,.., X } = E c for every E G n,. Thus, denominator in (3) Prob { i= i= X i > ME 2 ) X i > M E 2 = Prob{E G n, ; M E > 2ct} min P E c. E G n, } span{x,.., X } = E C M E P E + M E 2 ( P E). } M E > 2ct E G n, ;M E>2ct P E dµ(e) 9

Combining this with (2) and (3) we get Prob {M E > 2ct} < c e t2 < e Ct2 for t >. By integration by parts we obtain the desired estimate. Proof of Proposition 3.. By Hölder inequality, the right hand side increases with 0, hence we may assume that 0 = 4. We shall rely on the main result in [K], which claims that for any centrally-symmetric convex body T R n, and for all 0 < l < n, v.rad.(t ) > C G n, v.rad.(t E) l diam(t E) n l dµ(e) /n. (4) We are going to apply (4) to T = K E, for subspaces E G n,0. Denote by G E, the grassmanian of all -dimensional subspaces of E, equipped with the unique rotational invariant probability measure. Then by (0), (4) and the rotational invariance of all measures, v.rad.(k E) 2 diam(k E) 2 dµ(e) E G n, = v.rad.(k F ) 2 diam(k F ) 2 dµ E (F )dµ(e) E G n,4 F G E, C E G 0 v.rad.(k E) 0 dµ(e) = C 0 x 0 dσ(x). n,0 S n Also, by the Cauchy-Schwartz inequality, diam(k E) dµ(e) E G n, E G n, v.rad.(k E) 2 diam(k E) 2 dµ(e) 2 dµ(e) E G n, v.rad.(k E) 2 We will use the standard inequality v.rad.(k E) M E, which follows directly from Hölder inequality. Then, diam(k E) dµ(e) E G n, ( ) 2 ( 0 C x 0 dσ(x) S n E Gn, and the proposition follows by Lemma 3.2. (M E ) 2 dµ(e) 2 2. 0

Proof of Theorem.3. It is sufficient to prove (6), since (5) follows by Chebychev s inequality. The left hand side inequality is clear. According to Proposition 3. and Proposition.2, the right hand side inequality of (6) follows, as l ( ) 2 C diam(k E) l dµ(e) < CM(K) C G n,l M(K) M(K). Remars.. (Optimality). The estimate l < cd(k) in Theorem.3 is essentially optimal, for any centrally-symmetric convex K R n. Indeed, suppose that l > d u (K) for some u. Then, ( S n x l dσ(x) l > u M e M. (5) Since we always have diam(k E) 2 v.rad.(k E), then by (5) and (0), G n,l diam(k E) l dµ(e) Thus, (6) cannot hold for l > d u (K) when u. l M. 2. (Example of the cube). Suppose K = B n is a cube, the unit ball of l n, and let us estimate d log n u(k). It is well-nown that c n < M(B n ) < log n C n. According to Lemma 2., we may equivalently carry out our computations in the Gaussian setting. Now, for t > 0, γ(t n log nb n ) = Prob{ X t log n} = ( 2Φ(t log n)) n i= where X N(0, ) is a standard normal random variable, and Φ(s) = s For s >, a crude estimate gives e cs2 < Φ(s) < e c2s2. Thus, when t > 0 log n, exp( c n c 2 t2 ) < γ(t log nb n ) < exp( c n c2t2 ). Therefore, by Lemma 2., c n c u 2 < d u (B n ) < c 2n c 2 u 2 for any u >. Theorem.3 implies that random sections of this (polynomial!) dimension d u (B n ) have a diameter M(B n ). 2π e u2 2 du.

3. (Example of the l p ball). Consider K = Bp n, the unit ball of lp n, for some fixed p <. In this case, the conclusion of Theorem.3 is wellnown. For p 2, we now that (Bp n ) > cn (e.g. [MS, 5.4]), hence also d(bp n ) > cn and the conclusion of the theorem follows from the classical Dvoretzy Theorem. In the case where 2 < p <, we have Bp n cp M(Bp n)dn, and thus the conclusion of Theorem.3 is obvious in this case (with constants depending on p), and d cp (B n p ) > C p n, for some constants c p, C p depending only on p. References [CFM] D.Cordero-Erausquin, M.Fradelizi, B. Maurey, The (B) conjecture for the Gaussian measure of dilates of symmetric convex sets and related problems, J. Funct. Anal. 24, no. 2 (2004), 40 427. [DZ] A.Dembo, O.Zeitouni, Large deviations techniques and applications. Second edition. Applications of Mathematics (New Yor), 38, Springer-Verlag, New Yor, 998 [GM] A.Giannopoulos, V.Milman, On the diameter of proportional sections of a symmetric convex body, International Mathematics Research Notices (997), 5 9 [GM2] A.Giannopoulos, V.Milman, Mean width and diameter of proportional sections of a symmetric convex body, J. Reine Angew. Math. 497 (998), 3 39 [GMT] A.Giannopoulos, V.Milman, A. Tsolomitis, Asymptotic formulas for the diameter of sections of symmetric convex bodies, preprint [G] O.Guédon, Kahane-Khinchine type inequalities for negative exponent, Mathematia 46 (999), 65 73 [dh] F. den Hollander, Large deviations. Fields Institute Monographs, 4. American Mathematical Society, Providence, RI, 2000 [K] B. Klartag, A geometric inequality and a low M estimate. Proc. Amer. Math. Soc. 32 (2004), 269 2628 [LO] R. Lata la, K.Olesziewitz, Small ball probability estimates in terms of width, preprint. http://www.arxiv.org/abs/math.pr/050268 [L] M.Ledoux, The concentration of measure phenomenon. Mathematical Surveys and Monographs, 89. American Mathematical Society, Providence, RI, 200 [LS] W.V.Li, Q.-M. Shao, Gaussian processes: inequalities, small ball probabilities and applications. Stochastic processes: theory and methods, 533 597, Handboo of Statistics, 9, North-Holland, Amsterdam, 200 [LoSi] L.Lovász, M.Simonovits, Random wals in a convex body and an improved volume algorithm, Random Structures Algorithms 4 (993), 359 42 2

[LMS] A. E. Litva, V.D. Milman, V. D., G. Schechtman, Averages of norms and quasi-norms. Math. Ann. 32 (998), 95 24 [LMT] A. E. Litva, P. Maniewicz, N. Tomcza-Jaegermann, Randomized isomorphic Dvoretzy theorem, C. R. Math. Acad. Sci. Paris 335, no. 4, (2002), 345 350. [M] V.D. Milman, A new proof of the theorem of A. Dvoretzy on sections of convex bodies, Func. Anal. Appl. 5 (97), 28 38 (translated from Russian) [M2] V.D. Milman, Isomorphic symmetrizations and geometric inequalities, Geometric aspects of functional analysis (986/87), Israel seminar, Lect. Notes in Math. Springer 37 (988), 07 3 [MS] V.D. Milman, G. Schechtman, Asymptotic theory of finite-dimensional normed spaces. Lecture Notes in Mathematics 200, Springer-Verlag, Berlin (986) [MS2] V.D. Milman, G. Schechtman, Global versus local asymptotic theories of finite-dimensional normed spaces. Due Math. J. 90 (997), 73 93 [MW] V.D. Milman, R. Wagner, Some remars on a lemma of Ran Raz, Geometric aspects of functional analysis. Proceedings of the Israel seminar (GAFA) 200-2002. Berlin: Springer. Lect. Notes Math. 807 (2003), 58 68 [R] [V] R. Raz, Exponential separation of quantum and classical communication complexity. Annual ACM Symposium on Theory of Computing, Atlanta, GA (999), 358 367 S.R.S. Varadhan, Large Deviations and Applications, CBMS-NSF regional conference series in applied mathematics, vol. 40 (993) 3