On large deviations of sums of independent random variables

Similar documents
arxiv: v1 [math.pr] 12 Jun 2012

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012

Edgeworth expansions for a sample sum from a finite set of independent random variables

1 Introduction. 2 Measure theoretic definitions

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

A Note on the Central Limit Theorem for a Class of Linear Systems 1

A Note on Extended Binomial Coefficients

Self-normalized Cramér-Type Large Deviations for Independent Random Variables

On large deviations for combinatorial sums

Stochastic Models (Lecture #4)

EXACT CONVERGENCE RATE AND LEADING TERM IN THE CENTRAL LIMIT THEOREM FOR U-STATISTICS

SOME CONVERSE LIMIT THEOREMS FOR EXCHANGEABLE BOOTSTRAPS

On rate of convergence in distribution of asymptotically normal statistics based on samples of random size

Two hours. Statistical Tables to be provided THE UNIVERSITY OF MANCHESTER. 14 January :45 11:45

A generalization of Cramér large deviations for martingales

WLLN for arrays of nonnegative random variables

Asymptotic results for empirical measures of weighted sums of independent random variables

Review: mostly probability and some statistics

18.175: Lecture 8 Weak laws and moment-generating/characteristic functions

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Chapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

RELATIVE ERRORS IN CENTRAL LIMIT THEOREMS FOR STUDENT S t STATISTIC, WITH APPLICATIONS

A NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES

Chapter 4. Continuous Random Variables 4.1 PDF

CHAPTER 3: LARGE SAMPLE THEORY

Zhishui Hu 1, John Robinson 2 and Qiying Wang 2

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Cramér large deviation expansions for martingales under Bernstein s condition

Formulas for probability theory and linear models SF2941

1 Solution to Problem 2.1

ON THE MOMENTS OF ITERATED TAIL

The Central Limit Theorem: More of the Story

Probability and Measure

1.1 Review of Probability Theory

STAT 200C: High-dimensional Statistics

Chapter 3: Random Variables 1

A NOTE ON A DISTRIBUTION OF WEIGHTED SUMS OF I.I.D. RAYLEIGH RANDOM VARIABLES

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Asymptotic behavior for sums of non-identically distributed random variables

Gaussian vectors and central limit theorem

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

Section 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University

Asymptotic results for empirical measures of weighted sums of independent random variables

Concentration inequalities and tail bounds

Uses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).

On the absolute constants in the Berry Esseen type inequalities for identically distributed summands

arxiv: v1 [math.pr] 13 Jul 2011

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

18.175: Lecture 15 Characteristic functions and central limit theorem

On probabilities of large and moderate deviations for L-statistics: a survey of some recent developments

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

Proving the central limit theorem

S1. Proofs of the Key Properties of CIATA-Ph

Bell-shaped curves, variance

Math 104: Homework 7 solutions

STAT 801: Mathematical Statistics. Inversion of Generating Functions = 1. e ikt P (X = k). X n = [nx]/n

Zeros of lacunary random polynomials

Lecture 3 - Expectation, inequalities and laws of large numbers

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Course 212: Academic Year Section 1: Metric Spaces

Discrete Probability Refresher

A log-scale limit theorem for one-dimensional random walks in random environments

A NON-PARAMETRIC TEST FOR NON-INDEPENDENT NOISES AGAINST A BILINEAR DEPENDENCE

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

A Class of Fractional Stochastic Differential Equations

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

On variable bandwidth kernel density estimation

Lecture 1 Measure concentration

Complete q-moment Convergence of Moving Average Processes under ϕ-mixing Assumption

1 Presessional Probability

We introduce methods that are useful in:

x log x, which is strictly convex, and use Jensen s Inequality:

Refining the Central Limit Theorem Approximation via Extreme Value Theory

Module 3. Function of a Random Variable and its distribution

MOMENT CONVERGENCE RATES OF LIL FOR NEGATIVELY ASSOCIATED SEQUENCES

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs

BASICS OF PROBABILITY

ELEC546 Review of Information Theory

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

MAS113 Introduction to Probability and Statistics

A Note on the Strong Law of Large Numbers

Chap 2.1 : Random Variables

Product measure and Fubini s theorem

6.1 Moment Generating and Characteristic Functions

Yunhi Cho and Young-One Kim

EXISTENCE THEOREMS FOR SOLUTIONS OF DIFFERENTIAL EQUATIONS OF NON-INTEGRAL ORDER*

CSCI-6971 Lecture Notes: Probability theory

Lecture 22: Final Review

The Azéma-Yor Embedding in Non-Singular Diffusions

Asymptotic Behavior of a Controlled Branching Process with Continuous State Space

Probability Transforms with Elliptical Generators

Chapter 3: Random Variables 1

Some Aspects of Universal Portfolio

Large deviations for weighted random sums

Concentration Inequalities

Random Variables (Continuous Case)

The Moment Method; Convex Duality; and Large/Medium/Small Deviations

Uniformly Accurate Quantile Bounds Via The Truncated Moment Generating Function: The Symmetric Case

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

Transcription:

On large deviations of sums of independent random variables Zhishui Hu 12, Valentin V. Petrov 23 and John Robinson 2 1 Department of Statistics and Finance, University of Science and Technology of China, Hefei, Anhui 2326, China, 2 School of Mathematics and Statistics, University of Sydney, NSW 26, Australia, 3 Faculty of Mathematics and Mechanics, St. Petersburg University, Stary Peterhof, St. Petersburg 19854, Russia Extensions of some limit theorems are proved for tail probabilities of sums of independent identically distributed random variables satisfying the one-sided or two-sided Cramér s condition. The large deviation x-region under consideration is broader than in the classical Cramér s theorem, and the estimate of the remainder is uniform with respect to x. The corresponding asymptotic expansion with arbitrarily many summands is also obtained. Running head On large deviations. Keywords Limit theorems; sums of independent random variables; large deviation; Cramér s condition; asymptotic expansions. Mathematics Subject Classification Primary 6F1; Secondary 6G5, 62E2. 1 Introduction and results Let X 1, X 2, be a sequence of independent random variables with a common distribution V (x), mean and variance 1. Assume that the following Address correspondence to John Robinson, School of Mathematics and Statistics, University of Sydney, NSW 26, Australia; E-mail: johnr@maths.usyd.edu.au 1

one-sided Cramér s condition is satisfied: Ee hx 1 < for h < H and some H >. (1.1) Introduce a sequence of conjugate i.i.d. random variables X 1h, X 2h, with the d.f. V h (x) = 1 x e hy dv (y), R(h) where h < H and R(h) = e hy dv (y). The distribution V h (x) has the same set of the growth points as V (x). Put where h < H. We have m(h) = EX 1h = m(h), VarX 1h = σ 2 (h), (1.2) xdv h (x) = 1 R(h) σ 2 (h) = E(X1h 2 ) (EX 1h) 2 = R (h) ( R R(h) (h) R(h) xe hx dv (x) = R (h) d log R(h) =, R(h) dh ) 2 dm(h) = dh. Denote by F n (x) the d.f. of n k=1 X k/ n. Then we have the following theorem. Theorem 1.1 Let X 1, X 2, be a sequence of i.i.d. random variables with mean, variance 1 and Suppose that conditions (1.1) and are satisfied and let < H < H. Then EX 4 1 <. (1.3) lim sup Ee itx 1 < 1 (1.4) t 1 F n (x) = 1 exp{n log R(h) nhm(h)}m(hσ(h) n)[1 + O(1/ n)] (1.5) 2

holds uniformly in x in the area < x m(h ) n. Here M(y) = e y2 /2 y e t2 /2 dt is the Mills ratio and h is the unique positive root of the equation m(h) = x/ n. Let us introduce a corollary to Theorem 1.1 that corresponds to the two-sided Cramér s condition: Ee hx 1 < for H < h < H and some H >. (1.6) Corollary 1.1 Let X 1, X 2, be a sequence of i.i.d. random variables satisfying conditions (1.4) and (1.6), and let EX 1 =, EX 2 1 = 1. If < H < H, then relation (1.5) holds uniformly in x in the area < x m(h ) n, where m(h) is defined in (1.2), and h is the unique positive root of the equation m(h) = x/ n. This result differs from the classical Cramér s theorem on large deviations (see, for example, Theorem 5.23 in Petrov(1995)), in particular, by a stronger uniform estimate of the remainder; the remainder term in Cramér s theorem is O(x/ n), and the area is 1 < x = o( n)(n ) instead of the broader one < x m(h ) n. Under the one-sided Cramér s condition (1.1) a series of limit theorems for probabilities of large deviations of sums of independent identically distributed random variables was proved for special x-regions in Petrov(1965), where references on earlier works are given. Contrary to condition (1.1), the two-sided Cramér s condition (1.6) implies the existence of moments of X 1 of all orders. The one-sided condition (1.1) imposes restrictions on the behavior of the distribution of the random variable X 1 only on the positive half-line. Our next theorem is an improvement of Theorem 5 in Petrov(1965). In this theorem, we replace the condition (1.4) by the condition that X 1 is nonlattice. Theorem 1.2 Let X 1, X 2, be a sequence of i.i.d. nonlattice random variables with mean, variance 1 and E X 1 3 <. Suppose that condition (1.1) is satisfied and let < H < H. Then 1 F n (x) = 1 exp{n log R(h) nhm(h)}m(hσ(h) n)[1 + o(1)] (1.7) 3

holds uniformly in < x m(h ) n, where M( ) is defined in Theorem 1.1 and h is the unique positive root of the equation m(h) = x/ n. From Theorem 1.2, we can get the following corollary directly. This corollary is also shown in Theorem A of Höglund (1979). Corollary 1.2 Let X 1, X 2, be a sequence of i.i.d. nonlattice random variables satisfying condition (1.6), and let EX 1 =, EX 2 1 = 1. If < H < H, then relation (1.7) holds uniformly in x in the area < x m(h ) n, where h is the unique positive root of the equation m(h) = x/ n. Remark 1. Since EX 1 =, EX 2 1 = 1, we have m() = and σ2 () = 1. Under the condition (1.4), X 1 is nonlattice and so is X 1h. Thus σ 2 (h) = Var(X 1h ) >. This together with the fact σ 2 (h) = m (h) implies m(h) is strictly increasing. So m(h) > for all < h < H and the root of the equation m(h) = x/ n is unique for < x m(h ) n. More information about the behavior of m(h) can be found in Petrov(1965). Let γ ν (h) be the cumulant of order ν of the random variable X 1h = (X 1h m(h))/σ(h), i.e. γ ν (h) = 1 d ν i ν dt ν lneeitx 1h. t= In particular, γ 1 (h) =, γ 2 (h) = 1, γ 3 (h) = E(X 1h )3, if E X 1 3 < and condition (1.1) is satisfied. Define the so-called Esscher functions: for λ >, put B k (λ) = λ exp( λx x 2 /2)H k (x)dx, k =, 1,, where H k (x) is the Chebyshev-Hermite polynomial of degree k. In particular, B (λ) = () 1/2 λm(λ), B 1 (λ) = λ(b (λ) () 1/2 ), B 2 (λ) = λ 2 (B (λ) () 1/2 ). For more about the Esscher functions, see, for instance, Section 2.1 of Jensen(1995). We have the following extension of Theorem 1.1. 4

Theorem 1.3 Let < H < H and let k 3 be an integer. Suppose that E X 1 k+1 <. Then under the conditions of Theorem 1.1, { B ( nhσ(h)) 1 F n (x) = exp{n log R(h) nhm(h)} nhσ(h) k 2 ν ( ) + n ν/2 1 γm+2 (h) km B ν+2s ( nhσ(h)) k ν=1 m=1 m! (m + 2)! nhσ(h) } + O(n (k 1)/2 ) (1.8) holds uniformly in x in the area < x m(h ) n, where h is the unique positive root of the equation m(h) = x/ n, the inner summation is extended over all non-negative integer solutions (k 1, k 2,, k ν ) of the equation k 1 + 2k 2 + + νk ν = ν and s = k 1 + k 2 + + k ν. Relation (1.8) is an extension of (2.2.6) in Jensen(1995), where the density function of X 1 is required. This paper is organized as follows. A smoothness condition on the conjugate variables is studied in Section 2. Proofs of theorems and corollaries in Section 1 are offered in Section 3. Throughout the paper we shall use A, A 1, A 2,... to denote absolute positive constants whose values may differ at each occurrence. 2 Smoothness for conjugate random variables Results on large deviations frequently use a smoothness condition on the conjugate variables. We will show that it is enough to assume the usual C-condition of Cramér. Theorem 2.1 Let X be a random variable with distribution function V (x) such that for h < H, R(h) = Ee hx = For any < c < C, if then, for any positive H < H, sup h H c t <C e hx dv (x) <. (2.1) sup Ee itx < 1, (2.2) c t <C sup Ee (h+it)x /Ee hx < 1. (2.3) 5

Theorem 2.1 can be obtained by using a method similar to that in the proof of Lemma 7.3 in Zhou and Jing (26). Here we give another proof which discloses the inner relationship between X and the corresponding conjugate random variable. The proof will depend on the following lemma which is closely modelled on a result of Weber and Kokic (1997). Lemma 2.1 Let X be a random variable with distribution function V (x). For any < c < C, the following are equivalent: (i) there exist ε > and δ > such that for all real y and all t satisfying c t < C, the set A(y, t, ε) = {x : xt y 2kπ > ε, for all integers k} has P(X A(y, t, ε)) > δ; (ii) sup c t <C Ee itx < 1. Proof. We first show that (i) implies (ii). Without loss of generality take ε < π/2 and assume that (i) holds. Given any t with c t < C, choose, for any x, integers k(t, x) such that z(t, x) = xt k(t, x) [, ). Then for all y [, ) and all t from the area c t < C, P(d(z(t, X) y) > ε) > δ, where d(z) = min{ z, z +, z }. So Ee itx = = e i(z(t,x) y) dv (x) cos(z(t, x) y)dv (x) + i Choose y = y(t) [, ) such that sin(y(t)) E sinz(t, X) = cos(y(t)) E cos z(t, X), then E sin(z(t, X) y(t)) =. Now for all t with c t < C, E cos(z(t, X) y(t)) sin(z(t, x) y)dv (x). (2.4) P(d(z(t, X) y(t)) ε) + P(d(z(t, X) y(t)) > ε)cos ε = 1 (1 cos ε)p(d(z(t, X) y(t)) > ε) 1 δ(1 cos ε). (2.5) 6

Next, we choose ỹ(t) [, ) such that ỹ(t) y(t) = π. Similarly to (2.5), we have E cos(z(t, X) y(t)) = E cos(z(t, X) ỹ(t)) 1 δ(1 cos ε). (2.6) Hence by (2.4), (2.5) and (2.6), sup Ee itx 1 δ(1 cos ε) c t <C and (ii) follows. Now we will show (ii) implies (i) by showing that complement of (i) implies that for any η >, sup Ee itx > 1 η. c t <C The complement of (i) may be stated as follows: for all ε > and δ >, there exist y = y(ε, δ) and t = t(ε, δ) with c t < C such that the set has A c (y, t, ε) = {x : xt y 2kπ ε for some integer k} P(X A c (y, t, ε)) 1 δ. (2.7) Write Ee itx = e iy dv (x) + (e itx e i(y+k(t,x)) )dv (x) A c (y,t,ε) A c (y,t,ε) + e itx dv (x) A(y,t,ε) := J 1 + J 2 + J 3. Then Ee itx J 1 J 2 J 3. (2.8) Using (2.7) and the inequality e iz 1 z, we have J 1 1 δ, J 3 δ, (2.9) and J 2 A c (y,t,ε) (e i(tx y k(t,x)) 1)dV (x) ε. (2.1) 7

So it follows from (2.8)-(2.1) that Ee itx 1 2δ ε > 1 η, by choosing δ > and ε > such that 2δ + ε < η. The proof of Lemma 2.1 is complete. Proof of Theorem 2.1. We prove Theorem 2.1 by showing that the complement of (2.3) implies that sup Ee itx = 1. (2.11) c t <C The complement of (2.3) is: there exists H with < H < H such that sup sup h H c t <C Ee (h+it)x /Ee hx = 1. So there exists h with h H such that sup Ee (h+it)x /Ee hx = 1. (2.12) c t <C Let V h (x) = 1 x e hu dv (u) R(h ) and let X h be the conjugate r.v. with distribution function V h (x). Then it follows from (2.12) and Lemma 2.1 that for all ε > and δ >, there exist y and t with c t < C such that P(X h A(y, t, ε)) δ, where A(y, t, ε) is defined in Lemma 2.1. For any δ >, choose B > such that P( X B) < δ /2. Then P(X A(y, t, ε)) P(X A(y, t, ε) ( B, B)) + δ /2 = R(h ) e hx dv h (x) + δ /2 A(y,t,ε) ( B,B) R(h )e h B P(X h A(y, t, ε)) + δ /2 R(h )e h B δ + δ /2 < δ, 8

by choice of δ < δ e h B /(2R(h )). So we have shown that the complement of (i) holds for X: for all ε > and δ >, there exist y and t with c t < C such that P(X A(y, t, ε)) < δ. This, from Lemma 2.1, implies (2.11). The proof of Theorem 2.1 is complete. 3 Proof of results Proof of Theorem 1.1. From Remark 1, we know that σ 2 (h) > for h < H. Since σ 2 (h) is continuous on [, H) and σ 2 () = 1, there exist positive constants c 1 and c 2, not depending on h, such that Also from Remark 1, the equation c 1 σ 2 (h) c 2 for h H. (3.1) m(h) = x/ n (3.2) has a unique positive root h (, H ] for any positive x m(h ) n. Conditions (1.1) and (1.3) imply that Indeed, for < p 4, sup E X 1h p <, < p 4. (3.3) h H sup E X 1h p = h H sup x p e hx dv (x) h H x p dv (x) + sup E X 1 p + c p 3 ec 3H + { c3 h H c 3 + c 3 e δx dv (x) < } x p e hx dv (x) if H < δ < H and c 3 is a positive constant, not depending on h, such that x p e H x e δx for x c 3. Define by F nh (x) the d.f. of n k=1 X kh / n, where Xkn = (X kh m(h))/σ(h). By the same argument as that in the proof of Theorem 5.23 in Petrov(1995), we have 1 F n (x) = e Λn(h) e nhσ(h)y df nh (y), (3.4) 9

where Λ n (h) = n log R(h) nhm(h) and h is the unique positive root of the equation (3.2). By a theorem of Osipov (see, for example, Theorem 1 of Chapter 6 in Petrov (1975) or Theorem 5.18 in Petrov (1995)), we have where F nh (y) = Φ(y) + P 1(y) n + Q n (y) (3.5) P 1 (y) = (1/(6 ))(1 y 2 )e y2 /2 E(X1h )3, { Q n (y) A n 1/2 (1 + y ) 3 E X1h 3 I( X1h n(1 + y )) +n 1 (1 + y ) 4 E X1h 4 I( X1h < n(1 + y )) +n 6 (1 + y ) 4( ) n } sup Ee itx 1h + 1/(2n) t T(h) (3.6) for all y and n, where T(h) = 1/(12E X 1h 3 ). By (3.1) and (3.3), we have 1/(12σ(h)E X 1h 3 ) c for < h H, where c is a positive constant not depending on h. Thus in the last summand on the right side of (3.6) we can replace X 1h by X 1h and the area t T(h) by the area t c. It follows from (3.4) and (3.6) that where 1 F n (x) = e Λn(h)( I 1 + I 2 + I 3 ), (3.7) I 1 = 1 e nhσ(h)y y 2 /2 dy, I 2 = 1 e nhσ(h)y dp 1 (y), n I 3 = Since for nhσ(h) 1, e nhσ(h)y dq n (y). hσ(h) ni 1 = 1 e y y2 /(2nh 2 σ 2 (h)) dy 1 e y y2 /2 dy >, and for < nhσ(h) < 1, I 1 1 e y y2 /2 dy >, 1

we have I 1 = 1 M(hσ(h) n) A 1 1 + nhσ(h). (3.8) where Note that Therefore 6 ni 2 = E(X1h )3 (y 3 3y)e nhσ(h)y y 2 /2 dy = E(X 1h )3 B 3 ( nhσ(h))/( nhσ(h)), B 3 (λ) = (λ/ ) B 3 (λ)/λ (y 3 3y)e λy y2 /2 dy. (y 3 + 3y)e y2 /2 dy <, for λ >. Note that B 3 (λ) = O(λ 1 ) as λ + (see, for instance, Lemma 2.1.2 of Jensen(1995)). Thus B 3 (λ)/λ A 2 /(1 + λ) holds uniformly for λ >. Then it follows from (3.3) that ( ) 1 I 2 = O (3.9) n(1 + nhσ(h)) uniformly in x in the area < h H. It remains to estimate I 3. By (3.6) and the argument below (3.6), we have with I 3 = Q n () Q n (y)de nhσ(h)y, (3.1) { Q n (y) A E X1h 4 n 1 + n 6( ) n } sup Ee itx 1h + 1/(2n), (3.11) t c for all y, n and < h H. Condition (1.4) implies that sup t c Ee itx 1 < 1 (see, for instance, p14 in Petrov (1995)). Thus by Theorem 2.1, we have sup sup Ee itx 1h < 1. (3.12) <h H t c 11

This, together with (3.3), (3.1) and (3.11), implies that uniformly in x in the area < h H. It follows from (3.7) that I 3 = O(1/n) (3.13) 1 F n (x) = e Λn(h) I 1 (1 + 1 ) (I 2 + I 3 ). I 1 By (3.1), (3.8)-(3.9) and (3.13), we have I 2 /I 1 = O(n 1/2 ), I 3 /I 1 = O(hn 1/2 ) uniformly in x in the area < x m(h ) n. Since for any positive x m(h ) n, the equation m(h) = x/ n has the unique positive root h H, we have I 3 /I 1 = O(n 1/2 ). Therefore 1 F n (x) = 1 exp{n log R(h) nhm(h)}m(hσ(h) n)[1 + O(1/ n)] uniformly in x in the area < x m(h ) n. Proof of Corollary 1.1. This follows immediately from Theorem 1.1. Proof of Theorem 1.2. Similarly to the proof of (3.3), we obtain sup E X 1h p <, < p 3. (3.14) h H Following the mainstream of the proof of Theorem 1.1, we only need to show that F nh (y) Φ(y) P ( 1(y) 1 ) = o n n holds uniformly for all y and < h H, where P 1 (y) is defined below (3.5). This follows from the proof of Theorem 1 and 2 of 42 in Gnedenko and Kolmogorov (1968) with some modification, by using (3.1), (3.14) and Theorem 2.1. Proof of Corollary 1.2. This follows immediately from Theorem 1.2. 12

Proof of Theorem 1.3. Let H m (y) be the Chebyshev-Hermite polynomial of degree m defined by the equality and let H m (y) = ( 1) m e y2 /2 dm dy me y2 /2, P ν (y) = 1 e y2 /2 H ν+2s 1 (y) ν m=1 1 k m! ( ) γm+2 (h) km (3.15) (m + 2)! where the summation is extended over all non-negative integer solutions (k 1, k 2,, k ν ) of the equation k 1 + 2k 2 + + νk ν = ν and s = k 1 + k 2 + + k ν. Similarly to the proof of (3.3), sup E X 1h p <, < p k + 1. h H Using a theorem of Osipov (see, for example, Theorem 1 of Chapter 6 in Petrov (1975) or Theorem 5.18 in Petrov (1995)), we have k 2 F nh (y) = Φ(y) + P ν (y)n ν/2 + Q n (y), ν=1 where Q n (y) { c(k) n (k 2)/2 (1 + y ) k E X1h k I( X1h n(1 + y )) +n (k 1)/2 (1 + y ) (k+1) E X1h k+1 I( X1h < n(1 + y )) +n k(k+1)/2 (1 + y ) (k+1)( ) n } sup Ee itx 1h + 1/(2n) t T(h) for all y and n, where T(h) = 1/(12E X1h 3 ) and c(k) is a positive constant depending only on k. By the argument below (3.6) and a simple calculation, we have Q n (y) c(k) {n k 1 2 E X 1h k+1 + n k(k+1) 2 Similarly to the proof of (3.7), we have ( sup t c 1 ) n } Ee itx 1h +. 2n 1 F n (x) = e Λn(h)( k 2 L 1 + L 2ν n ν/2 + L 3 ), (3.16) ν=1 13

where Λ n (h) = n log R(h) nhm(h) and L 1 = L 2ν = L 3 = 1 e nhσ(h)y y 2 /2 dy, e nhσ(h)y dp ν (y), ν = 1,, k 2, e nhσ(h)y dq n (y). Using Lemma 2.1.1 in Jensen(1995) and noting that we have and L 2ν = 1 ν = d(h ν 1 (y)e y2 /2 )/dy = H ν (y)e y2 /2, ν 1, L 1 = B ( nhσ(h))/( nhσ(h)), (3.17) m=1 1 ν = ν m=1 1 k m! ( ) γm+2 (h) km (m + 2)! exp( nhσ(h)y)d(h ν+2s 1 (y)e y2 /2 ) m=1 1 k m! ( ) 1 γm+2 (h) km H ν+2s (y)e nhσ(h)y y 2 /2 dy k m! (m + 2)! ( ) γm+2 (h) km B ν+2s ( nhσ(h)), (3.18) (m + 2)! nhσ(h) where the summations have the same meaning as that in (3.15). Similarly to the proof of (3.13), we have L 3 = O(n (k 1)/2 ) (3.19) uniformly in x in the area < h H. Then (1.8) follows from (3.16)-(3.19). The proof of Theorem 1.3 is complete. References Cramér, H. (1938). Sur un nouveau théorème-limite de la théorie des probabilités. Actual. Sci. Indust. (Paris), 736: 5-23. 14

Gnedenko, B. V., Kolmogorov, A. N. (1968). Limit distributions for sums of independent random variables, (2nd edn). Addison-Wesley, Reading, MA. Höglund, T. (1979). A unified formulation of the central limit theorem for small and large deviations from the mean. Z. Wahrsch. Verw. Gebiete 49: 15-117. Jensen, J. L. (1995). Saddlepoint approximations. Oxford University Press, New York. Petrov, V. V. (1965). On the probabilities of large deviations for sums of independent random variables. Theory Probab. Appl. 1: 287-298. Petrov, V. V. (1975). Sums of independent random variables. Springer, New York. Petrov, V. V. (1995). Limit theorems of probability theory. Oxford University Press, New York. Weber, N. C., Kokic, P. N. (1997). On Cramér s condition for Edgeworth expansions in the finite population case. Theory of Stochastic Processes 3: 468-474. Zhou, W., Jing, B. Y. (26). Tail probability approximations for Student s t- statistics. To appear in Probab. Theory Relat. Fields. 15