Normal Approximation for Non-linear Statistics Using a Concentration Inequality Approach
|
|
- Sabrina Fields
- 5 years ago
- Views:
Transcription
1 Normal Approximation for Non-linear Statistics Using a Concentration Inequality Approach Louis H.Y. Chen 1 National University of Singapore and Qi-Man Shao 2 Hong Kong University of Science and Technology, University of Oregon,and Zhejiang University Abstract. Let T be a general sampling statistic that can be written as a linear statistic plus an error term. Uniform and non-uniform Berry-Esseen type bounds for T are obtained. The bounds are best possible for many known statistics. Applications to U-statistic, multi-sample U-statistic, L-statistic, random sums, and functions of non-linear statistics are discussed. Nov. 15, 2005 AMS 2000 subject classification: Primary 62E20, 60F05; secondary 60G50. Key words and phrases. Normal approximation, uniform Berry-Esseen bound, non-uniform Berry- Esseen bound, concentration inequality approach, nonlinear statistics, U-statistics, multi-sample U- statistics, L-statistic, random sums, functions of non-linear statistics. 1 Research is partially supported by grant R at the National University of Singapore. 2 Research is partially supported by NSF grant DMS and grant R at the National University of Singapore. 1
2 1 Introduction Let X 1, X 2,..., X n be independent random variables and let T : T (X 1,..., X n be a general sampling statistic. In many cases T can be written as a linear statistic plus an error term, say T W +, where W g i (X i, : (X 1,..., X n T W and g i : g n,i are Borel measurable functions. Typical cases include U-statistics, multi-sample U- statistics, L-statistics, and random sums. Assume that (1.1 E(g i (X i 0 for i 1, 2,..., n, and E(gi 2 (X i 1. It is clear that if 0 in probability as n, then we have the following central limit theorem (1.2 sup P (T z Φ(z 0 z where Φ denotes the standard normal distribution function, provided that the Lindeberg condition holds: ε > 0, Egi 2 (X i I( g i (X i > ε 0. If in addition, E p < for some p > 0, then by the Chebyshev inequality, one can obtain the following rate of convergence: (1.3 sup z P (T z Φ(z sup P (W z Φ(z + 2(E p 1/(1+p. z The first term on the right hand side of (1.3 is well-understood via the Berry-Esseen inequality. For example, using Stein s method, Chen and Shao (2001 obtained ( n (1.4 sup P (W z Φ(z 4.1 Egi 2 (X i I( g i (X i > 1 + z E g i (X i 3 I( g i (X i 1. However, the bound (E p 1/(1+p is in general not sharp for many commonly used statistics. Many authors have worked towards obtaining better Berry-Esseen bounds. For example, sharp Berry-Esseen bounds have been obtained for general symmetric statistics in van Zwet (1984 and Friedrich (1989. An Edgeworth expansion with remainder O(n 1 for symmetric statistics was proved by Bentkus, Götze and Zwet (1997. The main purpose of this paper is to establish uniform and non-uniform Berry-Esseen bounds for general nonlinear statistics. The bounds are best possible for many known statistics. Our proof is 2
3 based on a randomized concentration inequality approach to bounding P (W + z P (W z. Since proofs of uniform and non-uniform bounds for sums of independent random variables can be proved via Stein s method [8], which is much neater and simpler than the traditional Fourier analysis approach, this paper provides a direct and unifying treatment towards the Berry-Esseen bounds for general non-linear statistics. This paper is organized as follows. The main results are stated in next section, five applications are presented in Section 3 and an example is given in Section 4 to show the sharpness of the main results. Proofs of the main results are given in Section 5, while proofs of other results including Example 4.1 are postponed to Section 6. Throughout this paper, C will denote an absolute constant whose value may change at each appearance. The L p norm of a random variable X is denoted by X p, i.e., X p (E X p 1/p for p 1. 2 Main results Let {X i, 1 i n, T, W, be defined as in Section 1. In the following theorems, we assume that (1.1 is satisfied. Put (2.1 β and let δ > 0 satisfy (2.2 E g i (X i 2 I( g i (X i > 1 + E g i (X i 3 I( g i (X i 1 E g i (X i min(δ, g i (X i 1/2. Theorem 2.1 For each 1 i n, let i be a random variable such that X i and ( i, W g i (X i are independent. Then (2.3 sup P (T z P (W z 4δ + E W + z for δ satisfying (2.2. In particular, we have E g i (X i ( i (2.4 and (2.5 sup P (T z P (W z 2β + E W + z sup P (T z Φ(z 6.1β + E W + z E g i (X i ( i E g i (X i ( i. 3
4 Next theorem provides a non-uniform bound. Theorem 2.2 For each 1 i n, let i be a random variable such that X i and ( i, {X j, j i are independent. Then for δ satisfying (2.2 and for z R 1, (2.6 P (T z P (W z γ z + e z /3 τ where (2.7 (2.8 γ z P ( > ( z + 1/3 + P ( g i (X i > ( z + 1/3 + P ( W g i (X i > ( z 2/3P ( g i (X i > 1, τ 21δ g i (X i 2 i 2. In particular, if E g i (X i p < for 2 < p 3, then (2.9 P (T z Φ(z P ( > ( z + 1/3 + C( z + 1 p( 2 + g i (X i 2 i 2 + E g i (X i p. A result similar to (2.5 has been obtained by Friedrich (1989 for g i E(T X i using the method of characteristic function. Our proof is direct and simpler and the bounds are easier to calculate. The non-uniform bounds in (2.6 and (2.9 for general non-linear statistics are new. Remark 2.1 Assume E g i (X i p < for p > 2. Let (2.10 ( 2(p 2 p 2 δ (p 1 p 1 E g i (X i p 1/(p 2. Then (2.2 is satisfied. This follows from the inequality (2.11 min(a, b a (p 2p 2 a p 1 (p 1 p 1 b p 2 for a 0 and b > 0. Remark 2.2 If β 1/2, then (2.2 is satisfied with δ β/2. 4
5 Remark 2.3 Let δ > 0 be such that Egi 2 (X i I( g i (X i > δ 1/2. Then (2.2 holds. In particular, if X 1, X 2,, X n are independent and identically distributed (i.i.d. random variables and g i g 1, then (2.2 is satisfied with δ c 0 / n, where c 0 is a constant such that E( ng 1 (X 1 2 I( ng 1 (X 1 > c 0 1/2. Remark 2.4 In Theorems 2.1 and 2.2, the choice of i is flexible. For example, one can choose i (X 1,..., X i 1, 0, X i+1,..., X n or i (X 1,..., X i 1, ˆX i, X i+1,..., X n, where { ˆX i, 1 i n is an independent copy of {X i, 1 i n. The choice of g i is also flexible. It can be more general than g i (x E(T X i x, which is commonly used by others in the literature. Remark 2.5 Let X 1,..., X n be independent normally distributed random variables with mean zero and variance 1/n, and let W, T and be as in Example 4.1. Then (2.12 E W + E X i 3 + E X i ( (X 1,..., X i,..., X n (X 1,..., 0,..., X n Cε 2/3 for (1/ε 4/3 n 16(1/ε 4/3. This together with (4.5 shows that the bound in (2.4 is achievable. Moreover, the term E g i (X i ( i in (2.4 can not be dropped off. 3 Applications Theorems 2.1 and 2.2 can be applied to a wide range of different statistics and provide bounds of the best possible order in many instances. To illustrate the usefulness and the generality of these results, we give five applications in this section. The uniform bounds refine many existing results with specifying absolute constants, while the non-uniform bounds are new for many cases. 3.1 U-statistics Let X 1, X 2,..., X n be a sequence of independent and identically distributed (i.i.d. random variables, and let h(x 1,..., x m be a real-valued Borel measurable symmetric function of m variables, where m (2 m < n may depend on n. Consider the Hoeffding (1948 U-statistic ( n 1 U n h(x m i1,..., X im. 1i 1 <...<i mn 5
6 The U-statistic elegantly and usefully generalizes the notion of a sample mean. Numerous investigations on the limiting properties of the U-statistics have been done during the last few decades. A systematic presentation of the theory of U-statistics was given in Koroljuk and Borovskikh (1994. We refer the study on uniform Berry-Esseen bound for U-statistics to Filippova (1962, Grams and Serfling (1973, Bickel (1974, Chan and Wierman (1977, Callaert and Janseen (1978, Serfling (1980,Van Zwet (1984, and Friedrich (1989. One can also refer to Wang, Jing and Zhao (2000 on uniform Berry-Esseen bound for studentized U-statistics. Applying Theorems 2.1 and 2.2 to the U-statistic, we have Theorem 3.1 Assume that Eh(X 1,..., X m 0 and σ 2 Eh 2 (X 1,..., X m <. Let g(x E(h(X 1, X 2,..., X m X 1 x and σ1 2 Eg2 (X 1. Assume that σ 1 > 0. Then ( n ( 1 n sup P U n z P g(x i z (1 + 2(m 1σ z mσ 1 nσ1 (m(n m + 1 1/2 + c 0 (3.1, σ 1 n where c 0 is a constant such that Eg 2 (X 1 I( g(x 1 > c 0 σ 1 σ 2 1 /2. If in addition E g(x 1 p < for 2 < p 3, then (3.2 sup P z ( n U n z Φ(z (1 + 2(m 1σ mσ 1 (m(n m + 1 1/ E g(x 1 p σ 1 n (p 2/2 σ p 1 and for z R 1, ( n (3.3 P U n z Φ(z mσ 1 9mσ 2 (1 + z 2 (n m + 1σ1 2 CE g(x 1 p + (1 + z p n (p 2/2 σ p e z /3 m 1/2 σ (n m + 1 1/2 σ 1 Moreover, if E h(x 1,..., X m p < for 2 < p 3, then for z R 1, ( n (3.4 P U n z Φ(z mσ 1 Cm1/2 E h(x 1,..., X m p (1 + z p (n m + 1 1/2 σ p 1 CE g(x 1 p + (1 + z p n (p 2/2 σ p. 1 Note that the error in (3.1 is of order O(n 1/2 only under the assumption of finite second moment of h. The result appears not known before. The uniform bound given in (3.2 is not new, however, the specifying constant for general m is new. Finite second moment of h is not the weakest assumption for the uniform bound. Friedrich (1989 obtained an order of O(n 1/2 when E h 5/3 < which is necessary for the bound as shown by Bentkus, Götze and Zitikis (
7 then (3.5 For the non-uniform bound, Zhao and Chen (1983 proved that if m 2, E h(x 1, X 2 3 <, P ( n U n z Φ(z An 1/2 (1 + z 3 mσ 1 for z R 1, where the constant A does not depend on n and z but the moment of h. Clearly, (3.4 refines Zhao and Chen s result specifying the relationship of the constant A with the moment condition. After we finished proving Theorem 3.1, Wang (2001 informed the second author that he also obtained (3.4 for m 2 and p 3. Remark 3.1 (3.3 implies that (3.6 P ( n U n z Φ(z mσ 1 Cm 1/2 σ 2 (1 + z 3 (n m + 1 1/2 σ 2 1 CE g(x 1 p + (1 + z p n (p 2/2 σ p 1 for z ((n m + 1/m 1/2. For z > ((n m + 1/m 1/2, the bound like (3.6 can be easily obtained by using the Chebyshev inequality. On the other hand, if (3.6 holds for any z R 1, then it appears necessary to assume E h(x 1,..., X m p <. 3.2 Multi-sample U-statistics Consider k independent sequences {X j1,..., X jnj of i.i.d. random variables, j 1,..., k. Let h(x jl, l 1,..., m j, j 1,..., k be a measurable function symmetric with respect to m j arguments of the j-th set, m j 1, j 1,..., k. Let θ Eh(X jl, l 1,..., m j, j 1,..., k. The multi-sample U-statistic is defined as { k U n ( nj m j 1 h(xjl, l i j1,..., i jmj, j 1,..., k where n (n 1,..., n k and the summation is carried out over all 1 i j1 <... < i jmj n j, n j 2m j, j 1,..., k. Clearly, U n is an unbiased estimate of θ. The two-sample Wilcoxon statistic and the two-sample ω 2 -statistic are two typical examples of the multi-sample U-statistics. Without loss of generality, assume θ 0. For j 1,..., k, define ( h j (x E h(x 11,..., X 1m1 ;... ; X k1,..., X kmk X j1 x 7
8 and let σ 2 j Eh2 j (X j1 and σ 2 n k m 2 j n j σ 2 j. A uniform Berry-Esseen bound with order O((min 1jk n j 1/2 for the multi-sample U-statistics was obtained by Helmers and Janssen (1982 and Borovskich (1983 (see, [Koroljuk and Borovskich (1994, pp ]. Next theorem refines their results. Theorem 3.2 Assume that θ 0, σ 2 : Eh 2 (X 11,..., X 1m1 ;... ; X k1,..., X kmk < and max 1jk σ j > 0. Then for 2 < p 3 (3.7 ( sup P z σ 1 n U n z Φ(z (1 + 2σ σ n k m 2 j n j σ p n k m p j n p 1 j E h j (X j1 p and for z R 1 (3.8 ( P σ 1 n U n z Φ(z 9σ 2 (1 + z 2 σ 2 n C + (1 + z p σ p n ( k k m 2 j n j e z /3 σ σ n m p j E h j(x j1 p. n p 1 j k m 2 j n j 3.3 L-statistics Let X 1,..., X n be i.i.d. random variables with a common distribution function F, and let F n be the empirical distribution function defined by F n (x n 1 n Let J(t be a real-valued function on [0, 1] and define T (G for non-decreasing measurable function G. Put σ 2 I(X i x for x R 1. xj(g(xdg(x J(F (sj(f (tf (min(s, t(1 F (max(s, tdsdt and g(x (I(x s F (sj(f (sds. 8
9 The statistic T (F n is called an L-statistic (see [Serfling (1980, Chapter 8]. Uniform Berry-Esseen bounds for L-statistic for smoothing J were given by Helmers (1977, and Helmers, Janssen and Serfling (1990. Applying Theorems 2.1 and 2.2 yields the following uniform and non-uniform bounds for L-statistic. Theorem 3.3 Let n 4. Assume that EX 2 1 < and E g(x 1 p < for 2 < p 3. If the weight function J(t is Lipschitz of order 1 on [0, 1], that is, there exists a constant c 0 such that (3.9 then (3.10 and (3.11 J(t J(s c 0 t s for 0 s, t 1 sup P ( nσ 1 (T (F n T (F z Φ(z (1 + 2c 0 X E g(x 1 p z nσ n (p 2/2 σ p P ( nσ 1 (T (F n T (F z Φ(z 9c 2 0 EX2 1 (1 + z 2 nσ 2 + C ( c0 X 1 2 (1 + z p + E g(x 1 p nσ n (p 2/2 σ p 3.4 Random sums of independent random variables with non-random centering Let {X i, i 1 be i.i.d. random variables with EX i µ and Var(X i σ 2, and let {N n, n 1 be a sequence of non-negative integer-valued random variables that are independent of {X i, i 1. Assume that EN 2 n < and N n EN n Var(Nn d. N(0, 1. Then by Robbins (1948, Nn X i (EN n µ σ 2 EN n + µ 2 Var(N n d. N(0, 1. This is a special case of limit theorems for random sums with non-random centering. This kind of problems arises in the study, for example, of Galton-Watson branching processes. We refer to Finkelstein, Kruglov and Tucker (1994 and references therein for recent developments in this area. As another application of our general result, we give a uniform Berry-Esseen bound for the random sum. 9
10 Theorem 3.4 Let {Y i, i 1 be i.i.d. non-negative integer-valued random variables with EY i ν and Var(Y i τ 2. Put N n n Y i. Assume that E X i 3 < and that {Y i, i 1 and {X i, i 1 are independent. Then (3.12 ( N n sup P X i nµν x n(νσ 2 + τ 2 µ 2 x Φ(x Cn 1/2( τ 2 ν 2 + E X 1 3 σ 3 + σ µ. ν 3.5 Functions of non-linear statistics Let X 1, X 2,..., X n be a random sample and ˆΘ n ˆΘ n (X 1,..., X n be a weak consistent estimator of an unknown parameter θ. Assume that ˆΘ n can be written as ˆΘ n θ + 1 ( n g i (X i + n where g i are Borel measurable functions with Eg i (X i 0 and n Eg 2 i (X i 1, and : n (X 1,..., X n 0 in probability. Let h be a real-valued function differentiable in a neighborhood of θ with g (θ 0. Then, it is known that n(h( ˆΘn h(θ h (θ d. N(0, 1 under some regularity conditions. When ˆΘ n is the sample mean, the Berry-Esseen bound and Edgeworth expansion have been well studied (see Bhattacharya and Ghosh (1978. The next theorem shows that the results in Section 3 can be extended to functions of non-linear statistics. Theorem 3.5 Assume that h (θ 0 and δ(c 0 sup x θ c0 h (x < for some c 0 > 0. Then for 2 < p 3, (3.13 where W n g i (X i. ( n(h( sup P ˆΘ n h(θ z h (θ (1 + c 0δ(c 0 h (θ ( E W + z Φ(z E g i (X i ( i +6.1 E g i (X i p + 4 c 2 0 n + 2E 4.4c3 p 0 δ(c 0 + c 0 n1/2 h (θ n (p 2/2, 10
11 4 An example In this section we give an example to show that the bound of (2.4 in Theorem 2.1 is achievable. Moreover, the term E g i (X i ( i in (2.4 can not be dropped off. The example also provides a counter-example to a result of Shorack (2000 and of Bolthausen and Götze (1993. Example 4.1 Let X 1,..., X n be independent normally distributed random variables with mean zero and variance 1/n. Define W X i, T : T ε W ε W 1/2 + ε c 0 and T W ε W 1/2 + ε c 0, where c 0 E( W 1/2 2/π 0 x 1/2 e x2 /2 dx. Let { ˆX i, 1 i n be an independent copy of {X i, 1 i n and define (4.1 α 1 E (X 1,..., X i,..., X n (X 1,..., n ˆX i,..., X n. Then ET 0 and for 0 < ε < 1/64 and n (1/ε 4 (4.2 (4.3 (4.4 where C is an absolute constant. P (T ε c 0 Φ(ε c 0 ε 2/3 /6, E W + E 7ε, E + E X i 3 + α Cε (4.5 Clearly, (4.2 implies that sup P (T ε z Φ(z ε 2/3 /6. z A result of Shorack (2000 (see Lemma , p. 261, [22] states that for any random variables W and, (4.6 sup z P (W + z Φ(z sup P (W z Φ(z + 4E W + 4E. z Another result which is in Theorem 2 of Bolthausen and Götze (1993 states that if ET 0, then (4.7 ( sup P (T z Φ(z C E + z where C is an absolute constant and α is defined in (4.1. E g i (X i 3 + α, In view of (4.3, (4.4 and (4.5, the result of Shrock and of Bolthausen and Götze can be shown to lead to a contradiction. 11
12 5 Proof of Main Theorems In this section we prove Theorems 2.1 and 2.2 and Remarks 2.1 and 2.2. Proof of Theorem 2.1. (2.5 follows from (2.4 and (1.4. When β > 1/2, (2.4 is trivial. For β 1/2, (2.4 is a consequence of (2.3 and Remark 2.2. Thus, we only need to prove (2.3. Note that (5.1 P (z W z P (T z P (W z P (z W z +. It suffices to show that (5.2 P (z W z + 4δ + E W + E g i (X i ( i and (5.3 where δ satisfies (2.2. Let (5.4 Let P (z W z 4δ + E W + E g i (X i ( i /2 δ for w z δ, f (w w 1 2 (2z + for z δ w z + + δ, /2 + δ for w > z + + δ. ξ i g i (X i, ˆMi (t ξ i {I( ξ i t 0 I(0 < t ξ i, M i (t E ˆM i (t, ˆM(t n ˆM i (t, M(t E ˆM(t. Since ξ i and f i (W ξ i are independent for 1 i n and Eξ i 0, we have (5.5 E{W f (W E{ξ i (f (W f (W ξ i 1in + E{ξ i (f (W ξ i f i (W ξ i 1in : H 1 + H 2. Using the fact that ˆM(t 0 and f (w 0, we have (5.6 H 1 1in 1in 0 E {ξ i f (W + tdt ξ i { E f (W + t ˆM i (tdt 12
13 { E { E f (W + t ˆM(tdt f (W + t ˆM(tdt t δ { E I(z W z + 1in H 1,1 H 1,2, t δ ˆM(tdt { E I(z W z + ξ i min(δ, ξ i where H 1,1 P (z W z + Eη i, H 1,2 E η i Eη i, η i ξ i min(δ, ξ i. 1in 1in By (2.2, Hence (5.7 Eη i 1/2. 1in H 1,1 (1/2P (z W z +. By the Cauchy-Schwarz inequality, (5.8 H 1,2 (E( η i Eη i 2 1/2 1in ( 1in Eη 2 i 1/2 δ. As to H 2, it is easy to see that Hence (5.9 Combining (5.5, (5.7, (5.8 and (5.9 yields f (w f i (w i /2 i /2. H 2 (1/2 E ξ i ( i. { P (z W z + 2 E W f (W + δ + (1/2 E ξ i ( i E W + 2δE W + 2δ + E ξ i ( i 4δ + E W + E ξ i ( i. 13
14 This proves (5.2. Similarly, one can prove (5.3 and hence Theorem 2.1. Proof of Theorem 2.2. First, we prove (2.9. For z 4, (2.9 holds by (2.5. For z > 4, consider two cases. Case 1. n E g i (X i p > 1/2. By the Rosenthal (1970 inequality, we have (5.10 Hence P ( W > ( z 2/3 P ( W > z /6 ( z /6 p E W p C( z + 1 p{( n p/2 Egi 2 n (X i + E g i (X i p n C( z + 1 p E g i (X i p. P (T z Φ(z P ( > ( z + 1/3 + P ( W > ( z 2/3 + P ( N(0, 1 > z n P ( > ( z + 1/3 + C( z + 1 p E g i (X i p, which shows that (2.9 holds. Case 2. n E g i (X i p 1/2. Similar to (5.10, we have P ( W g i (X i > ( z 2/3 C( z + 1 p{( n p/2 Egj 2 n (X j + E g j (X j p C( z + 1 p and hence γ z P ( > ( z + 1/3 + (( z + 1/3 p E g i (X i p + C( z + 1 p E g i (X i p n P ( > ( z + 1/3 + C( z + 1 p E g i (X i p. By Remark 2.1, we can choose δ ( 2(p 2 p 2 (p 1 p 1 E g i (X i p 1/(p 2 2(p 2p 2 (p 1 p 1 E g i (X i p. Combining the above inequalities with (2.6 and the non-uniform Berry-Essee bound for independent random variables yields (
15 Next we prove (2.6. The main idea of the proof is first to truncate g i (X i and then adopt the proof of Theorem 2.1 to the truncated sum. Without loss of generality, assume z 0 as we can simply apply the result to T. By (5.1, it suffices to show that (5.11 and (5.12 P (z W z γ z + e z/3 τ P (z W z + γ z + e z/3 τ. Since the proof of (5.12 is similar to that of (5.11, we only prove (5.11. It is easy to see that P (z W z P ( > (z + 1/3 + P (z W z, (z + 1/3. Now (5.11 follows directly by Lemmas 5.1 and 5.2 below. This completes the proof of Theorem 2.2. Lemma 5.1 Let Then ξ i g i (X i, ξi ξ i I(ξ i 1, W n ξ i. (5.13 P (z W z, (z + 1/3 P (z W z, (z + 1/3 + P (ξ i > (z + 1/3 + P (W ξ i > (z 2/3P ( ξ i > 1. Proof. We have P (z W z, (z + 1/3 P (z W z, (z + 1/3, max 1in ξ i 1 +P (z W z, (z + 1/3, max ξ i > 1 1in P (z W z, (z + 1/3 + P (W > (2z 1/3, ξ i > 1 and P (W > (2z 1/3, ξ i > 1 15
16 as desired. P (ξ i > (z + 1/3 + P (W > (2z 1/3, ξ i (z + 1/3, ξ i > 1 P (ξ i > (z + 1/3 + P (W ξ i > (z 2/3, ξ i > 1 P (ξ i > (z + 1/3 + P (W ξ i > (z 2/3P ( ξ i > 1, Lemma 5.2 We have (5.14 P (z W z, (z + 1/3 e z/3 τ. Proof. Noting that E ξ i 0, e s 1 + s + s 2 (e a 1 aa 2 for s a and a > 0 and that a ξ i a, we have for a > 0 (5.15 Ee a W n Ee a ξ i n ( 1 + a E ξ i + (e a 1 ae ξ i 2 ( exp (e a 1 a E ξ i 2 ( exp (e a 1 a Eξi 2 exp(e a 1 a. In particular, we have Ee W /2 exp(e 1/ If δ 0.07, then This proves (5.14 when δ P (z W z, (z + 1/3 P ( W > (2z 1/3 e z/3+1/6 Ee W /2 e z/3 exp(e.5 4/3 1.38e z/3 20δ e z/3. For δ < 0.07, let 0 for w z δ, (5.16 f (w e w/2 (w z + + δ for z δ w z + δ, e w/2 ( + 2δ for w > z + δ. Put M i (t ξ i {I( ξ i t 0 I(0 < t ξ i, 16 M(t n M i (t.
17 By (5.5 and similar to (5.6, we have (5.17 E{W f ( W { E f ( W + t M(tdt { + E ξ i (f ( W ξ i f i ( W ξ i : G 1 + G 2, It follows from the fact that M(t 0, f (w ew/2 for z δ w z + δ and f (w 0 for all w, (5.18 { G 1 E t δ f ( W + t M(tdt { E e W /2 I(z W z, (z + 1/3 { E e W /2 I(z W z, (z + 1/3 { +E e W /2 I(z W z, (z + 1/3 G 1,1 G 1,2, t δ t δ t δ M(tdt E M(tdt ( M(t E M(tdt where G 1,1 e z/3 1/6 P (z W z, (z + 1/3 { G 1,2 E e W /2 M(t E M(t dt. t δ t δ E M(tdt, By (2.2 and the assumption that δ 0.07, Hence (5.19 t δ E M(tdt E ξ i min(δ, ξ i E ξ i min(δ, ξ i 1/2. ( G 1,1 (1/2e z/3 1/6 P z W z, (z + 1/3. By (5.15, we have Ee W exp(e 2 < It follows from the Cauchy-Schwarz inequality that (5.20 G 1,2.5 t δ ( 0.5Ee W + 2E ˆM(t M(t 2 dt { δ + 2 Eξi 2 (I( ξ i t 0 + I(0 < t ξ i dt t δ 17
18 { δ + 2 { δ + 2δ Eξi 2 Eξi 2 min(δ, ξ i 2.03δ. As to G 2, it is easy to see that f (w f i (w e w/2 i e w/2 i. Hence, by the Hölder inequality, (5.15 and the assumption that ξ i and W ξ i are independent (5.21 G 2 E ξ i e ( W ξ i /2 ( i (Eξ i 2 e W 1/2 ξ i (E( i 2 1/2 ( Eξi 2 Ee W ξ i 1/2 i ξ i 2 i 2. Following the proof of (5.15 and by using e s 1 s (e a 1/a for s a and a > 0, we have EW 2 e W Eξi 2 e ξ i Ee W ξ i + Eξ i (e ξ i 1Eξ j (e ξ j 1Ee W ξ i ξ j 1i jn 2.06 e Eξi (e 1 2 Eξi 2 Eξj 2 1i jn Thus, we obtain 2.06 e (e 1 2 < (5.22 E{W f ( W E W e W /2 ( + 2δ { ( 2 + 2δ E(W 2 e W 1/2 3.42( 2 + 2δ. Combining (5.17, (5.19, (5.20, (5.21 and (5.22 yields P (z W z, (z + 1/3 2e z/3+1/6{ 3.42( 2 + 2δ δ ξ i 2 i 2 e z/3{ 21δ ξ i 2 i 2 e z/3 τ. 18
19 This proves (5.14. Proof of Remark 2.1. It is known that for x 0, y 0, α > 0, γ > 0 with α + γ 1 x α y γ αx + γy, ( which yields with α (p 2/(p 1, γ 1/(p 1, x b(p 1/(p 2 and y a x α y γ αx + γy b + (p 2p 2 a p 1 (p 1 p 1 b p 2 ( p 2 p 1 p 2 ap 1 b p 2 1/(p 1, or On the other hand, it is clear that b a (p 2p 2 a p 1 (p 1 p 1 b p 2. a a (p 2p 2 a p 1 (p 1 p 1 b p 2. This proves (2.11. Now (2.2 follows directly from (2.11, (2.10 and the assumption (1.1. Proof of Remark 2.2. Note that δ β/2 1/4. Applying (2.11 with p 3 yields E g i (X i min(δ, g i (X i E g i (X i I( g i (X i 1 min(δ, g i (X i { Egi 2 (X i I( g i (X i 1 E g i (X i 3 I( g i (X i 1/(4δ ( 1 4δ Egi 2 (X i I( g i (X i > 1 + E g i (X i 3 I( g i (X i 1 /(4δ 1 β/(4δ 1/2. This proves Remark Proofs of Other Theorems In this section, we prove Theorems
20 6.1 Proof of Theorem 3.1 For 1 k m, let h k (x 1,..., x k E(h(X 1,..., X m X 1 x 1,..., X k x k and h k (x 1,..., x k h k (x 1,..., x k k g(x i. Observing that U n n 1 m ( n 1 g(x i + h m m (X i1,..., X im, 1i 1 <...<i mn we have where n mσ 1 U n W +, W 1 nσ1 n n mσ 1 ( n m g(x i, 1 1i 1 <...<i mn h m (X i1,..., X im. Let l n ( n 1 h mσ 1 m m (X i1,..., X im. 1i 1 <...<i m,i j l for all j By Theorems 2.1 and 2.2 (with Remark 2.3 for proof of (3.1, it suffices to show that (6.1 E 2 (m 12 σ 2 m(n m + 1σ 2 1 and (6.2 E l 2 It is known that (see, e.g., [18], p.271 2(m 1 2 σ 2 nm(n m + 1σ1 2. (6.3 ( E 1i 1 <...<i mn 2 h m (X i1,..., X im ( n m ( m ( n m E h 2 m j m j j(x 1,..., X j. j2 Note that (6.4 E h 2 j(x 1,..., X j j j Eh 2 j(x 1,..., X j 2 E[g(X i h k (X 1,..., X j ] + E( g(x i 2 Eh 2 j(x 1,..., X j 2jE[g(X 1 E(h(X 1,..., X m X 1,..., X j ] + keg 2 (X 1 20
21 Eh 2 j(x 1,..., X j 2jE[g(X 1 h(x 1,..., X m ] + jeg 2 (X 1 Eh 2 j(x 1,..., X j 2jEg 2 (X 1 + jeg 2 (X 1 Eh 2 j(x 1,..., X j jeg1(x 2 1. We next prove that for 2 j m (6.5 Eh 2 j 1(X 1..., X j 1 j 1 Eh 2 j j(x 1,..., X j Since E h 2 2 (X 1, X 2 0, (6.5 holds for j 2 by (6.4. Assume that (6.5 is true for j. Then (6.6 E(h j+1 (X 1,..., X j+1 h j (X 1,..., X j h j (X 2,..., X j+1 2 On the other hand, we have Eh 2 j+1(x 1,..., X j+1 4E[h j+1 (X 1,..., X j+1 h j (X 1,..., X j ] +2Eh 2 j(x 1,..., X j + 2Eh j (X 1,..., X j h j (X 2,..., X j+1 Eh 2 j+1(x 1,..., X j+1 2Eh 2 j(x 1,..., X j ( +2E E(h j (X 1,..., X j h j (X 2,..., X j+1 X 2,..., X j Eh 2 j+1(x 1,..., X j+1 2Eh 2 j(x 1,..., X j + 2Eh 2 j 1(X 1,..., X j 1. (6.7 E(h j+1 (X 1,..., X j+1 h j (X 1,..., X j h j (X 2,..., X j+1 2 ( E E(h j+1 (X 1,..., X j+1 h j (X 1,..., X j h j (X 2,..., X j+1 X 1,..., X j 2 Eh 2 j 1(X 1,..., X j 1. Combining (6.5 and (6.6 yields 2Eh 2 j(x 1,..., X j Eh 2 j+1(x 1,..., X j+1 + Eh 2 j 1(X 1,..., X j 1 Eh 2 j+1(x 1,..., X j+1 + j 1 Eh 2 j j(x 1,..., X j by the induction hypothesis, which in turn reduces to (6.4 for j + 1. This proves (6.4. It follows from (6.4 that (6.8 Eh 2 j(x 1,..., X j j m Eh2 m(x 1,..., X m j m σ2. (6.9 To complete the proof of (6.1, we need the following two inequalities: m ( m j j2 ( n m j m(m 12 n m j m (n m + 1n( m 21
22 and (6.10 m 1 for n > m 2. In fact, we have and m 1 ( m 1 ( n m j + 1 2(m 12 n j m 1 j m (n m + 1n( m m ( m j j2 ( n m j m j m m j2 m 1 ( m 1 j 1 ( n m m j ( m 1 ( n m j m 1 j ( n 1 m 1 ( n 1 {1 m 1 ( n 1 { 1 m 1 ( n m m 1 (n m!/(n m m + 1! (n 1!/(n m! n 1 jn m+1 ( n 1 n 1 m 1 m 1 jn m+1 j ( n 1 (m 1 2 m 1 n m + 1 (m 1 2 m (n m + 1n( n m ( m 1 ( n m j + 1 j m 1 j m m 1 m m 1 m m 1 m 2 j0 m 1 ( n 2 m m 2 ( n 1 ( (m 1 2 m 1 m(n 1. 2(m 1 2 (n m + 1n( n m (1 m 1 j ( m 1 ( n m j j m 1 j m m ( m 2 ( n m + 1 j m 2 j m + 1 ( n 1 m { m 1 + (m 12 (n m + 1m m 1 ( n m m 1 m 1 ( m 1 ( n m j m 1 j ( m 1 ( n m j m 1 j 22
23 (6.11 From (6.8 and (6.9 we obtain that This proves (6.1. Similarly, by (6.10 (6.12 E( l 2 E 2 n ( n 2E { 2 h m 2 σ1 2 m m (X 1,..., X m 1i 1 <...<i mn m 1 n m 2 σ 2 1 ( n m nσ2 m 2 σ 2 1 ( n m j2 m 1 j2 (m 1 2 σ 2 (n m + 1 m σ1 2. ( m j ( m j ( n m m j ( n m j m j m Eh 2 j(x 1,..., X j n ( n 2E ({ 2 m 2 σ1 2 hm (X m i1,..., X im 1i 1 <...<i mn 1i 1 <...<i mn, all i j l n ( n 2E ( 2 h m 2 σ1 2 m m (X i1,..., X im 1, X m 1i 1 <...<i m 1 n 1 n ( n 2 ( n 1 m 2 σ1 2 m m 1 n ( n 2 ( n 1 m 2 σ1 2 m m 1 σ2 mσ 2 1 ( n m 1 m 1 2(m 1 2 σ 2 m n(n m + 1 σ1 2. m 1 m 1 ( m 1 ( n m E h 2 j m 1 j j+1(x 1,..., X j ( m 1 ( n m Eh 2 j m 1 j j+1(x 1,..., X j ( m 1 ( n m j + 1 j m 1 j m This proves (6.2 and hence completes the proof of Theorem Proof of Theorem 3.2 We follow a similar argument as that in the proof of Theorem 3.1. For 1 j k, let X j (X j1,..., X jmj and x j (x j1,..., x jmj and define k m j h(x 1,..., x k h(x 1,..., x k h j (x ji. For the given U-statistic U n, we define its projection k n j Û n l1 23 E(U n X jl.
24 Since we have m j /n j k n j Û n l1 ( nj 1 ( nj /, m j 1 m j m j n j h j (X jl. The difference U n Û n can be rewritten as { k U n Û n ( nj m j 1 h(x1i 1,..., X ki k, where X ji j (X jij1,..., X jijmj and the summation is carried out over all indices 1 i j1 < i j2 <... < i jmj n j, j 1, 2,..., k. Thus, we have with Let W σ 1 n n k j l1 σ 1 n U n W + m j n j h j (X jl, { σ 1 k ( nj 1 n h(x1i m j 1,..., X ki k. { jl σ 1 k ( nv 1 n (jl h(x1i m v1 v 1,..., X ki k, where the summation is carried out over all indices 1 i v1 < i v2 <... < i vmv n v, 1 v k, v j and 1 i j1 < i j2 <... < i jmj n j with i js l for 1 s m j. By Theorems 2.1 and 2.2, it suffices to show that (6.13 E 2 σ2 ( k m 2 j 2 σ 2 n n j and (6.14 For 0 d j m j, 1 j k let E jl 2 2σ2 m 2 j n 2 j σ2 n k v1 m 2 v n v Y d1,...,d k (x ji, 1 i d j, 1 j k E h(x j1,..., x jd1, X jdj +1,..., X jmj, 1 j k 24
25 and Noting that y d1,...,d k EY 2 d 1,...,d k (X ji, 1 i d j, 1 j k. E( h(x 1i 1,..., X ki k X jl 0 for every 1 l m j, 1 j k, we have (see (4.5.8 in [18] (6.15 ( 2 E U n Û n { k σ 2{ k ( nj m j 1 σ 2( k ( nj m j 1 m 2 j n j 2, d d k 2 0 d j m j, 1 j k d d k 2 0 d j m j, 1 j k k {( mj d j ( nj m j m j d j y d1,...,d k k {( mj ( nj m j d j m j d j where in the last inequality we used the fact that (6.16 d d k 2 0 d j m j, 1 j k k {( mj ( nj m ( j k m 2 j 2 d j m j d j n j k ( nj m j. (See below for proof. This proves (6.13. As to (6.14, consider j 1 only. Similar to (6.12 and (6.15, we have (with X 1i 1 (X 1i1,1,..., X 1i1,m1 1, X 1,m 1 σ 2 ne 1l 2 { k v1 σ 2{ k ( nv m v 2E ( v1 1 i v1 < i v2 <... < i vmv n v, 2 v k 1 i 1,1 < i 1,2 <... < i 1,mj 1 n j 1 ( nv 2 ( n1 1 k ( nv m v m 1 1 m v2 v d d k 1 0 d 1 m 1 1, 0 d v m v, 2 v k ( m1 1 d 1 ( n1 m 1 m 1 1 d 1 k 2 h(x 1i, X 1 2i 2,..., X ki k v2 ( mv ( nv m v d v m v d v 25
26 σ2 m { 1 k ( nv 1 { n 1 m v1 v 1d 1 m vk d 1 0,1d vm v σ2 m { 1 k ( nv 1 { (m1 1 2 ( n1 1 k ( nj n 1 m v1 v n 1 m m 1 1 m j2 j σ2 m 2 1 n 2 1 1vk for n 1 2m 1. This proves (6.14. ( n1 m + 1 m 2 v m 1 1 n 2vk v m v + 1 m 2 v n v m v + 1 2σ2 m 2 1 n 2 1 1vk m 2 v n v Now we prove (6.16. Consider two cases in the summation: Case 1: At least one of d j 2, say d 1 2. In this case, by (6.9 for n 1 2m 1. d d j m j, 1 j k { k j2 { k m4 1 n 2 1 j2 ( nj m j ( nj m j m 2 k 2d 1 m 1 m 2 1 (m n 1 (n 1 m k ( nj m j k j2 {( mj ( nj m j d j m j d j 2d 1 m 1 k ( m1 ( n1 m 1 d 1 m 1 d 1 ( nj m j ( m1 d 1 ( n1 m 1 m 1 d 1 d1 m ( nj m j Case 2. At least two of {d j are equal to 1, say d 1 d 2 1. Then k {( mj ( nj m j d d 1 d 2 1 j m j d j m 1 m 2 ( n1 m 1 m 1 1 m2 1 m2 2 n 1 n 2 k ( nj m j. ( n2 m 2 m 2 1 k j3 ( nj m j 26
27 Thus, we have d d k 2 1 d j m j, 1 j k ( k m 4 j n 2 + j ( k m 2 j 2 k n j k 1i jk ( nj m j. {( mj ( nj m j d j m j d j m 2 i m2 j n i n j k This proves (6.16. Now the proof of Theorem 3.2 is complete. 6.3 Proof of Theorem 3.3 Let ψ(t t 0 and hence J(sds. As in [Serfling (1980, p.265], we have ( nj m j T (F n T (F [ψ(f n (x ψ(f (x]dx nσ 1 (T (F n T (F W +, where Let where F n,i (x 1 n W 1 n nσ (I(X i x F (xj(f (xdx nσ 1 [ψ(f n (x ψ(f (x (F n (x F (xj(f (x]dx. η i (x I(X i x F (x, g i (X i 1 (I(X i x F (xj(f (xdx, nσ i nσ 1 [ψ(f n,i (x ψ(f (x (F n,i (x F (xj(f (x]dx, {F (x + 1jn,j i I(X j x. We only need to prove (6.17 and (6.18 σ 2 E 2 c 2 0 n 1 EX 2 1 σ 2 E i 2 2c 2 0n 2 EX
28 Observe that the Lipschitz condition (3.9 implies (6.19 t s ψ(t ψ(s (t sj(s (J(u + s J(sdu 0.5 c 0 (t s 2 0 for 0 s, t 1. Hence Observe that and ( σ 2 E c 2 0 ne (F n (x F (x 2 dx ( 0.25c 2 0 n 3 n 2dxdy E η i (xη j (y ( 0.25c 2 0 n 3 3n 2 Eη1(xEη 2 1(y 2 + ne{η1(xη 2 1(y 2 dxdy. ( Eη1(xEη 2 1(ydxdy 2 2 F (x(1 F (xdx (E X1 2 EX1 2 E{η 2 1(xη 2 1(ydxdy xy xy E{η 2 1(xη 2 1(ydxdy { (1 F (x 2 (1 F (y 2 F (x +F 2 (x(1 F (y 2 (F (y F (x + F 2 (xf 2 (y(1 F (y dxdy 2 F (x(1 F (ydxdy 2 { 2 xy { x0 xy0 + 0<xy x F (xdx + y 0 + x0,y>0 { 2 E(X1 2 + E(X EX1 EX+ 1 F (x(1 F (ydxdy y(1 F (ydy + x0 F (xdx (1 F (ydy y>0 4EX 2 1 This proves (6.17. Next we prove (6.18. Observe that σ i n [ψ(f n (x ψ(f n,i (x (F n (x F n,i (xj(f n,i (xdx + (F n (x F n,i (x[j(f n,i (x J(F (x]dx 28
29 ( 2 E ηi (xdx 2 0.5c 0 (F n (x F n,i (x 2 dx +c 0 F n (x F n,i (x F n,i (x F (x dx 0.5c 0 n 2 (I(X i x F (x 2 dx +c 0 n 2 I(X i x F (x j x F (xdx j i{i(x 0.5c 0 n 2 ηi 2 (xdx + c 0 n 2 η i (x η j (x dx, j i Eη 2 1(xη 2 1(ydxdy 4EX 2 1 and ( E η i (x 2 η j (x dx j i (n 1 E{ η i (x j i η j (x η i (y j i E η i (xη i (y E{ j i η j (x j i η j (y dxdy η j (y dxdy η i (x 2 η i (y 2 η j (x 2 η j (y 2 dxdy j i j i η i (x 2 2 η i (y 2 2dxdy (n 1(E X 1 2 (n 1EX 2 1. Therefore ( σ 2 E i 2 n 3 c 2 0E 0.5 ηi 2 (xdx + η i (x 2 η j (x dx j i { ( 2 ( ηi 2 (xdx + 1.5E n 3 c E { n 3 c 2 0 3EX (n 1EX1 2 2n 2 EX 2 1. η i (x 2 η j (x dx j i This proves (6.18 and hence the theorem. 29
30 6.4 Proof of Theorem 3.4 Let Z 1 and Z 2 be independent standard normal random variables that are independent of {X i and {Y i. Put and write b Nn νσ 2 + τ 2 µ 2, T n X i nµν Nn, H n X i N n µ nb Nn σ T n Nn σ H n + (N n nνµ Nn σ, T n (Z 1 Z 1 + (N n nνµ. nb nb nb nb Applying the Berry-Esseen bound to H n for given N n yields (6.20 sup P (T n x P (T n (Z 1 x x ( X1 3 P ( N n nν > nν/2 + CE Nn σ 3 I{ N n nν nν/2 4n 1 ν 2 τ 2 + Cn 1/2 ν 1/2 σ 3 E X 1 3. Let x.5nν x 1.5nν W n N n nν nτ, for x <.5nν for.5nν x 1.5nν for x > 1.5nν T n (Z 1 : N n σ nb Z 1 + (N n nνµ nb τµ b (W n + σ ν τµ Z 1 +, where Let ( N n nνσz 1 nτµ. i ( (N n Y i + ν nνσz 1 nτµ. Then E( W n Z 1 Z ( 1 σ Wn (N n nν E Z 1 σ nτµ nν nµ ν and 1 nτ E( (Y i ν( i Z 1 Z 1 n 3/2 τ 2 µ ν E(Y i ν 2 Z 1 σ n 3/2 µ ν. 30
31 Now letting T n (Z 1, Z 2 τµ b and applying Theorem 2.1 for given Z 1 yields (Z 2 + σ ν τµ Z 1 (6.21 sup P (T n (Z 1 x P (T n (Z 1, Z 2 x x P ( N n nν > 0.5nν + sup P ( T n (Z 1 x P (T n (Z 1, Z 2 x x 4τ 2 ( nν 2 + C E X1 3 n 1/2 σ 3 + E Z 1 σ n 1/2 µ ν Cn 1/2( τ 2 ν 2 + E X 1 3 σ 3 + σ µ ν It is clear that T n (Z 1, Z 2 has a standard normal distribution. This proves (3.12 by (6.20 and ( Proof of Theorem 3.5 Since (3.13 is trivial if n E g i (X i p > 1/6, we assume (6.22 E g i (X i p 1/6. Let W n g i (X i. It is known that for 2 < p 3 (6.23 Observe that n(h( ˆΘn h(θ (6.24 h (θ where E W p 2(EW 2 p/2 + E g i (X i p 2.2. n ( ˆΘn θ h h (θ( (θ ˆΘ n θ + W + + : W + Λ + R, n h (θ 0 n 1/2 (W + 0 [h (θ + t h (θ]dt [h (θ + t h (θ]dt n (n 1/2 W +(n 1/2 Λ + h [h (θ + t h (θ]dt, (θ 0 n n 1/2 (W + R h [h (θ + t h (θ]dt, (θ (n 1/2 W +(n 1/2 c 0 /2 for x < c 0 /2, x x for c 0 /2 x c 0 /2, c 0 /2 for x > c 0 /2. 31
32 Clearly, n 1/2 W c 0 /2 and n 1/2 c 0 /2 imply R 0. Hence (6.25 P ( R > 0 P ( W > c 0 n 1/2 /2 + P ( > c 0 n 1/2 /2 4/(c 2 0n + 2E /(c 0 n 1/2. To apply Theorem 2.1, let W i W g(x i and Noting that Λ i i + n (n 1/2 W i +(n 1/2 i h [h (θ + t h (θ]dt. (θ 0 (6.26 (n 1/2 W +(n 1/2 [h (θ + t h (θ]dt 0 0.5δ(c 0 ((n 1/2 W + (n 1/2 2 δ(c 0 ((n 1/2 W 2 + (n 1/2 2 ( δ(c 0 (c 0 /2 3 p (n 1/2 W p 1 + (c 0 /2n 1/2, we have (6.27 E W Λ E W + (c 0/2 3 p δ(c 0 h (θ n (p 2/2 E W p + c 0δ(c 0 h E W (θ (1 + c 0δ(c 0 E W h + 2.2c3 p 0 δ(c 0 (θ h (θ n (p 2/2. Similar to (6.26, From this we obtain (n 1/2 W +(n 1/2 [h (θ + t h (θ]dt (n 1/2 W i +(n 1/2 i ( δ(c 0 (c 0 3 p (n 1/2 W (n 1/2 W i (n 1/2 W + (n 1/2 W i p 2 +c 0 (n 1/2 (n 1/2 i ( δ(c 0 c 3 p 0 n (p 1/2 g(x i (2 W i p 2 + g(x i p 2 + c 0 n 1/2 i. (6.28 E g(x i (Λ Λ i E g(x i ( i nδ(c0 + {(c h 0 3 p n (p 1/2 n ( E g(x i 2 (2 W i p 2 + g(x i p 2 (θ 32
33 n +c 0 n 1/2 E g(x i ( i (1 + c 0δ(c 0 n h E g(x i ( i (θ + c3 p 0 δ(c 0 h (θ n (p 2/2 (2Eg 2 (X i + E g(x i p (1 + c 0δ(c 0 n h E g(x i ( i + 2.2c3 p 0 δ(c 0 (θ h (θ n (p 2/2. This proves (3.13 by (2.5, (6.25, (6.27 and ( Proofs of Example 4.1 and Remark 2.5 First we prove Example 4.1. Let Z denote a standard normally distributed random variable and φ(x be the standard normal density function. Observe that 0 < c 0 < 2 and P (T ε c 0 Φ(ε c 0 P (Z ε/ Z 1/2 0 Φ(ε c 0 which proves (4.2. Clearly, we have P (Z 0 + P (Z 3/2 ε, Z > 0 Φ(ε c 0 ε 2/3 0 ε 2/3 2ε φ(tdt εc0 0 φ(tdt φ(tdt (ε 2/3 2ε/3 ε 2/3 /6, E W + E εe c 0 Z Z Z 1/2 c0 + εe Z 1/2 ε(c c 0 7ε by the fact that c 0 < 2. This proves (4.3. As to (4.4, observe that (6.29 E + E X i 3 2cε + 4n 1/2 8ε provided n > ε 2. Below we bound α. Since {X i are i.i.d., we have α εe W 1/2 ˆX 1 + X X n 1/2. 33
34 Let Y and Z be independent standard normal random variables, and let r (n 1/n and s 1 r 2. Noting that EW ( ˆX 1 + X X n r and EZ(sY + rz r, we see that (W, ˆX 1 + X X n and (Z, sy + rz have the same distribution. Hence (6.30 Write E W 1/2 ˆX 1 + X X n 1/2 E Z 1/2 sy + rz 1/2 sy + rz Z E Z 1/2 sy + rz 1/2 ( Z 1/2 + sy + rz 1/2 { Y se Z 1/2 sy + rz 1/2 ( Z 1/2 + sy + rz 1/2 { Z +(1 re Z 1/2 sy + rz 1/2 ( Z 1/2 + sy + rz 1/2 : sr 1 + (1 rr 2. R 1 E{ I( rz s Y /2 + E{ I(s Y /2 < rz 2s Y + E{ I( rz > 2s Y : R 1,1 + R 1,2 + R 1,3. Let C denote an absolute constant. Then { Y I( rz s Y /2 R 1,1 2E Z 1/2 sy (4/sE(s Y /(2r 1/2 4s 1/2, R 1,2 { Y I(s Y /2 < rz 2s Y 4E sy sy + rz 1/2 Cs 1/2 and Thus, we have R 1,3 { Y I( rz > 2s Y 2E Z 3/2 { Y CE (s Y 1/2 Cs 1/2. R 1 Cs 1/2. As to R 2, we have R 2 E(1/ sy + rz 1/2 c 0 34
35 Note that r 1 and s 2/n as n. Combining the above inequalities yields (6.31 α Cεn 1/4. By (6.29 and (6.31, we have E + E X i 3 + α 8ε + Cε 1/2 n 1/8 (8 + Cε provided that n > (1/ε 4. This proves (4.4. Finally, we prove (2.12 in Remark 2.5. Let θ (n 1/n, ρ (1/n 1/2, and let Y and Z be independent standard normal random variables. By (6.29, (6.32 E + E X i 3 4ε + 4n 1/2 8ε 2/3 for n (1/ε 4/3. Following the proof of (6.30, we have (6.33 E X i ( (X 1,..., X i,..., X n (X 1,..., 0,..., X n nεe X 1 ( X 1 + X X n 1/2 X X n 1/2 nεe ρy ( ρy + θz 1/2 θz 1/2 { nερ 2 Y 2 E ρy + θz 1/2 θz 1/2 ( ρy + θz 1/2 + θz 1/2 { Y 2 I( θz ρ Y /2 εe ρy + θz θz 1/2 { Y 2 I(ρ Y /2 < θz 2ρ Y +εe ρy + θz 1/2 θz { Y 2 I( θz > 2ρ Y +εe ρy + θz 1/2 θz { 2ερ 1 E Y θz 1/2 I( θz ρ Y /2 { Y I(ρ Y /2 < θz 2ρ Y +2ερ 1 E αy + θz 1/2 { Y 2 I( θz > 2ρ Y +2εE θz 3/2 Cερ 1/2 Cεn 1/4 2Cε 2/3 for an absolute constant C, provided that n 16(1/ε 4/3. This proves (2.12, by (6.32 and (
36 Acknowledgments. The authors are thankful to Xuming He for his contribution to the construction of the example in Section 4. References [1] Bentickus, V., Götze, F. and Zitikis, M. (1994. Lower estimates of the convergence rate for U-statistics. Ann. Probab. 22, [2] Bhattacharya, R.N. and Ghosh, J.K. (1978. On the validity of the formal Edgeworth expansion. Ann. Statist. 6, [3] Bickel,P. (1974. Edgeworth expansion in nonparametric statistics. Ann., Statist. 2, [4] Bolthausen, E. and Götze,F. (1993. The rate of convergence for multivariate sampling statistics. Ann. Statist [5] Borovskich, Yu.V. (1983. Asymptotics of U-statistics and von Mises functionals. Sov. Math. Dokl. 27, [6] Callaert, H. and Janssen, P. (1978. The Berry-Esseen theorem for U-statistics. Ann., Statist. 6, [7] Chan, Y.K. and Wierman, J. (1977. On the Berry-Esseen theorem for U-statistics. Ann. Probab. 5, [8] Chen, L.H.Y. and Shao, Q.M. (2001. A non-uniform Berry-Esseen bound via Stein s method. Probab. Theory Related Fields 120, [9] Figiel, T., Hitczenko, P., Johnson, W.B., Schechtman, G., and Zinn, J. (1997. Extremal properties of Rademacher functions with applications to the Khintchine and Rosenthal inequalities. Trans. Amer. Math. Soc. 349, [10] Filippova, A.A. (1962. Mises s theorem of the asympototic behavior of functionals of empirical distribution functions and its statistical applications. Theory Probab. Appl. 7,
37 [11] Finelstein, M., Kruglov, V.M., and Tucker, H.G. (1994. Convergence in law of random sums with non-random centering. J. Theoret. Probab. 3, [12] Friedrich, K. O. (1989. A Berry-Esseen bound for functions of independent random variables. Ann. Statist. 17, [13] Grams, W.F. and Serfling, R.J. (1973. Convergence rates for U-statistics and related statistics. Ann., Statist. 1, [14] Helmers, R. (1977. The order of the normal approximation for linear combinations of order statistics with smooth weight functions. Ann. Probab. 5, [15] Helmers, R. and Janssen, P. (1982. On the Berry-Esseen theorem for multivariate U-statistics. Math. Cent. Rep. SW 90/82, Mathematisch Centrum, Amsterdam, pp [16] Helmers, R., Janssen, P. and Serfling, R. J. (1990. Berry-Esseen bound and bootstrap results for genalized L-statistics. Scand. J. Statist. bf 17, [17] Hoeffding, W. (1948. A class of statististics with asymptotically normal distribution. Ann. Math. Statist. 19, [18] Koroljuk, V.S. and Borovskich, Yu. V. (1994. Theory of U-statistics. Kluwer Academic Publishers, Boston. [19] Robbins, H. (1948. The asymptotic distribution of the sum of a random numbers of random random variables. Bull. A.M.S. 54, [20] Rosenthal, H.P. (1970. On the subspaces of L p (p > 2 spanned by sequences of independent random variables. Israel J. Math. 8, [21] Serfling, R.J. (1980. Approximation Theorems of Mathematical Statisitics. Wiley, New York. [22] Shorack, G.R. (2000. Probability for Statisticians. Springer, New York. [23] Wang, Q. (2001. Non-uniform Berry-Esséen Bound for U-Statistics. Statist. Sinica (to appear [24] Wang, Q., Jing, B.Y. and Zhao, L. (2000. The Berry-Esseen bound for studentized statistics. Ann. Probab. 28,
38 [25] Zhao, L.C. and Chen, X.R. (1983. Non-uniform convergence rates for distributions of U- statistics. Sci. Sinica (Ser. A 26, [26] van Zwet, W.R. (1984. A Berry-Esseen bound for symmetric statisics. Z. Wahrsch. Verw. Gebiete 66, Louis H.Y. Chen Qi-Man Shao Institute for Mathematical Sciences Department of Mathematics National University of Singapore Hong Kong University of Science and Technology Singapore Clear Water Bay, Kowloon Republic of Singapore Hong Kong lhychen@ims.nus.edu.sg maqmshao@ust.hk and Department of Mathematics Department of Mathematics Department of Statistics & Applied Probability University of Oregon National University of Singapore Eugene, OR Singapore USA Republic of Singapore and Department of Mathematics Zhejiang University Hangzhou, Zhejiang China 38
EXACT CONVERGENCE RATE AND LEADING TERM IN THE CENTRAL LIMIT THEOREM FOR U-STATISTICS
Statistica Sinica 6006, 409-4 EXACT CONVERGENCE RATE AND LEADING TERM IN THE CENTRAL LIMIT THEOREM FOR U-STATISTICS Qiying Wang and Neville C Weber The University of Sydney Abstract: The leading term in
More informationIntroduction to Self-normalized Limit Theory
Introduction to Self-normalized Limit Theory Qi-Man Shao The Chinese University of Hong Kong E-mail: qmshao@cuhk.edu.hk Outline What is the self-normalization? Why? Classical limit theorems Self-normalized
More informationNon-uniform Berry Esseen Bounds for Weighted U-Statistics and Generalized L-Statistics
Coun Math Stat 0 :5 67 DOI 0.007/s4004-0-009- Non-unifor Berry Esseen Bounds for Weighted U-Statistics and Generalized L-Statistics Haojun Hu Qi-Man Shao Received: 9 August 0 / Accepted: Septeber 0 / Published
More informationBounds of the normal approximation to random-sum Wilcoxon statistics
R ESEARCH ARTICLE doi: 0.306/scienceasia53-874.04.40.8 ScienceAsia 40 04: 8 9 Bounds of the normal approximation to random-sum Wilcoxon statistics Mongkhon Tuntapthai Nattakarn Chaidee Department of Mathematics
More informationEdgeworth expansions for a sample sum from a finite set of independent random variables
E l e c t r o n i c J o u r n a l o f P r o b a b i l i t y Vol. 12 2007), Paper no. 52, pages 1402 1417. Journal URL http://www.math.washington.edu/~ejpecp/ Edgeworth expansions for a sample sum from
More informationSelf-normalized Cramér-Type Large Deviations for Independent Random Variables
Self-normalized Cramér-Type Large Deviations for Independent Random Variables Qi-Man Shao National University of Singapore and University of Oregon qmshao@darkwing.uoregon.edu 1. Introduction Let X, X
More informationEMPIRICAL EDGEWORTH EXPANSION FOR FINITE POPULATION STATISTICS. I. M. Bloznelis. April Introduction
EMPIRICAL EDGEWORTH EXPANSION FOR FINITE POPULATION STATISTICS. I M. Bloznelis April 2000 Abstract. For symmetric asymptotically linear statistics based on simple random samples, we construct the one-term
More informationBahadur representations for bootstrap quantiles 1
Bahadur representations for bootstrap quantiles 1 Yijun Zuo Department of Statistics and Probability, Michigan State University East Lansing, MI 48824, USA zuo@msu.edu 1 Research partially supported by
More informationMOMENT CONVERGENCE RATES OF LIL FOR NEGATIVELY ASSOCIATED SEQUENCES
J. Korean Math. Soc. 47 1, No., pp. 63 75 DOI 1.4134/JKMS.1.47..63 MOMENT CONVERGENCE RATES OF LIL FOR NEGATIVELY ASSOCIATED SEQUENCES Ke-Ang Fu Li-Hua Hu Abstract. Let X n ; n 1 be a strictly stationary
More informationAN EDGEWORTH EXPANSION FOR SYMMETRIC FINITE POPULATION STATISTICS. M. Bloznelis 1, F. Götze
AN EDGEWORTH EXPANSION FOR SYMMETRIC FINITE POPULATION STATISTICS M. Bloznelis 1, F. Götze Vilnius University and the Institute of Mathematics and Informatics, Bielefeld University 2000 (Revised 2001 Abstract.
More informationCramér type moderate deviations for trimmed L-statistics
arxiv:1608.05015v1 [math.pr] 17 Aug 2016 Cramér type moderate deviations for trimmed L-statistics Nadezhda Gribkova St.Petersburg State University, Mathematics and Mechanics Faculty, 199034, Universitetskaya
More informationConvergence rates in weighted L 1 spaces of kernel density estimators for linear processes
Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,
More informationRELATIVE ERRORS IN CENTRAL LIMIT THEOREMS FOR STUDENT S t STATISTIC, WITH APPLICATIONS
Statistica Sinica 19 (2009, 343-354 RELATIVE ERRORS IN CENTRAL LIMIT THEOREMS FOR STUDENT S t STATISTIC, WITH APPLICATIONS Qiying Wang and Peter Hall University of Sydney and University of Melbourne Abstract:
More informationRIESZ BASES AND UNCONDITIONAL BASES
In this paper we give a brief introduction to adjoint operators on Hilbert spaces and a characterization of the dual space of a Hilbert space. We then introduce the notion of a Riesz basis and give some
More informationOn large deviations of sums of independent random variables
On large deviations of sums of independent random variables Zhishui Hu 12, Valentin V. Petrov 23 and John Robinson 2 1 Department of Statistics and Finance, University of Science and Technology of China,
More informationA Gentle Introduction to Stein s Method for Normal Approximation I
A Gentle Introduction to Stein s Method for Normal Approximation I Larry Goldstein University of Southern California Introduction to Stein s Method for Normal Approximation 1. Much activity since Stein
More informationMODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT
Statistica Sinica 23 (2013), 1523-1540 doi:http://dx.doi.org/10.5705/ss.2012.203s MODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT Louis H. Y. Chen 1, Xiao Fang 1,2 and Qi-Man Shao 3 1 National
More informationOn large deviations for combinatorial sums
arxiv:1901.0444v1 [math.pr] 14 Jan 019 On large deviations for combinatorial sums Andrei N. Frolov Dept. of Mathematics and Mechanics St. Petersburg State University St. Petersburg, Russia E-mail address:
More informationA BERRY ESSEEN BOUND FOR STUDENT S STATISTIC IN THE NON-I.I.D. CASE. 1. Introduction and Results
A BERRY ESSEE BOUD FOR STUDET S STATISTIC I THE O-I.I.D. CASE V. Bentkus 1 M. Bloznelis 1 F. Götze 1 Abstract. We establish a Berry Esséen bound for Student s statistic for independent (non-identically)
More informationCramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics. Wen-Xin Zhou
Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics Wen-Xin Zhou Department of Mathematics and Statistics University of Melbourne Joint work with Prof. Qi-Man
More informationTHE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION
THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION JAINUL VAGHASIA Contents. Introduction. Notations 3. Background in Probability Theory 3.. Expectation and Variance 3.. Convergence
More informationAlmost Sure Central Limit Theorem for Self-Normalized Partial Sums of Negatively Associated Random Variables
Filomat 3:5 (207), 43 422 DOI 0.2298/FIL70543W Published by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat Almost Sure Central Limit Theorem
More informationSOME CONVERSE LIMIT THEOREMS FOR EXCHANGEABLE BOOTSTRAPS
SOME CONVERSE LIMIT THEOREMS OR EXCHANGEABLE BOOTSTRAPS Jon A. Wellner University of Washington The bootstrap Glivenko-Cantelli and bootstrap Donsker theorems of Giné and Zinn (990) contain both necessary
More informationA NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES
Bull. Korean Math. Soc. 52 (205), No. 3, pp. 825 836 http://dx.doi.org/0.434/bkms.205.52.3.825 A NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES Yongfeng Wu and Mingzhu
More informationPRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS
PRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS REZA HASHEMI and MOLUD ABDOLAHI Department of Statistics, Faculty of Science, Razi University, 67149, Kermanshah,
More informationSubmitted to the Brazilian Journal of Probability and Statistics
Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a
More informationAsymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University
More informationSelf-normalized Cramér type moderate deviations for the maximum of sums
Bernoulli 193, 2013, 1006 1027 DOI: 10.3150/12-BEJ415 arxiv:1307.6044v1 [math.st] 23 Jul 2013 Self-normalized Cramér type moderate deviations for the maximum of sums WEIDONG LIU 1, QI-MAN SHAO 2 and QIYING
More informationGoodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach
Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach By Shiqing Ling Department of Mathematics Hong Kong University of Science and Technology Let {y t : t = 0, ±1, ±2,
More informationORTHOGONAL DECOMPOSITION OF FINITE POPULATION STATISTICS AND ITS APPLICATIONS TO DISTRIBUTIONAL ASYMPTOTICS. and Informatics, 2 Bielefeld University
ORTHOGONAL DECOMPOSITION OF FINITE POPULATION STATISTICS AND ITS APPLICATIONS TO DISTRIBUTIONAL ASYMPTOTICS M. Bloznelis 1,3 and F. Götze 2 1 Vilnius University and Institute of Mathematics and Informatics,
More informationPhenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012
Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM
More informationOn probabilities of large and moderate deviations for L-statistics: a survey of some recent developments
UDC 519.2 On probabilities of large and moderate deviations for L-statistics: a survey of some recent developments N. V. Gribkova Department of Probability Theory and Mathematical Statistics, St.-Petersburg
More informationAN INEQUALITY FOR TAIL PROBABILITIES OF MARTINGALES WITH BOUNDED DIFFERENCES
Lithuanian Mathematical Journal, Vol. 4, No. 3, 00 AN INEQUALITY FOR TAIL PROBABILITIES OF MARTINGALES WITH BOUNDED DIFFERENCES V. Bentkus Vilnius Institute of Mathematics and Informatics, Akademijos 4,
More informationHan-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek
J. Korean Math. Soc. 41 (2004), No. 5, pp. 883 894 CONVERGENCE OF WEIGHTED SUMS FOR DEPENDENT RANDOM VARIABLES Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek Abstract. We discuss in this paper the strong
More informationENTROPY-BASED GOODNESS OF FIT TEST FOR A COMPOSITE HYPOTHESIS
Bull. Korean Math. Soc. 53 (2016), No. 2, pp. 351 363 http://dx.doi.org/10.4134/bkms.2016.53.2.351 ENTROPY-BASED GOODNESS OF FIT TEST FOR A COMPOSITE HYPOTHESIS Sangyeol Lee Abstract. In this paper, we
More informationLIFE SPAN OF BLOW-UP SOLUTIONS FOR HIGHER-ORDER SEMILINEAR PARABOLIC EQUATIONS
Electronic Journal of Differential Equations, Vol. 21(21), No. 17, pp. 1 9. ISSN: 172-6691. URL: http://ejde.math.txstate.edu or http://ejde.math.unt.edu ftp ejde.math.txstate.edu LIFE SPAN OF BLOW-UP
More informationAlmost sure limit theorems for U-statistics
Almost sure limit theorems for U-statistics Hajo Holzmann, Susanne Koch and Alesey Min 3 Institut für Mathematische Stochasti Georg-August-Universität Göttingen Maschmühlenweg 8 0 37073 Göttingen Germany
More informationSpectral Gap and Concentration for Some Spherically Symmetric Probability Measures
Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures S.G. Bobkov School of Mathematics, University of Minnesota, 127 Vincent Hall, 26 Church St. S.E., Minneapolis, MN 55455,
More informationStein s Method and the Zero Bias Transformation with Application to Simple Random Sampling
Stein s Method and the Zero Bias Transformation with Application to Simple Random Sampling Larry Goldstein and Gesine Reinert November 8, 001 Abstract Let W be a random variable with mean zero and variance
More informationWasserstein-2 bounds in normal approximation under local dependence
Wasserstein- bounds in normal approximation under local dependence arxiv:1807.05741v1 [math.pr] 16 Jul 018 Xiao Fang The Chinese University of Hong Kong Abstract: We obtain a general bound for the Wasserstein-
More informationGaussian vectors and central limit theorem
Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables
More informationBounds on the Constant in the Mean Central Limit Theorem
Bounds on the Constant in the Mean Central Limit Theorem Larry Goldstein University of Southern California, Ann. Prob. 2010 Classical Berry Esseen Theorem Let X, X 1, X 2,... be i.i.d. with distribution
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationAdditive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535
Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June
More informationJournal of Inequalities in Pure and Applied Mathematics
Journal of Inequalities in Pure and Applied Mathematics ON A HYBRID FAMILY OF SUMMATION INTEGRAL TYPE OPERATORS VIJAY GUPTA AND ESRA ERKUŞ School of Applied Sciences Netaji Subhas Institute of Technology
More informationStrong log-concavity is preserved by convolution
Strong log-concavity is preserved by convolution Jon A. Wellner Abstract. We review and formulate results concerning strong-log-concavity in both discrete and continuous settings. Although four different
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley
More informationSelected Exercises on Expectations and Some Probability Inequalities
Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ
More informationCHAPTER 3: LARGE SAMPLE THEORY
CHAPTER 3 LARGE SAMPLE THEORY 1 CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 2 Introduction CHAPTER 3 LARGE SAMPLE THEORY 3 Why large sample theory studying small sample property is usually
More informationChp 4. Expectation and Variance
Chp 4. Expectation and Variance 1 Expectation In this chapter, we will introduce two objectives to directly reflect the properties of a random variable or vector, which are the Expectation and Variance.
More informationArithmetic progressions in sumsets
ACTA ARITHMETICA LX.2 (1991) Arithmetic progressions in sumsets by Imre Z. Ruzsa* (Budapest) 1. Introduction. Let A, B [1, N] be sets of integers, A = B = cn. Bourgain [2] proved that A + B always contains
More informationTail bound inequalities and empirical likelihood for the mean
Tail bound inequalities and empirical likelihood for the mean Sandra Vucane 1 1 University of Latvia, Riga 29 th of September, 2011 Sandra Vucane (LU) Tail bound inequalities and EL for the mean 29.09.2011
More informationA generalization of Cramér large deviations for martingales
A generalization of Cramér large deviations for martingales Xiequan Fan, Ion Grama, Quansheng Liu To cite this version: Xiequan Fan, Ion Grama, Quansheng Liu. A generalization of Cramér large deviations
More informationLecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf
Lecture 13: 2011 Bootstrap ) R n x n, θ P)) = τ n ˆθn θ P) Example: ˆθn = X n, τ n = n, θ = EX = µ P) ˆθ = min X n, τ n = n, θ P) = sup{x : F x) 0} ) Define: J n P), the distribution of τ n ˆθ n θ P) under
More informationFOURIER TRANSFORMS OF STATIONARY PROCESSES 1. k=1
FOURIER TRANSFORMS OF STATIONARY PROCESSES WEI BIAO WU September 8, 003 Abstract. We consider the asymptotic behavior of Fourier transforms of stationary and ergodic sequences. Under sufficiently mild
More informationThe Central Limit Theorem: More of the Story
The Central Limit Theorem: More of the Story Steven Janke November 2015 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November 2015 1 / 33 Central Limit Theorem Theorem (Central Limit
More informationA CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE
Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received
More informationALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR
Periodica Mathematica Hungarica Vol. 51 1, 2005, pp. 11 25 ALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR István Berkes Graz, Budapest, LajosHorváth Salt Lake City, Piotr Kokoszka Logan Qi-man Shao
More informationSHARP BOUNDARY TRACE INEQUALITIES. 1. Introduction
SHARP BOUNDARY TRACE INEQUALITIES GILES AUCHMUTY Abstract. This paper describes sharp inequalities for the trace of Sobolev functions on the boundary of a bounded region R N. The inequalities bound (semi-)norms
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationLimit Theorems for Exchangeable Random Variables via Martingales
Limit Theorems for Exchangeable Random Variables via Martingales Neville Weber, University of Sydney. May 15, 2006 Probabilistic Symmetries and Their Applications A sequence of random variables {X 1, X
More informationECE 4400:693 - Information Theory
ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationA regeneration proof of the central limit theorem for uniformly ergodic Markov chains
A regeneration proof of the central limit theorem for uniformly ergodic Markov chains By AJAY JASRA Department of Mathematics, Imperial College London, SW7 2AZ, London, UK and CHAO YANG Department of Mathematics,
More informationA generalization of Strassen s functional LIL
A generalization of Strassen s functional LIL Uwe Einmahl Departement Wiskunde Vrije Universiteit Brussel Pleinlaan 2 B-1050 Brussel, Belgium E-mail: ueinmahl@vub.ac.be Abstract Let X 1, X 2,... be a sequence
More informationKolmogorov Berry-Esseen bounds for binomial functionals
Kolmogorov Berry-Esseen bounds for binomial functionals Raphaël Lachièze-Rey, Univ. South California, Univ. Paris 5 René Descartes, Joint work with Giovanni Peccati, University of Luxembourg, Singapore
More informationOptimal global rates of convergence for interpolation problems with random design
Optimal global rates of convergence for interpolation problems with random design Michael Kohler 1 and Adam Krzyżak 2, 1 Fachbereich Mathematik, Technische Universität Darmstadt, Schlossgartenstr. 7, 64289
More informationOn the Error Bound in the Normal Approximation for Jack Measures (Joint work with Le Van Thanh)
On the Error Bound in the Normal Approximation for Jack Measures (Joint work with Le Van Thanh) Louis H. Y. Chen National University of Singapore International Colloquium on Stein s Method, Concentration
More informationA CLT FOR MULTI-DIMENSIONAL MARTINGALE DIFFERENCES IN A LEXICOGRAPHIC ORDER GUY COHEN. Dedicated to the memory of Mikhail Gordin
A CLT FOR MULTI-DIMENSIONAL MARTINGALE DIFFERENCES IN A LEXICOGRAPHIC ORDER GUY COHEN Dedicated to the memory of Mikhail Gordin Abstract. We prove a central limit theorem for a square-integrable ergodic
More informationConcentration inequalities and the entropy method
Concentration inequalities and the entropy method Gábor Lugosi ICREA and Pompeu Fabra University Barcelona what is concentration? We are interested in bounding random fluctuations of functions of many
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More informationarxiv: v1 [math.pr] 6 Jan 2014
On Cox-Kemperman moment inequalities for independent centered random variables P.S.Ruzankin 1 arxiv:141.17v1 [math.pr] 6 Jan 214 Abstract In 1983 Cox and Kemperman proved that Ef()+Ef() Ef(+) for all functions
More informationTHE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX
THE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX MARK RUDELSON AND ROMAN VERSHYNIN Abstract. We prove an optimal estimate on the smallest singular value of a random subgaussian matrix, valid
More informationThe Convergence Rate for the Normal Approximation of Extreme Sums
The Convergence Rate for the Normal Approximation of Extreme Sums Yongcheng Qi University of Minnesota Duluth WCNA 2008, Orlando, July 2-9, 2008 This talk is based on a joint work with Professor Shihong
More informationNormal Approximation for Hierarchical Structures
Normal Approximation for Hierarchical Structures Larry Goldstein University of Southern California July 1, 2004 Abstract Given F : [a, b] k [a, b] and a non-constant X 0 with P (X 0 [a, b]) = 1, define
More informationProperties of Random Variables
Properties of Random Variables 1 Definitions A discrete random variable is defined by a probability distribution that lists each possible outcome and the probability of obtaining that outcome If the random
More informationHERMITE MULTIPLIERS AND PSEUDO-MULTIPLIERS
PROCEEDINGS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 124, Number 7, July 1996 HERMITE MULTIPLIERS AND PSEUDO-MULTIPLIERS JAY EPPERSON Communicated by J. Marshall Ash) Abstract. We prove a multiplier
More informationON THE COMPLETE CONVERGENCE FOR WEIGHTED SUMS OF DEPENDENT RANDOM VARIABLES UNDER CONDITION OF WEIGHTED INTEGRABILITY
J. Korean Math. Soc. 45 (2008), No. 4, pp. 1101 1111 ON THE COMPLETE CONVERGENCE FOR WEIGHTED SUMS OF DEPENDENT RANDOM VARIABLES UNDER CONDITION OF WEIGHTED INTEGRABILITY Jong-Il Baek, Mi-Hwa Ko, and Tae-Sung
More informationS. Lototsky and B.L. Rozovskii Center for Applied Mathematical Sciences University of Southern California, Los Angeles, CA
RECURSIVE MULTIPLE WIENER INTEGRAL EXPANSION FOR NONLINEAR FILTERING OF DIFFUSION PROCESSES Published in: J. A. Goldstein, N. E. Gretsky, and J. J. Uhl (editors), Stochastic Processes and Functional Analysis,
More informationA NOTE ON A DISTRIBUTION OF WEIGHTED SUMS OF I.I.D. RAYLEIGH RANDOM VARIABLES
Sankhyā : The Indian Journal of Statistics 1998, Volume 6, Series A, Pt. 2, pp. 171-175 A NOTE ON A DISTRIBUTION OF WEIGHTED SUMS OF I.I.D. RAYLEIGH RANDOM VARIABLES By P. HITCZENKO North Carolina State
More informationWittmann Type Strong Laws of Large Numbers for Blockwise m-negatively Associated Random Variables
Journal of Mathematical Research with Applications Mar., 206, Vol. 36, No. 2, pp. 239 246 DOI:0.3770/j.issn:2095-265.206.02.03 Http://jmre.dlut.edu.cn Wittmann Type Strong Laws of Large Numbers for Blockwise
More informationlarge number of i.i.d. observations from P. For concreteness, suppose
1 Subsampling Suppose X i, i = 1,..., n is an i.i.d. sequence of random variables with distribution P. Let θ(p ) be some real-valued parameter of interest, and let ˆθ n = ˆθ n (X 1,..., X n ) be some estimate
More informationINEQUALITIES FOR SUMS OF INDEPENDENT RANDOM VARIABLES IN LORENTZ SPACES
ROCKY MOUNTAIN JOURNAL OF MATHEMATICS Volume 45, Number 5, 2015 INEQUALITIES FOR SUMS OF INDEPENDENT RANDOM VARIABLES IN LORENTZ SPACES GHADIR SADEGHI ABSTRACT. By using interpolation with a function parameter,
More informationLaplace s Equation. Chapter Mean Value Formulas
Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic
More informationThe strong law of large numbers for arrays of NA random variables
International Mathematical Forum,, 2006, no. 2, 83-9 The strong law of large numbers for arrays of NA random variables K. J. Lee, H. Y. Seo, S. Y. Kim Division of Mathematics and Informational Statistics,
More informationAdvanced Probability Theory (Math541)
Advanced Probability Theory (Math541) Instructor: Kani Chen (Classic)/Modern Probability Theory (1900-1960) Instructor: Kani Chen (HKUST) Advanced Probability Theory (Math541) 1 / 17 Primitive/Classic
More informationPiecewise Smooth Solutions to the Burgers-Hilbert Equation
Piecewise Smooth Solutions to the Burgers-Hilbert Equation Alberto Bressan and Tianyou Zhang Department of Mathematics, Penn State University, University Park, Pa 68, USA e-mails: bressan@mathpsuedu, zhang
More informationEXISTENCE OF MOMENTS IN THE HSU ROBBINS ERDŐS THEOREM
Annales Univ. Sci. Budapest., Sect. Comp. 39 (2013) 271 278 EXISTENCE OF MOMENTS IN THE HSU ROBBINS ERDŐS THEOREM O.I. Klesov (Kyiv, Ukraine) U. Stadtmüller (Ulm, Germany) Dedicated to the 70th anniversary
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationDEPARTMENT MATHEMATIK ARBEITSBEREICH MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE
Estimating the error distribution in nonparametric multiple regression with applications to model testing Natalie Neumeyer & Ingrid Van Keilegom Preprint No. 2008-01 July 2008 DEPARTMENT MATHEMATIK ARBEITSBEREICH
More information5 Operations on Multiple Random Variables
EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y
More informationAsymptotically Efficient Nonparametric Estimation of Nonlinear Spectral Functionals
Acta Applicandae Mathematicae 78: 145 154, 2003. 2003 Kluwer Academic Publishers. Printed in the Netherlands. 145 Asymptotically Efficient Nonparametric Estimation of Nonlinear Spectral Functionals M.
More informationARCS IN FINITE PROJECTIVE SPACES. Basic objects and definitions
ARCS IN FINITE PROJECTIVE SPACES SIMEON BALL Abstract. These notes are an outline of a course on arcs given at the Finite Geometry Summer School, University of Sussex, June 26-30, 2017. Let K denote an
More information1 Exercises for lecture 1
1 Exercises for lecture 1 Exercise 1 a) Show that if F is symmetric with respect to µ, and E( X )
More informationTesting Goodness-of-Fit for Exponential Distribution Based on Cumulative Residual Entropy
This article was downloaded by: [Ferdowsi University] On: 16 April 212, At: 4:53 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 172954 Registered office: Mortimer
More informationOn lower and upper bounds for probabilities of unions and the Borel Cantelli lemma
arxiv:4083755v [mathpr] 6 Aug 204 On lower and upper bounds for probabilities of unions and the Borel Cantelli lemma Andrei N Frolov Dept of Mathematics and Mechanics St Petersburg State University St
More informationZeros of lacunary random polynomials
Zeros of lacunary random polynomials Igor E. Pritsker Dedicated to Norm Levenberg on his 60th birthday Abstract We study the asymptotic distribution of zeros for the lacunary random polynomials. It is
More informationPointwise convergence rates and central limit theorems for kernel density estimators in linear processes
Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract Convergence
More informationONE DIMENSIONAL MARGINALS OF OPERATOR STABLE LAWS AND THEIR DOMAINS OF ATTRACTION
ONE DIMENSIONAL MARGINALS OF OPERATOR STABLE LAWS AND THEIR DOMAINS OF ATTRACTION Mark M. Meerschaert Department of Mathematics University of Nevada Reno NV 89557 USA mcubed@unr.edu and Hans Peter Scheffler
More informationKrzysztof Burdzy University of Washington. = X(Y (t)), t 0}
VARIATION OF ITERATED BROWNIAN MOTION Krzysztof Burdzy University of Washington 1. Introduction and main results. Suppose that X 1, X 2 and Y are independent standard Brownian motions starting from 0 and
More information