Cramér type moderate deviations for trimmed L-statistics

Similar documents
On probabilities of large and moderate deviations for L-statistics: a survey of some recent developments

EXACT CONVERGENCE RATE AND LEADING TERM IN THE CENTRAL LIMIT THEOREM FOR U-STATISTICS

Self-normalized Cramér-Type Large Deviations for Independent Random Variables

Introduction to Self-normalized Limit Theory

THE ASYMPTOTICS OF L-STATISTICS FOR NON I.I.D. VARIABLES WITH HEAVY TAILS

MOMENT CONVERGENCE RATES OF LIL FOR NEGATIVELY ASSOCIATED SEQUENCES

The Convergence Rate for the Normal Approximation of Extreme Sums

Cramér large deviation expansions for martingales under Bernstein s condition

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

High Breakdown Analogs of the Trimmed Mean

AN INEQUALITY FOR TAIL PROBABILITIES OF MARTINGALES WITH BOUNDED DIFFERENCES

Bahadur representations for bootstrap quantiles 1

PRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS

Almost Sure Central Limit Theorem for Self-Normalized Partial Sums of Negatively Associated Random Variables

SOME CONVERSE LIMIT THEOREMS FOR EXCHANGEABLE BOOTSTRAPS

A generalization of Cramér large deviations for martingales

Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek

(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS

On rate of convergence in distribution of asymptotically normal statistics based on samples of random size

On large deviations of sums of independent random variables

On large deviations for combinatorial sums

Journal of Inequalities in Pure and Applied Mathematics

RELATIVE ERRORS IN CENTRAL LIMIT THEOREMS FOR STUDENT S t STATISTIC, WITH APPLICATIONS

Exact Asymptotics in Complete Moment Convergence for Record Times and the Associated Counting Process

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½

arxiv: v1 [math.st] 2 Apr 2016

Large deviations for weighted random sums

EMPIRICAL EDGEWORTH EXPANSION FOR FINITE POPULATION STATISTICS. I. M. Bloznelis. April Introduction

Edgeworth expansions for a sample sum from a finite set of independent random variables

Statistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it

Stein s Method and Characteristic Functions

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

Refining the Central Limit Theorem Approximation via Extreme Value Theory

Lecture 16: Sample quantiles and their asymptotic properties

University of California San Diego and Stanford University and

A NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES

Precise Asymptotics of Generalized Stochastic Order Statistics for Extreme Value Distributions

STAT 200C: High-dimensional Statistics

A Central Limit Theorem for the Sum of Generalized Linear and Quadratic Forms. by R. L. Eubank and Suojin Wang Texas A&M University

COVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE

Journal of Inequalities in Pure and Applied Mathematics

THE DVORETZKY KIEFER WOLFOWITZ INEQUALITY WITH SHARP CONSTANT: MASSART S 1990 PROOF SEMINAR, SEPT. 28, R. M. Dudley

5 Introduction to the Theory of Order Statistics and Rank Statistics

Estimation of the Bivariate and Marginal Distributions with Censored Data

KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS

Random Bernstein-Markov factors

Bounds on expectation of order statistics from a nite population

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON

On lower and upper bounds for probabilities of unions and the Borel Cantelli lemma

Zeros of lacunary random polynomials

Accepted for publication in: Comm. Statist. Theory and Methods April 14, 2015 ASYMPTOTICS OF GOODNESS-OF-FIT TESTS BASED ON MINIMUM P-VALUE STATISTICS

Normal approximation of Poisson functionals in Kolmogorov distance

Asymptotic Statistics-VI. Changliang Zou

Extreme Value for Discrete Random Variables Applied to Avalanche Counts

Tail bound inequalities and empirical likelihood for the mean

Optimal global rates of convergence for interpolation problems with random design

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Stein s Method and the Zero Bias Transformation with Application to Simple Random Sampling

Estimation of a quadratic regression functional using the sinc kernel

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

Asymptotic efficiency of simple decisions for the compound decision problem

Soo Hak Sung and Andrei I. Volodin

Lecture 8: Information Theory and Statistics

SDS : Theoretical Statistics

On the Bennett-Hoeffding inequality

Convergence of Multivariate Quantile Surfaces

The tail does not determine the size of the giant

MAXIMAL CLADES IN RANDOM BINARY SEARCH TREES

Entropy and Ergodic Theory Lecture 15: A first look at concentration

CHARACTERIZATIONS OF UNIFORM AND EXPONENTIAL DISTRIBUTIONS VIA MOMENTS OF THE kth RECORD VALUES RANDOMLY INDEXED

Advanced Probability Theory (Math541)

Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics. Wen-Xin Zhou

ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE. By Michael Nussbaum Weierstrass Institute, Berlin

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

Limiting Distributions

Asymptotic normality of the L k -error of the Grenander estimator

arxiv: v2 [math.ca] 23 Feb 2016

Notes on Poisson Approximation

On the singular values of random matrices

Convergence at first and second order of some approximations of stochastic integrals

A Law of the Iterated Logarithm. for Grenander s Estimator

Mean convergence theorems and weak laws of large numbers for weighted sums of random variables under a condition of weighted integrability

W. Lenski and B. Szal ON POINTWISE APPROXIMATION OF FUNCTIONS BY SOME MATRIX MEANS OF CONJUGATE FOURIER SERIES

arxiv: v5 [math.na] 16 Nov 2017

Random Graphs. EECS 126 (UC Berkeley) Spring 2019

Multiplicative Order of Gauss Periods

Strong log-concavity is preserved by convolution

Asymptotics of minimax stochastic programs

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example

Nonparametric estimation of tail risk measures from heavy-tailed distributions

Chapter 2 Asymptotics

Gaussian vectors and central limit theorem

Theoretical Statistics. Lecture 1.

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011

On the Error Bound in the Normal Approximation for Jack Measures (Joint work with Le Van Thanh)

Wittmann Type Strong Laws of Large Numbers for Blockwise m-negatively Associated Random Variables

Bootstrap inference for the finite population total under complex sampling designs

Finite-time Ruin Probability of Renewal Model with Risky Investment and Subexponential Claims

ON THE SHARPNESS OF ONE INEQUALITY OF DIFFERENT METRICS FOR ALGEBRAIC POLYNOMIALS arxiv: v1 [math.ca] 5 Jul 2016

Transcription:

arxiv:1608.05015v1 [math.pr] 17 Aug 2016 Cramér type moderate deviations for trimmed L-statistics Nadezhda Gribkova St.Petersburg State University, Mathematics and Mechanics Faculty, 199034, Universitetskaya nab. 7/9, St. Petersburg, Russia Abstract. We establish Cramér type moderate deviation (MD) results for heavy trimmed L-statistics; we obtain our results under a very mild smoothness condition on the inversion F 1 (F is the underlying distribution of i.i.d. observations) near two points, where trimming occurs, we assume also some smoothness of weights of the L- statistic. Our results complement previous work on Cramér type large deviations (LD) for trimmed L-statistics by Gribkova (2016) and Callaert et al. (1982). Keywords: moderate deviations; large deviations; trimmed L-statistics Mathematics Subject Classification (2010): 60F10, 62G30, 62G20, 62G35. 1 Introduction and main results The theory of large deviations is one of the main branches in the probability theory and its applications. There is an extensive literature on this subject for the various classes of statistics, especially for the classical case of sums of independent random variables (see, e.g., Petrov (1975); Saulis and Statulevičius (1991)) and for some types of sums of dependent variables, e.g., for U-statistics (see, e.g., Borovskikh and Weber (2003); Lai et al. (2011), and the references therein). In contrast, there are only a few papers on this topic for L-statistics. In the case of non-trimmed L-statistics with coefficients generated by a smooth on (0, 1) weight function, the Cramér type large and moderate deviations were studied by Vandemaele and Veraverbeke (1982); Aleskeviciene (1991). A highly sharp result on Cramér type large deviations for non-trimmed L-statistics with a smooth weight function was established by Bentkus and Zitikis (1990). For the case of heavy truncated L-statistics, i.e., the case when the weight function is zero outside some interval [α,β] (0,1), a result on Cramér type large deviations was first obtained by Callaert et al. (1982); more recently, the latter result was extended and strengthened in Gribkova (2016), where a different approach than in Callaert et al. (1982) was proposed and implemented. 1

To conclude this introduction we want to mention a paper by Gao and Zhao (2011), where a general delta method in the theory of Chernoff s type large and moderate deviations suggested and illustrated by many examples including M- estimators and L-statistics. Some interesting results on Chernoff s type large deviations for (non-trimmed) L-statistics with smooth weight function were obtained also by Boistard (2007). In this article we supplement our previous work on Cramér type large deviations for trimmed L-statistics (cf. Gribkova (2016)) by some results on moderate deviations. Our approach here is the same as in Gribkova (2016): we approximate the trimmed L-statistic by a non-trimmed L-statistic with coefficients generated by a smooth on (0, 1) weight function, where the approximating (non-trimmed) L-statistic is based on order statistics corresponding to a sample of auxiliary i.i.d. Winsorized observations. We apply a result on moderate deviations due to Vandemaele and Veraverbeke (1982) to the approximating L-statistic and estimate suitably the remainder term of our approximation. Let X 1,X 2,... be a sequence of independent identically distributed (i.i.d.) real-valued random variables (r.v. s) with common distribution function F, and for each integer n 1 let X 1:n X n:n denote the order statistics based on the sample X 1,...,X n. Consider the trimmed L-statistic given by n m n L n = n 1 i=k n+1 c i,n X i:n, (1.1) where c i,n R, k n, m n are two sequences of integers such that 0 k n < n m n n. Put α n = k n /n, β n = m n /n. It will be assumed throughout this paper that α n α, β n β, 0 < α < 1 β < 1, (1.2) as n, i.e. we focus on the case of heavy trimmed L-statistic. Define the left-continuous inverse of F: F 1 (u) = inf{x : F(x) u}, 0 < u 1, F 1 (0) = F 1 (0 + ), and let F n, Fn 1 denote the empirical distribution function and its inverse respectively. We will consider also the trimmed L-statistics with coefficients generated by a weight function: n m L 0 n = n n 1 i=k n+1 c 0 i,n X i:n = 1 βn α n J(u)Fn 1 (u)du, (1.3) where c 0 i,n = n i/n J(u)du, and J is a function defined in an open set I such (i 1)/n that [α,1 β] I (0,1). To state our results, we will need the following set of assumptions. (i) J is Lipschitz in I, i.e. there exists a constant C 0 such that J(u) J(v) C u v, u, v I. (1.4) 2

(ii) There exists a positive ε such that for each t R F 1( α+t logn/n ) F 1( α ) = O ( (logn) (1+ε)), F 1( 1 β +t logn/n ) F 1( 1 β ) = O ( (logn) (1+ε)) (1.5) as n. (iii) max( α n α, β n β ) = O ( ) logn n as n. (iv) For some ε > 0 n m n i=k n+1 c i,n c 0 i,n = O( 1 Define a sequence of centering constants µ n = 1 βn log ε n n ). (1.6) logn α n J(u)F 1 (u)du. (1.7) Since α n α, β n β as n, both variables L 0 n and µ n are well defined for all sufficiently large n. It is well known (see, e.g., Mason and Shorack (1990); Stigler (1974); van der Vaart (1998)) that when the inverse F 1 is continuous at two points α and 1 β, the smoothness condition (1.4) implies the weak convergence to the normal law: n(l 0 n µ n) N(0,σ 2 ), where σ 2 = σ 2 (J,F) = 1 β 1 β α α J(u)J(v)(u v uv)df 1 (u)df 1 (v), (1.8) where u v = min(u,v); we will also use the notation u v for max(u,v). Here and in the sequel, we use the convention that b = when integrating a [a,b) withrespect totheleftcontinuousintegratorf 1. Allalongthearticle, weassume that σ > 0. Define the distribution functions of the normalized L n and L 0 n respectively F Ln (x) = P{ n(l n µ n )/σ x}, F L 0 n (x) = P{ n(l 0 n µ n)/σ x}. (1.9) Let Φ denote the standard normal distribution function. Here is our first result on Cramér type moderate deviations for L n. Theorem 1.1 Suppose that F 1 satisfies condition (ii) and that condition (iii) holds for the sequences α n and β n. In addition, assume that there exists a function J satisfying condition (i) such that (iv) holds for the weights c i,n. Then 1 F Ln (x) = [1 Φ(x)](1+o(1)), F Ln ( x) = Φ( x)(1+o(1)), (1.10) as n, uniformly in the range A x c logn, for each c > 0 and A > 0. 3

The proof of our results is relegated to Section 3. Theorem 1.1 directly implies the following corollary. Corollary 1.1 Let c i,n = c 0 i,n = n i/n J(u)du (k (i 1)/n n +1 i n m n ), where J is a function satisfying (i). Assume that conditions (ii) and (iii) are satisfied. Then relations (1.10) with L n = L 0 n hold true, for each c > 0 and A > 0, uniformly in the range A x c logn. Finally, we state a version of Theorem 1.1, where the scale factor σ/n 1/2 is replaced by Var(L n ). A very mild moment condition will be required now to ensure the existence of the variance of L n. Theorem 1.2 Suppose that the conditions of Theorem 1.1 hold true. In addition, assume that E X 1 γ < for some γ > 0. Then Var(Ln ) σ/ = 1+O ( (logn) (1+2ν), (1.11) n where ν = ε ε (ε, ε are as in (1.6) and (1.5) respectively). Moreover, relations (1.10) remain valid for each c > 0 and A > 0, uniformly in the range A x c logn, if we replace σ/n 1/2 in definition of F Ln (x) (cf. (1.9)) by Var(L n ). 2 Stochastic approximation for L 0 n Let ξ ν = F 1 (ν), 0 < ν < 1, be the ν-th quantile of F and W i denote X i Winsorized outside of (ξ α,ξ 1 β ]. In other words ξ α, X i ξ α, W i = X i, ξ α < X i ξ 1 β, (2.1) ξ 1 β, ξ 1 β < X i. Let W i:n denote the order statistics, corresponding to W 1,...,W n, the sample of n i.i.d. auxiliary random variables. Similarly as in Gribkova (2016), we will approximate L n by a linear combination of the order statistics W i:n with the coefficients generated by the following weight function J(α), u α, J w (u) = J(u), α < u 1 β, (2.2) J(1 β), 1 β < u defined in[0,1]. It isobvious thatif J is Lipschitz ini, i.e. satisfies condition(1.4) withsomepositiveconstant C, thenthefunctionj w islipschitz in[0,1]withsome constant C w C. Consider the auxiliary non-truncated L-statistic given by n L n = n 1 c i,n W i:n = i=1 4 1 0 J w (u)g 1 n (u)du, (2.3)

where c i,n = n i/n (i 1)/n J w(u)du. Define the centering constants 1 = µ Ln 0 J w (u)g 1 (u)du. (2.4) Since W i has the finite moments of any order and because J w is Lipschitz, the distribution of the normalized L n tends to the standard normal law (see, e.g., Stigler (1974)) n( Ln µ L n )/σ(j w,g) N(0,1), where the asymptotic variance is given by σ 2 (J w,g) = 1 1 0 0 J w (u)j w (v)(u v uv)dg 1 (u)dg 1 (v). (2.5) Observe that for u (α,1 β] we have J w (u) = J(u), G 1 (u) = F 1 (u), and that dg 1 (u) 0 for u / (α,1 β]. This yields the equality of the asymptotic variances σ 2 (J w,g) = σ 2 (J,F) = σ 2 (2.6) of the truncated L-statistic L 0 n and the non-truncated L-statistic L n based on the Winsorized random variables. Define the binomial random variable N ν = {i : X i ξ ν }, where 0 < ν < 1. Put A n = N α /n, B n = (n N 1 β )/n. The following lemma provides us a useful representation which is crucial in our proofs. This lemma is proved in (Gribkova, 2016, Lemma 2.1), therefore here we present only its statement. Lemma 2.1 (Gribkova (2016)) where R n = R (1) n +R (2) n, and R (1) n = An α R (2) n = α L 0 n µ n = L n µ Ln +R n, (2.7) 1 Bn J w (u)[fn 1 (u) ξ α]du J w (u)[fn 1 (u) ξ 1 β]du (2.8) 1 β 1 β J(u)[Fn 1 (u) F 1 (u)]du α n 1 β n J(u)[F 1 n (u) F 1 (u)]du. (2.9) Remark 2.1 It should be noted that the method based on the L-statistic approximation was first applied in Gribkova (2016); it can be viewed as a development of the approach proposed in Gribkova and Helmers(2006, 2007, 2014), where the second order asymptotic properties (Berry Esseen bounds and one term Edgeworth type expansions) for (intermediate) trimmed means and their Studentized and 5

bootstrapped versions were established. In the articles mentioned we constructed U-statistic type approximations for (intermediate) trimmed means, where we used sums of auxiliary i.i.d. Winsorized observations as the linear terms; in order to get the second (quadratic) U-statistic terms, we applied some special Bahadur Kiefer representations of von Mises statistic type for (intermediate) sample quantiles (cf. Gribkova and Helmers (2012)). 3 Proofs Proof of Theorem 1.1. Obviously it suffices to prove the first of relations(1.10). Set n m V n = L n L 0 n = n n 1 (c i,n c 0 i,n )X i:n. (3.1) Lemma 2.1 and relation (3.1) yield i=k n+1 L n µ n = L n µ L n +R n +V n. (3.2) An application of the classical Slutsky argument to (3.2) gives that, for δ > 0, 1 F Ln (x) is bounded above and below by and P{ n( L n µ L n )/σ > x 2δ}+P{ n R n /σ > δ}+p{ n V n /σ > δ} (3.3) P{ n( L n µ L n )/σ > x+2δ} P{ n R n /σ > δ} P{ n V n /σ > δ} (3.4) respectively. Fix arbitrary c > 0 and A > 0. Set δ = δ n = ( log(n + 1) ) 1/2 ε 1, where 0 < ε 1 < ε ε, and ε, ε are as in conditions ( ii) and (iv) respectively(cf (1.5)-(1.6)). From (3.3) and (3.4) it immediately follows that to prove our theorem it suffices to show that P{ n( L n µ Ln )/σ > x±2δ n } = [1 Φ(x)](1+o(1)), (3.5) P{ n R n /σ > δ n } = [1 Φ(x)]o(1), (3.6) P{ n V n /σ > δ n } = [1 Φ(x)]o(1), (3.7) uniformly in the range A x c logn. Let us prove (3.5). Observe that L n represents a non-truncated L-statistic based on the sample W 1,...,W n of i.i.d. bounded random variables; and since its weight function J w is Lipschitz in [0,1], we can apply a result by Vandemaele and Veraverbeke (1982). Set = 2sup n 1 δ n = 2/(log2) 1/2+ε 1. Since E W i p M < (for each p > 0 with some M > 0), by Theorem 1 (i) of Vandemaele and Veraverbeke (1982) P{ n( L n µ Ln )/σ > x±2δ n } = [1 Φ(x±2δ n )](1+o(1)), (3.8) 6

uniformly with respect to x such that (A + ) x ± 2δ n c 1 logn, where we may take c 1 > c. Hence (3.8) holds uniformly in the range A x c logn for all sufficiently large n. Further, we apply Lemma A1 of Vandemaele and Veraverbeke (1982) (in which the required asymptotic property of Φ is given in a very convenient form). Since δ n logn = o(1), due to that lemma we obtain that 1 Φ(x±δ n ) = [1 Φ(x)](1+o(1)) uniformly in the range A x c logn. Summarizing, we find that (3.5) is valid, uniformly in the range required. Let us prove (3.6). First, we argue similarly to the corresponding place in (Gribkova, 2016, Theorem 1.1). Let I (j) 1 and I (j) 2 denote the first and the second terms of R (j) n (cf. (2.8) (2.9)) respectively, j = 1,2. Then R n = I (1) 1 I (1) 2 +I (2) I (2) 2 and P{ n R n /σ > δ n } 2 k=1 Notice that for x [ A,c logn] P{ n I (1) k /σ > δ n/4}+ 2 k=1 1 P{ n I (2) k /σ > δ n/4}. (3.9) 1 1 Φ(x) 1 1 Φ(c logn) c lognn c2 /2. (3.10) Hence, it suffices to show that for each positive C (in particular for C = σ/4), P{ ( ) n I (j) k > Cδ n} = o (logn) 1/2 n c2 /2, k,j = 1,2, (3.11) as n. We will prove (3.11) for I (1) 1 and I (2) 1 (the treatment of I (1) 2 and I (2) 2 is similar and therefore omitted). Consider I (1) 1. First, note that if α < A n, then max u (α,an) Fn 1 (u) ξ α = ξ α X [nα]+1:n ξ α X [nα]:n, as Fn 1 is monotonic. Here and in what follows [x] represents the greatest integer function. Similarly we find that if A n α, then max u (An,α) Fn 1 (u) ξ α = X [nα]:n ξ α. Furthermore, by the Lipschitz condition for J, there exists a positive K such that max u [0,1] J w (u) sup u I J(u) K. This yields An I (1) 1 = J w (u)[fn 1 (u) ξ α ]du K A n α X [nα]:n ξ α. (3.12) α Let U 1,...,U n be a sample of independent (0,1)-uniform distributed random variables, U i:n the corresponding order statistics. Set M α = {i : U i α}. Since the joint distribution of X i:n and N α coincides with the joint distribution of F 1 (U i:n ) and M α, i = 1,...,n, we have P{ n I (1) 1 > Cδ n } P{n 1/2 M α αn F 1 (U [nα]:n ) F 1 (α) > Cδ n } P 1 +P 2, (3.13) 7

where P 1 =P{ M α αn > c 1 nlogn}, P 2 =P{ F 1 (U [nα]:n ) F 1 (α) > C(log(n+1)) (1+ε 1) }, where we choose c 1 > c. Here and in the sequel, C stands for a positive constant not depending on n, which may change its value from line to line. For P 1 by Bernstein s inequality we obtain c 2 1 logn P 1 2exp( h n ), (3.14) with h n = c2 1 logn. Hence P 2[1+O( logn/n)] 2 1 = o ( (logn) 1/2 n ) c2 /2. Next we estimate P 2 on the r.h.s. in (3.13). To shorten notation, let k α = [nα], p α = EU kα:n = k α /(n+1), and note that 0 < α p α (α+1)/(n+1) = O(1/n). Define V n (p α ) = n(u kα:n p α ) and let E denote the event { V n (p α ) c 1 pα logn}, where as before c 1 is an arbitrary number such that c 1 > c. Put M n = F 1 (p α + c 1 pα logn/n) F 1 (p α ) F 1 (p α c 1 pα logn/n) F 1 (p α ). Then we have P 2 P{M n > C(log(n+1)) (1+ε 1) }+P{E}. (3.15) Bycondition(ii), andbecauseofε 1 > ε, thefirstprobabilityonther.h.s. in(3.15) is zero for all sufficiently large n. In order to estimate the second probability on the r.h.s. in (3.15), we can apply Inequality 1 given in (Shorack and Wellner, 1986, page 453). Then we obtain [ P{E} exp c 2 logn ] [ 1 +exp c 2 logn ] 1 ψ(t n ), (3.16) 2 2 where ψ is the function defined in (Shorack and Wellner, 1986, page 453, formula (2)), t n = c 1. Since t n n 0 as n, hence t n > 1 for all logn sufficiently large n, and by Proposition 1 in (Shorack and Wellner, 1986, page 455, relation (12)), we find that ψ(t n ) 1 1+2t n/3. This and relation (3.16) together imply that [ P{E} 2exp c 2 logn ] 2 = 2n c2 2 /2, (3.17) 2 for each c 2 such that c < c 2 < c 1 and for all sufficiently large n. Summarizing, we get that P 2 = o ( (logn) 1/2 n ) c2 /2, and the desired bound (3.11) for I (1) 1 follows. Next we prove (3.11) for I (2) 1. Define a sequence of intervals Γ n = [α α n,α α n +1/n), then we obtain α I (2) 1 = J(u)[Fn 1 (u) F 1 (u)]du K α n α D n, (3.18) α n where D n = max i:i/n Γn X i:n F 1 (i/n) X i:n F 1 ((i 1)/n). By condition (iii), the estimate (3.18) implies that P{ n I (2) 1 > Cδ n} P{D n,u > C(log(n+1)) (1+ε 1) }, (3.19) 8

where D n,u = max i:i/n Γn F 1 (U i:n ) F 1 (i/n) F 1 (U i:n ) F 1 ((i 1)/n). Define V n (p i ) = n(u i:n p i ), where p i = EU i:n = i/(n+1), and let E i denote the event { V n (p i ) c 1 pi logn}, where now we take c 1 such that c 2 1 > c 2 +1/2. Then by condition (ii), we find that P{ n I (2) 1 > Cδ n} P{ E i } P(E i ). (3.20) i:i/n Γ n i:i/n Γ n Similarly as before, using Inequality 1 from Shorack and Wellner (1986), for each i : i/n Γ n, we obtain that [ P{E i } 2exp c 2 logn ] 2 = 2n c2 2 /2, 2 with some c 2 such that c 2 1 > c2 2 > c2 + 1/2, and since by condition (iii) {i : i/n Γ n } = O( nlogn), it follows from (3.20) that P{ n I (2) 1 > Cδ n} = o ( (logn) 1/2 n ) c2 /2. This completes the proof of (3.11), which implies that (3.6) holds true uniformly in the range A x c logn. Let us finally prove that (3.7) is valid uniformly in the range A x c logn. By condition (iv), there exists b > 0 such that n Vn b(logn) (1/2+ ε)( X ) (kn+1):n X (n mn):n, for all sufficiently large n. Thus, P ( n V n /σ > δ n ) P ( X(kn+1):n X (n mn):n > C(log(n+1)) ε ε 1 ) P 3 +P 4, where P 3 = P ( X (kn+1):n > C(log(n+1)) ε ε 1), P4 = P ( X (n mn):n > C(log(n+ 1)) ε ε 1), and ε ε1 > 0 by the choice of ε 1. Let us estimate P 3 (the treatment for P 4 is same and therefore omitted). We have P 3 = P ( F 1 (U (kn+1):n) > C(log(n+1)) ε ε 1 ) P ( F 1 (U (kn+1):n) F 1 (α) + F 1 (α) > C(log(n+1)) ε ε 1 ), = P ( F 1 (U (kn+1):n) F 1 (p αn ) > C(log(n+1)) ε ε 1 (1+o(1)) ), (3.21) where p αn = EU (kn+1):n. Arguing similarly as when estimating P 2 (cf. (3.15)- (3.17)), we find that the r.h.s. of (3.21) is o ( (logn) 1/2 n c2 /2 ). This completes the proof of (3.7) and the theorem. Proof of Theorem 1.2. Let us first prove relation (1.11). By Lemma 2.1 and relation (3.2), we have Var(L n ) = Var( L n )+Var(R n +V n )+2cov( L n,r n +V n ). Since W i are bounded, the conditions in (Vandemaele and Veraverbeke, 1982, Theorem 2(ii), page 431) are satisfied, and hence σ 1 n 1/2 Var( L n ) = 1+O(n 1/2 ) 9

Further, we have n cov( L n,r n +V n ) n[var( L n )Var(R n +V n )] 1/2 = σ[nvar(r n +V n )] 1/2 (1+O(n 1/2 )). The latter three relations imply that in order to prove (1.11), it suffices to show that nvar(r n +V n ) = O ( (logn) (1+2ν), (3.22) where ν = ε ε, and ε, ε are the constants from conditions (ii) and (iv) respectively. We have [ 2 nvar(r n +V n ) ne(r n +V n ) 2 5n k,j=1 E ( I (j) ) ] 2 k +EVn 2, (3.23) where I (j) k are as in (3.9)-(3.11). We will show that ne ( I (j) ) 2 1 = O ( (logn) 2(1+ε)), ne ( I (j) ) 2 2 = O ( (logn) (1+2ε)), j = 1,2, (3.24) and that nev 2 n = O( (logn) (1+2 ε)). (3.25) Relations (3.23)-(3.25) imply the desired bound (3.22). Let us prove the first relation in (3.24). We will consider in detail only the case k = 1 (the treatment in the case k = 2 is same and therefore omitted). Let as before k α = [αn] and k n = α n n. By (3.12) and the Schwarz inequality, we have E ( I (1) ) 2 1 K 2 [E(A n α) 4 E(X kα:n ξ α ) 4 ] 1/2 =K 2 n 2 [E(N α αn) 4 E(X kα:n ξ α ) 4 ] 1/2. By well-known formula for 4-th moments of a binomial random variable, we have E(N α αn) 4 = 3α 2 (1 α 2 )n 2 (1+o(1)). Thus, there exists a positive constant C independent of n such that ne ( I (1) 1 ) 2 C[E(Xkα:n ξ α ) 4 ] 1/2 (3.26) for all sufficiently large n. Fix arbitrary c > 0, A > 0. Let p α, V n (p α ) and the event E be as when estimating P 2 in the proof of Theorem 1.1, and c 1 is an arbitrary constant such that c 1 > c. Then we can write E(X kα:n ξ α ) 4 = E[(F 1 (U kα:n) F 1 (α)) 4 1 E ]+E[(F 1 (U kα:n) F 1 (α)) 4 1 E ]. By a well known property of the order statistics (see, e.g., (Gribkova, 1995, Thoerem 1), and due to our moment assumption, E F 1 (U kα:n k is bounded from above for each k > 0. Then by condition (ii), the latter quantity is of the order O ( (logn) 4(1+ε) +P(E) ) = O ( (logn) 4(1+ε)) (cf. (3.17) ). This bound and (3.26) together imply that ne ( I (1) ) 2 1 = O ( (logn) 2(1+ε)). 10

Consider I (2) 1. By condition (iii), there exists L > 0 such that (α n α) 2 Llogn/n, for all sufficiently large n. Then in view of (3.18) we obtain ne ( I (2) 1 ) 2 nk 2 (α n α) 2 ED 2 n LK2 logned 2 n. (3.27) Hence, to get the second bound in (3.24), it suffices to show that ED 2 n = O( (logn) 2(1+ε)), (3.28) and since D d n,u = D n,u, it suffices to prove (3.28) for D n,u. Let p i, V n (p i ) and the event E i be as in the proof of Theorem 1.1, when estimating of I (2) 1 (cf. (3.19)- (3.20), where we now take c 1 such that c 2 1 > c 2 +1/2. Let 1 Ei denote the indicator of the event E i Then we have ED 2 n,u = E ( D 2 n,u1 i:i/n Γn E i) +E ( D 2 n,u 1 i:i/n Γn E i). (3.29) By condition (ii) the first term on the r.h.s. in (3.29) is O ( (logn) 2(1+ε)), and since ED k n,u is bounded from above for each k > 0, there exists a positive constant M, not depending on n, such that for all sufficiently large n E ( ) { } D 2 n,u1 i:i/n Γn E i MP M i:i/n Γ n E i i:i/n Γ n P{E i }. (3.30) Similarly to the proof of Theorem 1.1 (cf. (3.20)), we find that the magnitude on the r.h.s. in (3.30) is of the order o ( (logn) 1/2 n c2 /2 ). Summarizing, we obtain the validity of the second relation in (3.24). We now turn to the proof of (3.25). We have ( n m n nevn 2 n 1 i=k n+1 ) 2E ( c i,n c 0 i,n X 2 k n+1:n Xn m n:n) 2. Due to our moment assumption we have E ( X 2 k n+1:n X2 n m n:n) = O(1), and by condition (iv) we get ( n m n i=k c n+1 i,n c 0 i,n )2 = O ( n(logn) (1+2 ε)). These bounds and the latter displayed estimate yield (3.25). Thus, relation (1.11) is proved. In order to complete the proof of our theorem, it remains to argue the possibility of the replacement σ/n 1/2 by Var(L n ) in (1.10) without affecting the result. We prove it for the first relation in (1.10), for the second one it will then follow from the first one if we replace c i,n by c i,n. Fix arbitrary c > 0 and A > 0, set λ n = σ 1 n 1/2 Var(L n ) and write P ( (L n µ n )/ Var(L n ) > x ) 1 Φ(x) = 1 F L n (λ n x) 1 Φ(λ n x) 1 Φ(λ n x) 1 Φ(x). (3.31) Set B = Asup n N λ n. Since λ n 1, the number B exists. Hence by Theorem 1.1, the first ratio on the r.h.s. in (3.31) tends to 1 as n, uniformly in x such that B λ n x c logn, hence in particular uniformly in the range A x c logn. Furthermore, we see that λ n 1 1/2 logn 0, which is due to the fact that λ n 1 1/2 = O ( (logn) (1/2+ν)). Hence, by Lemma A1 from Vandemaele and Veraverbeke (1982), the second ratio on the r.h.s. in (3.31) also tends to 1, uniformly in the range A x c logn. The theorem is proved. 11

References Aleskeviciene, A. (1991). Large and moderate deviations for L-statistics. Lithuanian Math. J., 33:145 156. Bentkus, V. and Zitikis, R.(1990). Probabilities of large deviations for L-statistics. Lithuanian Math. J., 30:215 222. Boistard, H. (2007). Large deviations for L-statistics. Statist. & Decis., 25:89 125. Borovskikh, Y. and Weber, N.(2003). Large deviations of U-statistics. I-II. Lithuanian Math. J., 43: 11 33, 241 261. Callaert, H., Vandemaele, M., and Veraverbeke, N. (1982). A Cramér type large deviations theorem for trimmed linear combinations of order statistics. Comm. Statist. Th. Meth., 11: 2689 2698. Gao, F. and Zhao, X. (2011). Delta method in large deviations and moderate deviations for estimators. Ann. Statist., 39: 1211 1240. Gribkova, N. (1995). Bounds for absolute moments of order statistics. In (Skorokhod, A.V., Borovskikh, Yu.V. eds.) Exploring Stochastic Laws: Festschrift in Honor of the 70th Birthday of Acad. V.S. Korolyuk, pages 129 134. VSP, Utrecht. Available at arxiv:1607.08066v2[math.pr]. Gribkova, N. V. (2016). Cramér type large deviations for trimmed L-statistics. Probab. Math. Statist. (to appiar). Available at arxiv:1507.02403[math.pr]. Gribkova, N. V. and Helmers, R. (2006). The empirical Edgeworth expansion for a Studentized trimmed mean. Math. Methods Statist., 15(1): 61 87. Gribkova, N. V. and Helmers, R. (2007). On the Edgeworth expansion and the M out of N bootstrap accuracy for a Studentized trimmed mean. Math. Methods Statist., 16(2): 142 176. Gribkova, N. V. and Helmers, R. (2012). On a Bahadur Kiefer representation of von Mises statistic type for intermediate sample quantiles. Probab. Math. Statist., 32(2): 255 279. Gribkova, N. V. and Helmers, R. (2014). Second order approximations for slightly trimmed means. Theory Probab. Appl., 58(3): 383 412. Lai, T., Shao, Q., and Wang, Q. (2011). Cramér type moderate deviations for Studentized U-statistics. ESAIM: Probability and Statistics, 15: 168 179. Mason, D. and Shorack, G. (1990). Necessary and sufficient conditions for asymptotic normality of trimmed L-statistics. J. Statist. Plan. Inference, 25: 111 139. Petrov, V. V. (1975). Sums of independent random variables. Springer-Verlag, New York. 12

Saulis, L. and Statulevičius, V.(1991). Limit theorems for large deviations. Kluwer Academic Publishers, Dordrecht. Shorack, G. R. and Wellner, J. A. (1986). Empirical processes with application in statistics. Wiley, New York. Stigler, S. M. (1974). Linear functions of order statistics with smooth weight functions. Ann. Statist., 2: 676 693. van der Vaart, A. (1998). Asymptotic statistics. Cambridge Univ. Press, Cambridge. Vandemaele, M. and Veraverbeke, N. (1982). Cramér type large deviations for linear combinations of order statistics. Ann. Probab., 10: 423 434. 13