The Central Limit Theorem Under Random Truncation

Size: px
Start display at page:

Download "The Central Limit Theorem Under Random Truncation"

Transcription

1 The Central Limit Theorem Under Random Truncation WINFRIED STUTE and JANE-LING WANG Mathematical Institute, University of Giessen, Arndtstr., D-3539 Giessen, Germany. Department of Statistics, University of California, Davis, CA 9566, USA Under left truncation data (X i, Y i ) are observed only when Y i X i. Usually the distribution function F of the X i is the target of interest. In this paper we study linear functionals ϕdf n of the nonparametric MLE of F, the Lynden- Bell estimator F n. A useful representation of ϕdf n is derived which yields asymptotic normality under optimal moment conditions on the score function ϕ. No continuity assumption on F is required. As a by-product we obtain the distributional convergence of the Lynden-Bell empirical process on the whole real line. Running Title: The CLT under random truncation Key words: CLT; Lynden-Bell integral; Truncated data. Introduction and main results In this paper we provide some further methodology for statistical analysis of data which are truncated from the left. To be more specific, let (X i, Y i ), i N, denote a sample of independent bivariate data such that, for each i, X i is also independent of Y i. Denote with F and G, respectively, the unknown distribution functions of X and Y. Typically, F is the target of interest. Now, under left truncation, X i is observed only when Y i X i. As a consequence, the empirical distribution of the X s is unavailable and cannot serve as a basic process to compute other statistics. The nonparametric maximum likelihood estimator of F for left-truncated data was first derived by Lynden-Bell (97). Its first mathematical investigation may be attributed to Woodroofe (985), who also reviewed some examples of truncated data from astronomy and economics. See also Wang (989) for applications in the analysis of AIDS data. Now, denoting with n the number of data which are actually observed, we have, by the Strong Law of Large Numbers (SLLN), n N α P(Y X) as N with probability one.

2 Without further mentioning we shall assume that α >, because otherwise nothing will be observed. Of course, α will be unknown. Conditionally on n, the observed data are still independent, but the joint distribution of X i and Y i becomes H (x, y) = P(X x, Y y Y X) = α G(y z)f (dz). (,x] The marginal distribution of the observed X s thus equals (x) α G(z)F (dz). (.) (,x] It may be consistently estimated by the empirical distribution function of the known X s: Fn(x) = n {Xi x}, x R. n i= The problem, however, is one of reconstructing F and not from the available data (X i, Y i ), i n. A crucial quantity in this context is the function where C(z) = P(Y z X Y X) = α G(z)[ F (z )], (.) F (z ) = lim x z F (x) is the left continuous version of F and F {z} = F (z) F (z ) is the F -mass at z. The function C may be consistently estimated by C n (z) = n n {Yi z X i }. It is very helpful to express the cumulative hazard function of F, F (dz) Λ(x) = F (z ), (,x] i= in terms of estimable quantities. For this, let a G = inf{x : G(x) > } be the largest lower bound for the support of G. Similarly for F. From (.) and (.) we obtain (a G,x] F (dz) F (z ) = (a G,x] (dz) C(z). (.3)

3 Provided that a G a F and F {a F } =, the left-hand side equals Λ(x). Otherwise this is no longer true, and as Woodroofe (985) pointed out, F cannot be fully recovered from the available data. The situation is similar for right-censored data. See Stute and Wang (993) for a detailed discussion for (upper) boundary effects there. Throughout the paper we shall therefore assume that a G a F and F {a F } =. (.4) If a G < a F the second assumption is superfluous, while it is automatically satisfied when F is continuous. For the moment, however, no other assumptions such as continuity of F or G will be needed. In differential terms, equation (.3) leads to F (dx) = ( F ()) (dx). (.5) The Lynden-Bell (97) estimator F n of F is obtained as the solution of the so-called self-consistency equation, i.e., as the solution of (.5) after having replaced and C with n and C n, respectively: F n (dx) = ( F n ()) n(dx) C n (x). (.6) Solving for F n yields the product-integral representation of F n : [ F n (x) = F ] n{x i }. (.7) C n (X i ) distinctx i x If there are no ties among the X s, (.7) simplifies to become F n (x) = [ ] ncn (X i ). (.8) nc n (X i ) X i x Note that nc n (X i ) so that each ratio is well defined. Since in this paper our objective will be to study general linear statistics based on F n, namely Lynden-Bell integrals ϕdf n, we also introduce the Lynden-Bell weights W in attached to each datum in the X-sample. For this, denote with X :n <... < X m:n the m distinct order statistics, with m possibly strictly less than n. From (.7) we obtain W in F n {X i:n } = [ F n (X i :n )] n{x i:n } C n (X i:n ) (.9) 3

4 and ϕdf n = m W in ϕ(x i:n ). (.) i= For ϕ = (,x], we are back at F n (x). As to other ϕ s we refer to Stute and Wang (993) or Stute (4), who considered possible applications of empirical integrals in the context of right censored data. More general statistical functionals often admit expansions, in which the leading term is of the form ϕdf n, with ϕ denoting the associated influence function. Since for a fully observable data set with Fn e being the classical empirical distribution function, ϕdfn e is just a sample mean to which, under a second moment assumption, the Central Limit Theorem (CLT) applies, distributional convergence of ϕdf n therefore constitutes an extension of the CLT to the left-truncation case. The corresponding SLLN has been studied in papers by He and Yang (998a,b). The CLT for censored data is due to Stute (995). In the present situation, the CLT is much more elusive than for randomly censored data. In the censored data case the right tails create technical difficulties, but not the left. For left truncation, however, both sides create problems. This is already seen now through the functions C and C n which decrease to zero on the left and on the right tail, and with C n appearing in denominators. The only trivial bound is nc n (X i ), which keeps everything from being not well defined. Moreover, C and C n are not monotone, again contrary to the censored data case, where the role of the C s is played by survival functions. This non-monotone feature of C n creates additional technical complications for truncated data. Worse than that, C n may become zero between the data points also in the central part. Keiding and Gill (99), who recognized the danger of these sets, have called these holes the empty inner risk sets. Not all authors seem to know about these problems since a detailed study of the C n process is sometimes missing. A consequence of these empty inner risk sets, as revealed by the representation (.7), is the loss of mass of F n after the first such hole. More specifically, suppose that, in terms of Keiding and Gill (99), a risky hole exists at X j. This means that X j is not covered by any other pair (X i, Y i ). Hence nc n (X j ) = nfn{x j }. From (.7) we get that all data points right to X j have mass zero under F n. In proofs this excludes one to incorporate exponential representations of the weights. To circumvent this difficulty we first study an asymptotically equivalent estimator ˆF n. This estimator is defined through: ϕd ˆF n = m i= ϕ(x i:n ) n{x i:n } C n (X i:n ) i j= [ nf ] n{x j:n }. (.) nc n (X j:n ) + The extra summand in the denominator allows for contributions of data which have holes on their left. As we shall see in a small simulation study 4

5 it may have a robustifying effect resulting in smaller MSE for small sample sizes. Asymptotically ϕdf n and ϕd ˆF n are equivalent at the n / - rate, which facilitates the asymptotic theory for ϕdf n. Another technical problem is caused by possible ties. In such a situation (.8) does not apply. Lemma. will show how the general case covered by (.7) and (.9) may be traced back to the case of a continuous. The main result of this paper, Theorem., provides a representation of ϕdfn as a sum of i.i.d. random variables under minimal assumptions on ϕ and the truncation mechanism. Usually linear i.i.d. representations of complicated estimators will include the Hájek projection of the statistic of interest. In the case of (.), however, this projection can be computed only up to remainders, and this is exactly what Theorem. does. More precisely, the proof of Theorem. proceeds in two steps. In the first step we have to identify all error terms which are negligible. The leading terms will turn out to be V-statistics. See Serfling (98). Finally, an application of the Hájek projection to the leading terms yields the desired i.i.d. representation. Needless to say asymptotic normality follows immediately. We shall also add some interesting comments on a so-called uniform representation. Proofs will be given in Section 3. Theorem. will hold under the following two assumptions: (A) (i) ϕ /GdF < (ii) df G < Condition (ii) already appeared in Woodroofe (985) in his study of the Lynden-Bell process, i.e., when he considered indicators ϕ = (,x ]. Nontechnically speaking, it is needed to guarantee enough information in the left tails to estimate F at the rate n /. Under slightly stronger assumptions, Stute (993) obtained an almost sure representation with sharp bounds on the remainder, again for indicators. See also Chao and Lo (988). Condition (i) guarantees, among other things, that the leading terms in the i.i.d. representation admit a finite second moment so that asymptotic normality holds. Since G it implies ϕ df <, which is the standard finite moment assumption when no truncation occurs. When ϕ has a finite second F -moment and is locally bounded in a neighborhood of a G, then (i) is implied by (ii). Note also that (i) and (ii) are always satisfied when a G < a F and ϕ df <. A CLT for truncated data is also contained in Sellero et al. (5). Apart from continuity assumptions they also need conditions which, in our notation, require finiteness of the integral ϕ (x)( F (x)) 5 F (dx), where ϕ ϕ. Since, however, this integral equals infinity for constant ϕ, their result is not applicable to bounded ϕ s, not to mention ϕ s which 5

6 increase to infinity as x. Rather, finiteness of the above integral is only obtained for ϕ s which converge to zero fast enough in the right tails. The focus of this paper is, however, on distributional convergence for which (A) will suffice. Theorem. is formulated for the case when F is continuous. This guarantees that among the observed X s there will be no ties, with probability one. In such a situation we obtain ϕd ˆF ϕ(x)γn (x) n = C n (x) n(dx), (.) with γ n (x) = exp n [ ln ] nc n (y) + Fn(dy). (.3) At the end of this section we shall show how general Lynden-Bell integrals may be traced back to the present case. Theorem.. Under Assumptions (A) and (.4), assume that F is continuous. Then we have ψ(y) ϕ[df n df ] = C(y) [F n(dy) ] Cn (y) C(y) ψ(y) + o C P (n / ), (y) where ψ(y) = ϕ(y)( F (y)) {y<x} ϕ(x)( F (x)) (dx) = {y<x} [ϕ(y) ϕ(x)]f (dx). For indicators ϕ = (,x ] the leading term already appears in Theorem in Stute (993). Remark.. If one checks the proof of Theorem. step by step, the following fact will be revealed: If rather than a single ϕ, one considers a collection {ϕ}, then the error terms are uniformly small in the sense that the remainder is o P (n / ) uniformly in ϕ, provided that ϕ ϕ for some ϕ satisfying ϕ df <. Actually, all remainders may be bounded from G above in absolute values by replacing ϕ by ϕ. Compared with Sellero et al. (5), no VC property for the ϕ s is needed for a representation as a V-statistic process. For the i.i.d. representation one has to guarantee that the errors in the Hájek projection are also uniformly small. These errors, however, form a class of degenerate U-Statistics. This kind of process was studied in Stute (994), and the achieved bounds are useful to handle the error terms in the second half of the proof. Details are omitted. 6

7 Corollary.. Under the assumptions of Theorem., we have n / ϕ[df n df ] N (, σ ) in distribution, with σ ψ(x) = Var C(X) X Y ψ(y) C (y). It is not difficult to see that σ < under (A). Finally, if in Remark. we take for {ϕ} the class of all indicators ϕ = (,x] and set ϕ, we obtain the following Corollary. Corollary.. Under df G have F n (x) F (x) = < and (.4), and for a continuous F, we ψx C [df n d ] Cn C ψ C x + o P (n / ) uniformly in x, where ψ x is the ψ belonging to ϕ = (,x]. Remark.. To make the point clear, Corollary. provides a representation which holds uniformly on the whole real line and not only on subintervals (, b] with b < b F = sup{x : F (x) < }, as is usually done in the literature. A major technical problem for proving Theorem. for a general F is caused by the fact that for discontinuous F ties may arise. As before, denote with X :n <... < X m:n the m ordered distinct data in the observed X-sample. To circumvent ties one may use the fact that each X i may be written in the form X i = (U i ), where U,..., U n is a sample of independent random variables with a uniform distribution on (,) and in the quantile function of. The construction of the U s is similar to the construction in Lemma.8 of Stute and Wang (993), with H there replaced by here. For the following recall that a quantile function is continuous from the left. Furthermore, with probability one, is also right-continuous at each U i. With this in mind, one can see that the corresponding truncating sample for the U i s are (Y i ), i n. If we denote with Cn U the C n -function corresponding to the pseudo observations (U i, (Y i )), i n, and if we let U i:n <... < U idi :n denote the ordered U s that satisfy (U ij:n ) = X i:n, with d i = nfn{x i:n }, then nc U n (U i:n ) = nc n (X i:n ) (.4) 7

8 and nc U n (U ij:n ) = nc U n (U i,j :n ), for j d i. (.5) Also note that the Lynden-Bell estimator Fn U for the pseudo observations satisfies (.8), i.e., Fn U (u) = [ ] nc U n (U i ). nc U U i u n (U i ) Introducing the function ϕ = ϕ, the analogue of (.) becomes ϕ df U n = = m d i i= j= ϕ( (U ij:n )) F U n (U ij:n ) nc U n (U ij:n ) m d i ϕ(x i:n ) i= j= Fn U (U ij:n ). ncn U (U ij:n ) In Lemma. we will show that for each i m and every j d i we have F U n (U ij:n ) nc U n (U ij:n ) = F n(x i :n ), (.6) nc n (X i:n ) where X :n =. It follows from (.9), (.) and (.6) that ϕ dfn U = ϕdf n with probability one. In conclusion the study of Lynden-Bell integrals may be traced back to the case when the variables of interest are uniform on (,) and therefore, with probability one, have no ties. At the same time our handling of ties does not require external randomization but maintains the product limit structure of the weights and hence of the estimators. Lemma.. For each i m and j d i, equation (.6) holds true. Simulations It is interesting to compare the small sample size behavior of F n and ˆF n. In a simulation study we considered ϕ(x) = x, i.e., the target was the mean lifetime of X. This ϕ is the canonical score function in the classical CLT. Recall that through truncation there is a sampling bias which would result in an upward bias if we would take the empirical distribution function of the X s and not F n or ˆF n. Introducing more or less complicated weights has the effect, among other things, that compared with the empirical distribution function the bias is reduced. 8

9 In the following we report on some simulation results which are part of a much more extensive study. Typically, for this ϕ, ϕd ˆF n outperforms ϕdfn in terms of the Mean Squared Error (MSE) when n 4 and truncation is heavy. For larger n(n 4) the difference is negligible. MSE was computed through Monte Carlo. In the following both X and Y were exponentially distributed with parameter : 3 Proofs Table. Comparison of MSE Sample Size MSE ( ϕdf n ) MSE( ϕd ˆF n ) n = n = n = n = n = n = n = n = 8.. n = 9.. n =..9 To prove Theorem., note that under continuity of F (.8) applies. Also we may assume w.l.o.g. that all data are nonnegative. Hence all appearing integrals will be over the positive real line hereafter. In our first lemma we provide a bound for the function C/C n. This will be needed in proofs to handle negligible terms. Though, by Glivenko-Cantelli, C n C uniformly, bounding the ratio is more delicate in view of possible holes and the non-monotonicity of C and C n. Lemma 3.. Assume F is continuous. For any λ such that αλ one has ( ) C(X i ) P sup X i :X i b C n (X i ) λ λe exp[ G(b)αλ]. Proof. The proof is similar to that of Lemma. in Stute (993), which provides an exponential bound for the sup extended over the left tails X i b rather than the right tails. Together with the above mentioned bound from Stute (993), Lemma 3. immediately implies sup i n C(X i ) C n (X i ) = O P() as n. (3.) 9

10 Assertion (3.) will be of some importance in forthcoming proofs, as it will allow us to replace the random C n appearing in denominators with the deterministic C. We are now ready to expand ϕd ˆF n. By (.) and (.3), ϕd ˆF n = ϕ(x) C n (x) exp n ln ϕ(x) C n (x) γ n(x) n(dx). [ ] nc n (y) + n(dy) F n(dx) Set γ(x) = F (x) = exp From Taylor s expansion we obtain x C(y). where γ n (x) = γ(x) + e n(x) [B n (x) + D n (x) + D n (x)], B n (x) = n D n (x) = D n (x) = ln [ ] nc n (y) + n(dy) + Fn(dy), C(y) [C n (y) + n C(y)] [C n (y) + n ]C(y) n(dy) Fn(dy) C n (y) +, n and n (x) is between the exponents of γ n (x) and γ(x). Particularly, we have n (x). Setting S n = S n = S n3 = ϕ(x) C n (x) e n(x) B n (x) n(dx) ϕ(x)γ(x) C n (x) n(dx) ϕ(x) C n (x) e n(x) D n (x) n(dx)

11 and we thus get S n4 = ϕ(x) C n (x) e n(x) D n (x) n(dx), ϕd ˆF n = S n + S n + S n3 + S n4. (3.) In the next lemma we study the functions D n and D n more closely. To motivate the following note that for each fixed x such that F (x ) < D n (x ) with probability one. Actually, by standard Glivenko-Cantelli arguments, D n (x) with probability one uniformly on x x. Similarly, when we consider the standardized processes x n / D n (x), it is easy to show their distributional convergence in the Skorokhod space D[, x ]. Things, however, change if we study D n on the whole support of. Since = C(y) we cannot expect uniform convergence on the whole support of. The situation is similar for the cumulative hazard function Λ, where uniform convergence of the Nelson-Aalen estimator Λ n may be obtained only on compact subsets of the support of F. When one evaluates these processes at x = X i it is known though, that the uniform deviation between Λ n and Λ does not got to zero but remains at least bounded. See Theorem. in Zhou (99). Similar things turn out to be true for D n and D n, as Lemma 3. will show. Our proofs are different though since compared with Zhou (99) we shall apply a truncation technique which in proofs guarantees that the sup of D n and D n are bounded on large but not too large sets. Lemma 3.. We have and sup D n (X i ) = O P () (3.3) i n sup D n (X i ) = O P (). (3.4) i n

12 Proof. Assume w.l.o.g. that has unbounded support on the right. Otherwise, replace by b F = sup{x : (x) < }. For a given ε >, one may find some small c = c ε and a sequence a n such that (a n ) = c n and P(X n:n a n ) ε. Actually, this follows from P(X n:n a n ) = [ c n ]n exp( c). It therefore suffices to bound D n and D n on (, a n ]. By Lemma 3., we may replace C n in the denominator by C. Hence with large probability and up to a constant factor, (3.4) is bounded from above by a n C n (y) C(y) C (y) n(dy) + n a n n(dy) C (y). Using (ii) of (A), it is easily seen that the expectation of the second term is bounded as n. The expectation of the first term is less than or equal to a n E C n (y) C(y) + n C (y) a n n C / (y) + n C (y) = O(). This proves (3.4). As for D n, we already mentioned that D n converges to zero with probability one uniformly on each interval [, x ] with F (x ) <. Also for each fixed x, we have ED n (x) = and VarD n (x) n x C (y). Moreover, by the construction of a n and the definition of and C, we have VarD n (a n ) = O(). Conclude that D n (a n ) = O P (). Apply standard tightness arguments for the (non-standardized) process D n, see Billingsley (968), p. 8, to get that D n is uniformly bounded on [, a n ]. This, however, implies (3.3) and completes the proof of Lemma 3.. Our next Lemma implies that S n is negligible. Lemma 3.3. Under the assumptions of Theorem., we have S n = o P (n / ).

13 Proof. We first bound B n (x). Since, for x, we have we obtain n x x x n(dy) C n(y) n x ( x) ln( x) x, n(dy) [nc n (y) + ] B n(x). (3.5) Recall that nc n on the support of n so that the above integrals are all well defined. Conclude from (3.5) that S n n ϕ(x) C n (x) e n(x) n(dy) C n(y) n(dx). Now, as in the proof of Lemma 3., for a given ε >, we may choose some small c = c ε and a sequence a n such that (a n ) = c n and P(X n:n a n ) ε. Hence on this event, n has all its mass on [, a n ], and integration with respect to n may thus be restricted to x a n. Furthermore, by Lemmas 3. and 3., the processes C/C n and D n + D n remain stochastically bounded when restricted to the support of n, as n. Together with B n (x) it therefore suffices to show that a n n / n(dy) C (y) n(dx) = o P (). The expectation of the left-hand side is less than or equal to a n n / For any fixed x with F (x ) <, the integral x C (y) (dx). C (y) (dx) is finite, since F (y) is bounded away from zero there and by assumption (dy) df G (y) = α G <. The same holds true for the integral a n x x 3 C (y) (dx).

14 It remains to study a n x x x C (y) (dx). This integral is bounded from above, however, by G (x ) Now apply Cauchy-Schwarz to get a n x ϕ(x) F (dx) F (x) x a n x ϕ(x) F (dx) F (x). ϕ (x)f (dx) / a n F (dx) [ F (x)] The second integral is less than or equal to [ F (a n )]. Since c n = (a n ) = α a n G(z)F (dz) α ( F (a n )), the second square root is O(n / ). On the other hand,the first integral can be made arbitrarily small by choosing x large enough. This concludes the proof of Lemma 3.3. Next we study S n. For this, the following Lemma will be crucial. Lemma 3.4. Under the assumptions of Theorem., we have ϕ γ[c Cn ] I n df C n = o P (n / ). C n /. Proof. As in the proof of Lemma 3.3 we have X i a n for all i =,..., n with large probability. Similarly, consider a sequence b n > such that (b n ) = c, c sufficiently small, n such that X i b n for i =,,..., n with large probability. In other words, up to an event of small probability, we may restrict integration in I n to the interval [b n, a n ]. Also, in view of Lemma 3., the C n in the denominator may be replaced by C (times a constant), with large probability. Hence on 4

15 a set with large probability we have n I n = n ϕ(x i ) γ(x i ) {bn X i a n } C (X i= i )C n (X i ) { n ϕ(x i ) γ(x i ) {...} n C 3 (X i ) n i= + n 3 n i= j i [ n ( {Yj X i X j } C(X i ) ) + n ( C(X i)) j i ( {...} C(X i ) )} ϕ(x i ) γ(x i ) {...} C (X i )C n (X i ). (3.6) The first sum has an expectation not exceeding n a n b n (dx). C (x) Fix a small positive x and, as in the previous proof, a large x. Assume w.l.o.g. that b n x x a n. The integral ] x x ϕ γ C d is finite. So, the middle part contributes an error O( ), which is more than n a n we want. The upper part,..., is dealt with as in the proof of Lemma x 3.3, yielding a bound o(n / ) as x gets large. The same holds true for the lower part. Finally, the second sum in (3.6) may be studied along the same lines as for the first, upon starting with the inequality nc n (X i ). The proof is complete. Corollary 3.. Under the assumptions of Theorem., we have S n = ϕγ C d n + ϕγ[c C n ] C d n + o P (n / ). We shall come back to S n later but first proceed to S n3. Recalling D n, we have S n3 = ϕ(x)e n(x) C n (x) Fn(dy) C(y) n(dx). To get an expansion for S n3, the next lemma will be crucial. 5

16 Lemma 3.5. Under the assumptions of Theorem., we have [ ] ϕ(x)γ(x) II n e n(x)+ C(y) F n(dy) C n (x) C(y) n(dx) = o P (n / ). Proof. By Cauchy-Schwarz, on a set with large probability, nii n n / a n a n n / C n (x) C n (x) [ n(dy) C(y) e F n(x)+ (dy) C(y) ] n(dx) n(dy). (3.7) By Lemma 3., we may again replace C n by C. The expectation of the resulting first integral is then less than or equal to a n n / C (y) (dx), which was already shown to be o() in the proof of Lemma 3.3. It therefore remains to show that also the second factor in (3.7) is o P (). Putting we may write z n (x) = n (x) + C(y), [ e z n (x) ] = z n (x)e z n(x), where z n (x) is between zero and z n (x). Recalling that n (x) is between the first integral in B n (x) and and B n (x), we may infer that C(y) z n (x) is uniformly bounded from above in probability, on the support of Fn, as n. Hence it suffices to bound the term a n n / which in turn is less than or equal to a n n / n / zn(x) n(dx), [B n (x) + D n (x) + D n (x)] n(dx) a n [B n(x) + (D n (x) + D n (x)) ]Fn(dx). 6

17 By (3.5), a n Bn(x) n(dx) n a n n(dy) C n(y) n(dx). To bound the right-hand side, replace C n with C first. Then the squared term may be viewed as a V-statistic, of which the leading term is a U- statistic. Its expectation is n n Thus we have to show that a n n ϕ(x) C (y) C (y). F (dx) = o(n / ). This follows as in the proof of Lemma 3.3. Only the powers of F (x) are different. Finally, the error terms involving D n and D n may be dealt with similarly. The proof is complete. Lemma 3.6. Under the assumptions of Theorem., we have ϕγ[cn C](x) CC n (x) Fn(dy) C(y) n(dx) = o P (n / ). Proof. As before. Lemmas 3.5 and 3.6 immediately imply the following Corollary, which brings us to the desired representation of S n3. Corollary 3.. Under the assumptions of Theorem., we have the representation S n3 = ϕ(x)γ(x) Fn(dy) C(y) n(dx) + o P (n / ). We now proceed to S n4. As a first step to get the desired representation, we shall need the following lemma. Lemma 3.7. Under the assumptions of Theorem., S n4 admits the expansion S n4 = ϕ(x)e n(x) C n (x) + o P (n / ). C n (y) + C(y) n C (y) n(dy)fn(dx) (3.8) 7

18 Proof. Recalling the definition of D n, the difference between S n4 and the leading term in (3.8) becomes ϕ(x)e n(x) C n (x) [C n (y) + n C(y)] C (y)[c n (y) + n ] n(dy) n(dx). Taking Lemma 3. into account, the last expression is bounded in absolute values from above, with large probability, by ϕ(x) e n(x) [C n (y) + n C(y)] C 3 (y) n(dy)fn(dx). In the proof of Lemma 3.5 we have argued that z n (x) = n (x) + C(y) is uniformly bounded from above on the support of n, as n, with probability one. Consequently we may replace exp( n (x)) by γ(x). Also we may wish to extend the integral only up to a n. The expectation of the resulting term then does not exceed a n ϕ(x) C(y) + 4 n n F (dx). C 3 (y) These terms already appeared in the proof of Lemma 3.3 and were shown to be o(n / ) there. The proof therefore is complete. In the following we shall omit the summand in (3.8), since its contribution n is also negligible. The next Lemma will enable us to replace exp( n (x)) by γ(x). Lemma 3.8. Under the assumptions of Theorem., we have III n ϕ(x)γ(x) C n (x) = o P (n / ). [ e F n(x)+ (dy) C(y) ] C n C df C n n(dx) Proof. Cauchy-Schwarz leads to a bound for niiin similar to (3.7) for niin. The second factor is exactly the same as before. The first factor is dealt with in the following Lemma. 8

19 Lemma 3.9. Under the assumptions of Theorem., we get C n (y) C(y) C (y) n(dy) Fn(dx) = o P (n / ). Proof. As in previous proofs it is enough to extend the x-integral up to a n. It is then easily checked that the expectation of the resulting term is less than or equal to (when n > 3) a n where ϕ(x) G(y z)( F (y z)) + 4 (n 3)α n (dz)f (dx), C (y)c (z) y z = min(y, z) and y z = max(y, z). Neglecting 4n for a moment, the rest equals a n G(y)( F (z)) ϕ(x) (dz)f (dx) (n 3)α C (y)c (z) = = (n 3) n 3 n 3 a n a n a n y ϕ(x) ϕ(x) ϕ(x) ( F (y)) C(z) z y F (dz)f (dy)f (dx) C(z) F (dy)f (dz)f (dx) ( F (y)) F (dz)f (dx). C(z)( F (z)) The last integral already appeared in the proof of Lemma 3.3 and was shown to be o(n / ). The error term 4 yields similar bounds so that the n proof is complete. In the final step C n in the denominator of (3.8) may be replaced with C. Details are omitted. Hence we come up with the following representation of S n4. Corollary 3.3. Under the assumptions of Theorem., we have S n4 = ϕ(x)γ(x) + o P (n / ). C n (y) C(y) C (y) n(dy)fn(dx) 9

20 We are now ready to give the Proof of Theorem.. Corollaries 3., 3. and 3.3 yield representations of the relevant terms S n, S n3 and S n4 as V-statistics. As is known, see Serfling (98), a V-statistic equals a U-statistic, up to an error o P (n / ). If, in S n, S n3 and S n4, we replace Fn by, we come up with degenerate U-statistics which are all of the order o P (n / ). Collecting the leading terms and an application of Fubini s theorem yield the proof of Theorem. with ˆF n instead of F n. As to F n, apply the inequality i i i a j b j a j b j, a j, b j j= j= j= to the weights of F n and ˆF n. It follows that ϕd ˆF n ϕdf n n 3 n i= ϕ(x i:n ) C n (X i:n ) = n ϕ(x) C n (x) i j= C n(x j:n ) C n(y) n(dy) n(dx). Use Lemma 3. again to replace C n with C and apply the SLLN (for U- statistics) to get that the last term is O P (n ). This completes the proof of Theorem.. Proof of Lemma.. The proof proceeds by induction on i. For i = and j =, the LHS of (.6) equals /nc U n (U :n ), and so does the RHS. The assertion then follows from (.4). For i = and j =, the LHS of (.6) equals, by (.8) and (.5), nc U n (U :n ) nc U n (U :n ) nc U n (U :n ) = nc U n (U :n ) = nc n (X :n ). For i = and j >, the proof follows the same pattern, by repeated use of (.5). Hence the assertion of Lemma. holds true for i =. Assuming that (.6) holds true for all indices less than or equal to i, we obtain, by (.7), F n (X i:n ) = F U n (U idi :n). This may be used as a starting point to show (.6) for i +. Again make repeated use of (.4) and (.5). Acknowledgement. The research of Jane-Ling Wang was supported by an NSF grant NSF

21 References Billingsley, P. (968) Convergence of probability meassures. Wiley, New York. Chao, M.-T. and Lo, S.-H. (988) Some representations of the nonparametric maximum likelihood estimators with truncated data. Ann. Statist.,, He, S. and Yang, G. L. (998a) The strong law under random truncation. Ann. Statist., 6, 99-. He, S. and Yang, G. L. (998b) Estimation of the truncation probability in the random truncation model. Ann. Statist., 6, -7. Keiding, N. and Gill, R. D. (99) Random truncation models and Markov processes. Ann. Statist., 8, Lynden-Bell, D. (97) A method of allowing for known observational selection in small samples applied to 3CR quasars. Monthly Notices Roy. Astronom. Soc., 55, Sellero, C. S., Manteiga, W. G. and Van Keilegom, I. (5) Uniform representation of product-limit integrals with applications. Scand. J. Statist., 3, Serfling, R. J. (98) Approximation theorems of mathematical statistics. Wiley, New York. Stute, W. (993) Almost sure representations of the product-limit estimator for truncated data. Ann. Statist.,, Stute, W. (994) U-Statistic processes: a martingale approach. Ann. Probab.,, Stute, W. (995) The central limit theorem under random censorship. Ann. Statist., 3, Stute, W. (4) Kaplan-Meier integrals. Handbook of Statistics, 3, Elsevier, Amsterdam. Stute, W. and Wang, J. L. (993) The strong law under random censorship. Ann. Statist.,, Wang, M. C. (989) A semiparametric model for randomly truncated data. J. Amer. Statist. Assoc., 84, Woodroofe, M. (985) Estimating a distribution function with truncated data. Ann. Statist., 3,

22 Zhou, M. (99) Some properties of the Kaplan-Meier estimator for independent nonidentically distributed random variables. Ann. Statist., 9,

Estimation of the Bivariate and Marginal Distributions with Censored Data

Estimation of the Bivariate and Marginal Distributions with Censored Data Estimation of the Bivariate and Marginal Distributions with Censored Data Michael Akritas and Ingrid Van Keilegom Penn State University and Eindhoven University of Technology May 22, 2 Abstract Two new

More information

GOODNESS-OF-FIT TEST FOR RANDOMLY CENSORED DATA BASED ON MAXIMUM CORRELATION. Ewa Strzalkowska-Kominiak and Aurea Grané (1)

GOODNESS-OF-FIT TEST FOR RANDOMLY CENSORED DATA BASED ON MAXIMUM CORRELATION. Ewa Strzalkowska-Kominiak and Aurea Grané (1) Working Paper 4-2 Statistics and Econometrics Series (4) July 24 Departamento de Estadística Universidad Carlos III de Madrid Calle Madrid, 26 2893 Getafe (Spain) Fax (34) 9 624-98-49 GOODNESS-OF-FIT TEST

More information

Lecture 21: Convergence of transformations and generating a random variable

Lecture 21: Convergence of transformations and generating a random variable Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous

More information

Empirical likelihood ratio with arbitrarily censored/truncated data by EM algorithm

Empirical likelihood ratio with arbitrarily censored/truncated data by EM algorithm Empirical likelihood ratio with arbitrarily censored/truncated data by EM algorithm Mai Zhou 1 University of Kentucky, Lexington, KY 40506 USA Summary. Empirical likelihood ratio method (Thomas and Grunkmier

More information

The main results about probability measures are the following two facts:

The main results about probability measures are the following two facts: Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a

More information

STAT 6385 Survey of Nonparametric Statistics. Order Statistics, EDF and Censoring

STAT 6385 Survey of Nonparametric Statistics. Order Statistics, EDF and Censoring STAT 6385 Survey of Nonparametric Statistics Order Statistics, EDF and Censoring Quantile Function A quantile (or a percentile) of a distribution is that value of X such that a specific percentage of the

More information

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with

More information

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June

More information

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky Empirical likelihood with right censored data were studied by Thomas and Grunkmier (1975), Li (1995),

More information

11 Survival Analysis and Empirical Likelihood

11 Survival Analysis and Empirical Likelihood 11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with

More information

1 Glivenko-Cantelli type theorems

1 Glivenko-Cantelli type theorems STA79 Lecture Spring Semester Glivenko-Cantelli type theorems Given i.i.d. observations X,..., X n with unknown distribution function F (t, consider the empirical (sample CDF ˆF n (t = I [Xi t]. n Then

More information

Smooth nonparametric estimation of a quantile function under right censoring using beta kernels

Smooth nonparametric estimation of a quantile function under right censoring using beta kernels Smooth nonparametric estimation of a quantile function under right censoring using beta kernels Chanseok Park 1 Department of Mathematical Sciences, Clemson University, Clemson, SC 29634 Short Title: Smooth

More information

STAT 7032 Probability Spring Wlodek Bryc

STAT 7032 Probability Spring Wlodek Bryc STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,

More information

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then

More information

Asymptotic statistics using the Functional Delta Method

Asymptotic statistics using the Functional Delta Method Quantiles, Order Statistics and L-Statsitics TU Kaiserslautern 15. Februar 2015 Motivation Functional The delta method introduced in chapter 3 is an useful technique to turn the weak convergence of random

More information

Product-limit estimators of the survival function with left or right censored data

Product-limit estimators of the survival function with left or right censored data Product-limit estimators of the survival function with left or right censored data 1 CREST-ENSAI Campus de Ker-Lann Rue Blaise Pascal - BP 37203 35172 Bruz cedex, France (e-mail: patilea@ensai.fr) 2 Institut

More information

Asymptotics of minimax stochastic programs

Asymptotics of minimax stochastic programs Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming

More information

The International Journal of Biostatistics

The International Journal of Biostatistics The International Journal of Biostatistics Volume 1, Issue 1 2005 Article 3 Score Statistics for Current Status Data: Comparisons with Likelihood Ratio and Wald Statistics Moulinath Banerjee Jon A. Wellner

More information

EMPIRICAL ENVELOPE MLE AND LR TESTS. Mai Zhou University of Kentucky

EMPIRICAL ENVELOPE MLE AND LR TESTS. Mai Zhou University of Kentucky EMPIRICAL ENVELOPE MLE AND LR TESTS Mai Zhou University of Kentucky Summary We study in this paper some nonparametric inference problems where the nonparametric maximum likelihood estimator (NPMLE) are

More information

Regression model with left truncated data

Regression model with left truncated data Regression model with left truncated data School of Mathematics Sang Hui-yan 00001107 October 13, 2003 Abstract For linear regression model with truncated data, Bhattacharya, Chernoff, and Yang(1983)(BCY)

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

Chapter 4: Asymptotic Properties of the MLE

Chapter 4: Asymptotic Properties of the MLE Chapter 4: Asymptotic Properties of the MLE Daniel O. Scharfstein 09/19/13 1 / 1 Maximum Likelihood Maximum likelihood is the most powerful tool for estimation. In this part of the course, we will consider

More information

4 Sums of Independent Random Variables

4 Sums of Independent Random Variables 4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables

More information

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University

More information

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests Biometrika (2014),,, pp. 1 13 C 2014 Biometrika Trust Printed in Great Britain Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests BY M. ZHOU Department of Statistics, University

More information

Laplace s Equation. Chapter Mean Value Formulas

Laplace s Equation. Chapter Mean Value Formulas Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic

More information

Statistical inference on Lévy processes

Statistical inference on Lévy processes Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline

More information

STAT Sample Problem: General Asymptotic Results

STAT Sample Problem: General Asymptotic Results STAT331 1-Sample Problem: General Asymptotic Results In this unit we will consider the 1-sample problem and prove the consistency and asymptotic normality of the Nelson-Aalen estimator of the cumulative

More information

Asymptotic Properties of Kaplan-Meier Estimator. for Censored Dependent Data. Zongwu Cai. Department of Mathematics

Asymptotic Properties of Kaplan-Meier Estimator. for Censored Dependent Data. Zongwu Cai. Department of Mathematics To appear in Statist. Probab. Letters, 997 Asymptotic Properties of Kaplan-Meier Estimator for Censored Dependent Data by Zongwu Cai Department of Mathematics Southwest Missouri State University Springeld,

More information

arxiv: v2 [math.st] 20 Nov 2016

arxiv: v2 [math.st] 20 Nov 2016 A Lynden-Bell integral estimator for the tail inde of right-truncated data with a random threshold Nawel Haouas Abdelhakim Necir Djamel Meraghni Brahim Brahimi Laboratory of Applied Mathematics Mohamed

More information

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1 Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be

More information

Extreme Value Analysis and Spatial Extremes

Extreme Value Analysis and Spatial Extremes Extreme Value Analysis and Department of Statistics Purdue University 11/07/2013 Outline Motivation 1 Motivation 2 Extreme Value Theorem and 3 Bayesian Hierarchical Models Copula Models Max-stable Models

More information

Marshall-Olkin Bivariate Exponential Distribution: Generalisations and Applications

Marshall-Olkin Bivariate Exponential Distribution: Generalisations and Applications CHAPTER 6 Marshall-Olkin Bivariate Exponential Distribution: Generalisations and Applications 6.1 Introduction Exponential distributions have been introduced as a simple model for statistical analysis

More information

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get 18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

Bahadur representations for bootstrap quantiles 1

Bahadur representations for bootstrap quantiles 1 Bahadur representations for bootstrap quantiles 1 Yijun Zuo Department of Statistics and Probability, Michigan State University East Lansing, MI 48824, USA zuo@msu.edu 1 Research partially supported by

More information

Inconsistency of the MLE for the joint distribution. of interval censored survival times. and continuous marks

Inconsistency of the MLE for the joint distribution. of interval censored survival times. and continuous marks Inconsistency of the MLE for the joint distribution of interval censored survival times and continuous marks By M.H. Maathuis and J.A. Wellner Department of Statistics, University of Washington, Seattle,

More information

1. Introduction In many biomedical studies, the random survival time of interest is never observed and is only known to lie before an inspection time

1. Introduction In many biomedical studies, the random survival time of interest is never observed and is only known to lie before an inspection time ASYMPTOTIC PROPERTIES OF THE GMLE WITH CASE 2 INTERVAL-CENSORED DATA By Qiqing Yu a;1 Anton Schick a, Linxiong Li b;2 and George Y. C. Wong c;3 a Dept. of Mathematical Sciences, Binghamton University,

More information

Functional Analysis Exercise Class

Functional Analysis Exercise Class Functional Analysis Exercise Class Week 9 November 13 November Deadline to hand in the homeworks: your exercise class on week 16 November 20 November Exercises (1) Show that if T B(X, Y ) and S B(Y, Z)

More information

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

FULL LIKELIHOOD INFERENCES IN THE COX MODEL October 20, 2007 FULL LIKELIHOOD INFERENCES IN THE COX MODEL BY JIAN-JIAN REN 1 AND MAI ZHOU 2 University of Central Florida and University of Kentucky Abstract We use the empirical likelihood approach

More information

Efficiency of Profile/Partial Likelihood in the Cox Model

Efficiency of Profile/Partial Likelihood in the Cox Model Efficiency of Profile/Partial Likelihood in the Cox Model Yuichi Hirose School of Mathematics, Statistics and Operations Research, Victoria University of Wellington, New Zealand Summary. This paper shows

More information

Real Analysis Math 131AH Rudin, Chapter #1. Dominique Abdi

Real Analysis Math 131AH Rudin, Chapter #1. Dominique Abdi Real Analysis Math 3AH Rudin, Chapter # Dominique Abdi.. If r is rational (r 0) and x is irrational, prove that r + x and rx are irrational. Solution. Assume the contrary, that r+x and rx are rational.

More information

3 Integration and Expectation

3 Integration and Expectation 3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ

More information

Practical conditions on Markov chains for weak convergence of tail empirical processes

Practical conditions on Markov chains for weak convergence of tail empirical processes Practical conditions on Markov chains for weak convergence of tail empirical processes Olivier Wintenberger University of Copenhagen and Paris VI Joint work with Rafa l Kulik and Philippe Soulier Toronto,

More information

Density estimators for the convolution of discrete and continuous random variables

Density estimators for the convolution of discrete and continuous random variables Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract

More information

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9 MAT 570 REAL ANALYSIS LECTURE NOTES PROFESSOR: JOHN QUIGG SEMESTER: FALL 204 Contents. Sets 2 2. Functions 5 3. Countability 7 4. Axiom of choice 8 5. Equivalence relations 9 6. Real numbers 9 7. Extended

More information

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the

More information

TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOZIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississip

TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOZIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississip TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississippi University, MS38677 K-sample location test, Koziol-Green

More information

Quantile-quantile plots and the method of peaksover-threshold

Quantile-quantile plots and the method of peaksover-threshold Problems in SF2980 2009-11-09 12 6 4 2 0 2 4 6 0.15 0.10 0.05 0.00 0.05 0.10 0.15 Figure 2: qqplot of log-returns (x-axis) against quantiles of a standard t-distribution with 4 degrees of freedom (y-axis).

More information

Limiting Distributions

Limiting Distributions Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Some basic elements of Probability Theory

Some basic elements of Probability Theory Chapter I Some basic elements of Probability Theory 1 Terminology (and elementary observations Probability theory and the material covered in a basic Real Variables course have much in common. However

More information

Krzysztof Burdzy University of Washington. = X(Y (t)), t 0}

Krzysztof Burdzy University of Washington. = X(Y (t)), t 0} VARIATION OF ITERATED BROWNIAN MOTION Krzysztof Burdzy University of Washington 1. Introduction and main results. Suppose that X 1, X 2 and Y are independent standard Brownian motions starting from 0 and

More information

A Conditional Approach to Modeling Multivariate Extremes

A Conditional Approach to Modeling Multivariate Extremes A Approach to ing Multivariate Extremes By Heffernan & Tawn Department of Statistics Purdue University s April 30, 2014 Outline s s Multivariate Extremes s A central aim of multivariate extremes is trying

More information

Lecture 4 Lebesgue spaces and inequalities

Lecture 4 Lebesgue spaces and inequalities Lecture 4: Lebesgue spaces and inequalities 1 of 10 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 4 Lebesgue spaces and inequalities Lebesgue spaces We have seen how

More information

Approximate Self Consistency for Middle-Censored Data

Approximate Self Consistency for Middle-Censored Data Approximate Self Consistency for Middle-Censored Data by S. Rao Jammalamadaka Department of Statistics & Applied Probability, University of California, Santa Barbara, CA 93106, USA. and Srikanth K. Iyer

More information

The best expert versus the smartest algorithm

The best expert versus the smartest algorithm Theoretical Computer Science 34 004 361 380 www.elsevier.com/locate/tcs The best expert versus the smartest algorithm Peter Chen a, Guoli Ding b; a Department of Computer Science, Louisiana State University,

More information

Lecture 6 Basic Probability

Lecture 6 Basic Probability Lecture 6: Basic Probability 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 6 Basic Probability Probability spaces A mathematical setup behind a probabilistic

More information

Goodness-of-fit tests for the cure rate in a mixture cure model

Goodness-of-fit tests for the cure rate in a mixture cure model Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,

More information

Characterization of Upper Comonotonicity via Tail Convex Order

Characterization of Upper Comonotonicity via Tail Convex Order Characterization of Upper Comonotonicity via Tail Convex Order Hee Seok Nam a,, Qihe Tang a, Fan Yang b a Department of Statistics and Actuarial Science, University of Iowa, 241 Schaeffer Hall, Iowa City,

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research

More information

Exercises in Extreme value theory

Exercises in Extreme value theory Exercises in Extreme value theory 2016 spring semester 1. Show that L(t) = logt is a slowly varying function but t ǫ is not if ǫ 0. 2. If the random variable X has distribution F with finite variance,

More information

AFT Models and Empirical Likelihood

AFT Models and Empirical Likelihood AFT Models and Empirical Likelihood Mai Zhou Department of Statistics, University of Kentucky Collaborators: Gang Li (UCLA); A. Bathke; M. Kim (Kentucky) Accelerated Failure Time (AFT) models: Y = log(t

More information

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

General Glivenko-Cantelli theorems

General Glivenko-Cantelli theorems The ISI s Journal for the Rapid Dissemination of Statistics Research (wileyonlinelibrary.com) DOI: 10.100X/sta.0000......................................................................................................

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Product measure and Fubini s theorem

Product measure and Fubini s theorem Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω

More information

CURRENT STATUS LINEAR REGRESSION. By Piet Groeneboom and Kim Hendrickx Delft University of Technology and Hasselt University

CURRENT STATUS LINEAR REGRESSION. By Piet Groeneboom and Kim Hendrickx Delft University of Technology and Hasselt University CURRENT STATUS LINEAR REGRESSION By Piet Groeneboom and Kim Hendrickx Delft University of Technology and Hasselt University We construct n-consistent and asymptotically normal estimates for the finite

More information

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974 LIMITS FOR QUEUES AS THE WAITING ROOM GROWS by Daniel P. Heyman Ward Whitt Bell Communications Research AT&T Bell Laboratories Red Bank, NJ 07701 Murray Hill, NJ 07974 May 11, 1988 ABSTRACT We study the

More information

5 Introduction to the Theory of Order Statistics and Rank Statistics

5 Introduction to the Theory of Order Statistics and Rank Statistics 5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order

More information

THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION

THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION JAINUL VAGHASIA Contents. Introduction. Notations 3. Background in Probability Theory 3.. Expectation and Variance 3.. Convergence

More information

2 Functions of random variables

2 Functions of random variables 2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as

More information

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 = Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values

More information

Extreme Value Theory and Applications

Extreme Value Theory and Applications Extreme Value Theory and Deauville - 04/10/2013 Extreme Value Theory and Introduction Asymptotic behavior of the Sum Extreme (from Latin exter, exterus, being on the outside) : Exceeding the ordinary,

More information

Self-normalized Cramér-Type Large Deviations for Independent Random Variables

Self-normalized Cramér-Type Large Deviations for Independent Random Variables Self-normalized Cramér-Type Large Deviations for Independent Random Variables Qi-Man Shao National University of Singapore and University of Oregon qmshao@darkwing.uoregon.edu 1. Introduction Let X, X

More information

WLLN for arrays of nonnegative random variables

WLLN for arrays of nonnegative random variables WLLN for arrays of nonnegative random variables Stefan Ankirchner Thomas Kruse Mikhail Urusov November 8, 26 We provide a weak law of large numbers for arrays of nonnegative and pairwise negatively associated

More information

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON GEORGIAN MATHEMATICAL JOURNAL: Vol. 3, No. 2, 1996, 153-176 THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON M. SHASHIASHVILI Abstract. The Skorokhod oblique reflection problem is studied

More information

Refining the Central Limit Theorem Approximation via Extreme Value Theory

Refining the Central Limit Theorem Approximation via Extreme Value Theory Refining the Central Limit Theorem Approximation via Extreme Value Theory Ulrich K. Müller Economics Department Princeton University February 2018 Abstract We suggest approximating the distribution of

More information

Chapter 2: Resampling Maarten Jansen

Chapter 2: Resampling Maarten Jansen Chapter 2: Resampling Maarten Jansen Randomization tests Randomized experiment random assignment of sample subjects to groups Example: medical experiment with control group n 1 subjects for true medicine,

More information

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema Chapter 7 Extremal Problems No matter in theoretical context or in applications many problems can be formulated as problems of finding the maximum or minimum of a function. Whenever this is the case, advanced

More information

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

A note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis

A note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis A note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis Jon A. Wellner and Vladimir Koltchinskii Abstract. Proofs are given of the limiting null distributions of the

More information

Random Bernstein-Markov factors

Random Bernstein-Markov factors Random Bernstein-Markov factors Igor Pritsker and Koushik Ramachandran October 20, 208 Abstract For a polynomial P n of degree n, Bernstein s inequality states that P n n P n for all L p norms on the unit

More information

Math212a1413 The Lebesgue integral.

Math212a1413 The Lebesgue integral. Math212a1413 The Lebesgue integral. October 28, 2014 Simple functions. In what follows, (X, F, m) is a space with a σ-field of sets, and m a measure on F. The purpose of today s lecture is to develop the

More information

Principle of Mathematical Induction

Principle of Mathematical Induction Advanced Calculus I. Math 451, Fall 2016, Prof. Vershynin Principle of Mathematical Induction 1. Prove that 1 + 2 + + n = 1 n(n + 1) for all n N. 2 2. Prove that 1 2 + 2 2 + + n 2 = 1 n(n + 1)(2n + 1)

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS

A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS Statistica Sinica 20 2010, 365-378 A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS Liang Peng Georgia Institute of Technology Abstract: Estimating tail dependence functions is important for applications

More information

Estimation of Dynamic Regression Models

Estimation of Dynamic Regression Models University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.

More information

ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE. By Michael Nussbaum Weierstrass Institute, Berlin

ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE. By Michael Nussbaum Weierstrass Institute, Berlin The Annals of Statistics 1996, Vol. 4, No. 6, 399 430 ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE By Michael Nussbaum Weierstrass Institute, Berlin Signal recovery in Gaussian

More information

Research Article Exponential Inequalities for Positively Associated Random Variables and Applications

Research Article Exponential Inequalities for Positively Associated Random Variables and Applications Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 008, Article ID 38536, 11 pages doi:10.1155/008/38536 Research Article Exponential Inequalities for Positively Associated

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

MATH 409 Advanced Calculus I Lecture 12: Uniform continuity. Exponential functions.

MATH 409 Advanced Calculus I Lecture 12: Uniform continuity. Exponential functions. MATH 409 Advanced Calculus I Lecture 12: Uniform continuity. Exponential functions. Uniform continuity Definition. A function f : E R defined on a set E R is called uniformly continuous on E if for every

More information

Problem 1: Compactness (12 points, 2 points each)

Problem 1: Compactness (12 points, 2 points each) Final exam Selected Solutions APPM 5440 Fall 2014 Applied Analysis Date: Tuesday, Dec. 15 2014, 10:30 AM to 1 PM You may assume all vector spaces are over the real field unless otherwise specified. Your

More information

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem

More information

FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH

FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH Jian-Jian Ren 1 and Mai Zhou 2 University of Central Florida and University of Kentucky Abstract: For the regression parameter

More information

Section 8.2. Asymptotic normality

Section 8.2. Asymptotic normality 30 Section 8.2. Asymptotic normality We assume that X n =(X 1,...,X n ), where the X i s are i.i.d. with common density p(x; θ 0 ) P= {p(x; θ) :θ Θ}. We assume that θ 0 is identified in the sense that

More information