The Central Limit Theorem Under Random Truncation
|
|
- Madeline Gilmore
- 5 years ago
- Views:
Transcription
1 The Central Limit Theorem Under Random Truncation WINFRIED STUTE and JANE-LING WANG Mathematical Institute, University of Giessen, Arndtstr., D-3539 Giessen, Germany. Department of Statistics, University of California, Davis, CA 9566, USA Under left truncation data (X i, Y i ) are observed only when Y i X i. Usually the distribution function F of the X i is the target of interest. In this paper we study linear functionals ϕdf n of the nonparametric MLE of F, the Lynden- Bell estimator F n. A useful representation of ϕdf n is derived which yields asymptotic normality under optimal moment conditions on the score function ϕ. No continuity assumption on F is required. As a by-product we obtain the distributional convergence of the Lynden-Bell empirical process on the whole real line. Running Title: The CLT under random truncation Key words: CLT; Lynden-Bell integral; Truncated data. Introduction and main results In this paper we provide some further methodology for statistical analysis of data which are truncated from the left. To be more specific, let (X i, Y i ), i N, denote a sample of independent bivariate data such that, for each i, X i is also independent of Y i. Denote with F and G, respectively, the unknown distribution functions of X and Y. Typically, F is the target of interest. Now, under left truncation, X i is observed only when Y i X i. As a consequence, the empirical distribution of the X s is unavailable and cannot serve as a basic process to compute other statistics. The nonparametric maximum likelihood estimator of F for left-truncated data was first derived by Lynden-Bell (97). Its first mathematical investigation may be attributed to Woodroofe (985), who also reviewed some examples of truncated data from astronomy and economics. See also Wang (989) for applications in the analysis of AIDS data. Now, denoting with n the number of data which are actually observed, we have, by the Strong Law of Large Numbers (SLLN), n N α P(Y X) as N with probability one.
2 Without further mentioning we shall assume that α >, because otherwise nothing will be observed. Of course, α will be unknown. Conditionally on n, the observed data are still independent, but the joint distribution of X i and Y i becomes H (x, y) = P(X x, Y y Y X) = α G(y z)f (dz). (,x] The marginal distribution of the observed X s thus equals (x) α G(z)F (dz). (.) (,x] It may be consistently estimated by the empirical distribution function of the known X s: Fn(x) = n {Xi x}, x R. n i= The problem, however, is one of reconstructing F and not from the available data (X i, Y i ), i n. A crucial quantity in this context is the function where C(z) = P(Y z X Y X) = α G(z)[ F (z )], (.) F (z ) = lim x z F (x) is the left continuous version of F and F {z} = F (z) F (z ) is the F -mass at z. The function C may be consistently estimated by C n (z) = n n {Yi z X i }. It is very helpful to express the cumulative hazard function of F, F (dz) Λ(x) = F (z ), (,x] i= in terms of estimable quantities. For this, let a G = inf{x : G(x) > } be the largest lower bound for the support of G. Similarly for F. From (.) and (.) we obtain (a G,x] F (dz) F (z ) = (a G,x] (dz) C(z). (.3)
3 Provided that a G a F and F {a F } =, the left-hand side equals Λ(x). Otherwise this is no longer true, and as Woodroofe (985) pointed out, F cannot be fully recovered from the available data. The situation is similar for right-censored data. See Stute and Wang (993) for a detailed discussion for (upper) boundary effects there. Throughout the paper we shall therefore assume that a G a F and F {a F } =. (.4) If a G < a F the second assumption is superfluous, while it is automatically satisfied when F is continuous. For the moment, however, no other assumptions such as continuity of F or G will be needed. In differential terms, equation (.3) leads to F (dx) = ( F ()) (dx). (.5) The Lynden-Bell (97) estimator F n of F is obtained as the solution of the so-called self-consistency equation, i.e., as the solution of (.5) after having replaced and C with n and C n, respectively: F n (dx) = ( F n ()) n(dx) C n (x). (.6) Solving for F n yields the product-integral representation of F n : [ F n (x) = F ] n{x i }. (.7) C n (X i ) distinctx i x If there are no ties among the X s, (.7) simplifies to become F n (x) = [ ] ncn (X i ). (.8) nc n (X i ) X i x Note that nc n (X i ) so that each ratio is well defined. Since in this paper our objective will be to study general linear statistics based on F n, namely Lynden-Bell integrals ϕdf n, we also introduce the Lynden-Bell weights W in attached to each datum in the X-sample. For this, denote with X :n <... < X m:n the m distinct order statistics, with m possibly strictly less than n. From (.7) we obtain W in F n {X i:n } = [ F n (X i :n )] n{x i:n } C n (X i:n ) (.9) 3
4 and ϕdf n = m W in ϕ(x i:n ). (.) i= For ϕ = (,x], we are back at F n (x). As to other ϕ s we refer to Stute and Wang (993) or Stute (4), who considered possible applications of empirical integrals in the context of right censored data. More general statistical functionals often admit expansions, in which the leading term is of the form ϕdf n, with ϕ denoting the associated influence function. Since for a fully observable data set with Fn e being the classical empirical distribution function, ϕdfn e is just a sample mean to which, under a second moment assumption, the Central Limit Theorem (CLT) applies, distributional convergence of ϕdf n therefore constitutes an extension of the CLT to the left-truncation case. The corresponding SLLN has been studied in papers by He and Yang (998a,b). The CLT for censored data is due to Stute (995). In the present situation, the CLT is much more elusive than for randomly censored data. In the censored data case the right tails create technical difficulties, but not the left. For left truncation, however, both sides create problems. This is already seen now through the functions C and C n which decrease to zero on the left and on the right tail, and with C n appearing in denominators. The only trivial bound is nc n (X i ), which keeps everything from being not well defined. Moreover, C and C n are not monotone, again contrary to the censored data case, where the role of the C s is played by survival functions. This non-monotone feature of C n creates additional technical complications for truncated data. Worse than that, C n may become zero between the data points also in the central part. Keiding and Gill (99), who recognized the danger of these sets, have called these holes the empty inner risk sets. Not all authors seem to know about these problems since a detailed study of the C n process is sometimes missing. A consequence of these empty inner risk sets, as revealed by the representation (.7), is the loss of mass of F n after the first such hole. More specifically, suppose that, in terms of Keiding and Gill (99), a risky hole exists at X j. This means that X j is not covered by any other pair (X i, Y i ). Hence nc n (X j ) = nfn{x j }. From (.7) we get that all data points right to X j have mass zero under F n. In proofs this excludes one to incorporate exponential representations of the weights. To circumvent this difficulty we first study an asymptotically equivalent estimator ˆF n. This estimator is defined through: ϕd ˆF n = m i= ϕ(x i:n ) n{x i:n } C n (X i:n ) i j= [ nf ] n{x j:n }. (.) nc n (X j:n ) + The extra summand in the denominator allows for contributions of data which have holes on their left. As we shall see in a small simulation study 4
5 it may have a robustifying effect resulting in smaller MSE for small sample sizes. Asymptotically ϕdf n and ϕd ˆF n are equivalent at the n / - rate, which facilitates the asymptotic theory for ϕdf n. Another technical problem is caused by possible ties. In such a situation (.8) does not apply. Lemma. will show how the general case covered by (.7) and (.9) may be traced back to the case of a continuous. The main result of this paper, Theorem., provides a representation of ϕdfn as a sum of i.i.d. random variables under minimal assumptions on ϕ and the truncation mechanism. Usually linear i.i.d. representations of complicated estimators will include the Hájek projection of the statistic of interest. In the case of (.), however, this projection can be computed only up to remainders, and this is exactly what Theorem. does. More precisely, the proof of Theorem. proceeds in two steps. In the first step we have to identify all error terms which are negligible. The leading terms will turn out to be V-statistics. See Serfling (98). Finally, an application of the Hájek projection to the leading terms yields the desired i.i.d. representation. Needless to say asymptotic normality follows immediately. We shall also add some interesting comments on a so-called uniform representation. Proofs will be given in Section 3. Theorem. will hold under the following two assumptions: (A) (i) ϕ /GdF < (ii) df G < Condition (ii) already appeared in Woodroofe (985) in his study of the Lynden-Bell process, i.e., when he considered indicators ϕ = (,x ]. Nontechnically speaking, it is needed to guarantee enough information in the left tails to estimate F at the rate n /. Under slightly stronger assumptions, Stute (993) obtained an almost sure representation with sharp bounds on the remainder, again for indicators. See also Chao and Lo (988). Condition (i) guarantees, among other things, that the leading terms in the i.i.d. representation admit a finite second moment so that asymptotic normality holds. Since G it implies ϕ df <, which is the standard finite moment assumption when no truncation occurs. When ϕ has a finite second F -moment and is locally bounded in a neighborhood of a G, then (i) is implied by (ii). Note also that (i) and (ii) are always satisfied when a G < a F and ϕ df <. A CLT for truncated data is also contained in Sellero et al. (5). Apart from continuity assumptions they also need conditions which, in our notation, require finiteness of the integral ϕ (x)( F (x)) 5 F (dx), where ϕ ϕ. Since, however, this integral equals infinity for constant ϕ, their result is not applicable to bounded ϕ s, not to mention ϕ s which 5
6 increase to infinity as x. Rather, finiteness of the above integral is only obtained for ϕ s which converge to zero fast enough in the right tails. The focus of this paper is, however, on distributional convergence for which (A) will suffice. Theorem. is formulated for the case when F is continuous. This guarantees that among the observed X s there will be no ties, with probability one. In such a situation we obtain ϕd ˆF ϕ(x)γn (x) n = C n (x) n(dx), (.) with γ n (x) = exp n [ ln ] nc n (y) + Fn(dy). (.3) At the end of this section we shall show how general Lynden-Bell integrals may be traced back to the present case. Theorem.. Under Assumptions (A) and (.4), assume that F is continuous. Then we have ψ(y) ϕ[df n df ] = C(y) [F n(dy) ] Cn (y) C(y) ψ(y) + o C P (n / ), (y) where ψ(y) = ϕ(y)( F (y)) {y<x} ϕ(x)( F (x)) (dx) = {y<x} [ϕ(y) ϕ(x)]f (dx). For indicators ϕ = (,x ] the leading term already appears in Theorem in Stute (993). Remark.. If one checks the proof of Theorem. step by step, the following fact will be revealed: If rather than a single ϕ, one considers a collection {ϕ}, then the error terms are uniformly small in the sense that the remainder is o P (n / ) uniformly in ϕ, provided that ϕ ϕ for some ϕ satisfying ϕ df <. Actually, all remainders may be bounded from G above in absolute values by replacing ϕ by ϕ. Compared with Sellero et al. (5), no VC property for the ϕ s is needed for a representation as a V-statistic process. For the i.i.d. representation one has to guarantee that the errors in the Hájek projection are also uniformly small. These errors, however, form a class of degenerate U-Statistics. This kind of process was studied in Stute (994), and the achieved bounds are useful to handle the error terms in the second half of the proof. Details are omitted. 6
7 Corollary.. Under the assumptions of Theorem., we have n / ϕ[df n df ] N (, σ ) in distribution, with σ ψ(x) = Var C(X) X Y ψ(y) C (y). It is not difficult to see that σ < under (A). Finally, if in Remark. we take for {ϕ} the class of all indicators ϕ = (,x] and set ϕ, we obtain the following Corollary. Corollary.. Under df G have F n (x) F (x) = < and (.4), and for a continuous F, we ψx C [df n d ] Cn C ψ C x + o P (n / ) uniformly in x, where ψ x is the ψ belonging to ϕ = (,x]. Remark.. To make the point clear, Corollary. provides a representation which holds uniformly on the whole real line and not only on subintervals (, b] with b < b F = sup{x : F (x) < }, as is usually done in the literature. A major technical problem for proving Theorem. for a general F is caused by the fact that for discontinuous F ties may arise. As before, denote with X :n <... < X m:n the m ordered distinct data in the observed X-sample. To circumvent ties one may use the fact that each X i may be written in the form X i = (U i ), where U,..., U n is a sample of independent random variables with a uniform distribution on (,) and in the quantile function of. The construction of the U s is similar to the construction in Lemma.8 of Stute and Wang (993), with H there replaced by here. For the following recall that a quantile function is continuous from the left. Furthermore, with probability one, is also right-continuous at each U i. With this in mind, one can see that the corresponding truncating sample for the U i s are (Y i ), i n. If we denote with Cn U the C n -function corresponding to the pseudo observations (U i, (Y i )), i n, and if we let U i:n <... < U idi :n denote the ordered U s that satisfy (U ij:n ) = X i:n, with d i = nfn{x i:n }, then nc U n (U i:n ) = nc n (X i:n ) (.4) 7
8 and nc U n (U ij:n ) = nc U n (U i,j :n ), for j d i. (.5) Also note that the Lynden-Bell estimator Fn U for the pseudo observations satisfies (.8), i.e., Fn U (u) = [ ] nc U n (U i ). nc U U i u n (U i ) Introducing the function ϕ = ϕ, the analogue of (.) becomes ϕ df U n = = m d i i= j= ϕ( (U ij:n )) F U n (U ij:n ) nc U n (U ij:n ) m d i ϕ(x i:n ) i= j= Fn U (U ij:n ). ncn U (U ij:n ) In Lemma. we will show that for each i m and every j d i we have F U n (U ij:n ) nc U n (U ij:n ) = F n(x i :n ), (.6) nc n (X i:n ) where X :n =. It follows from (.9), (.) and (.6) that ϕ dfn U = ϕdf n with probability one. In conclusion the study of Lynden-Bell integrals may be traced back to the case when the variables of interest are uniform on (,) and therefore, with probability one, have no ties. At the same time our handling of ties does not require external randomization but maintains the product limit structure of the weights and hence of the estimators. Lemma.. For each i m and j d i, equation (.6) holds true. Simulations It is interesting to compare the small sample size behavior of F n and ˆF n. In a simulation study we considered ϕ(x) = x, i.e., the target was the mean lifetime of X. This ϕ is the canonical score function in the classical CLT. Recall that through truncation there is a sampling bias which would result in an upward bias if we would take the empirical distribution function of the X s and not F n or ˆF n. Introducing more or less complicated weights has the effect, among other things, that compared with the empirical distribution function the bias is reduced. 8
9 In the following we report on some simulation results which are part of a much more extensive study. Typically, for this ϕ, ϕd ˆF n outperforms ϕdfn in terms of the Mean Squared Error (MSE) when n 4 and truncation is heavy. For larger n(n 4) the difference is negligible. MSE was computed through Monte Carlo. In the following both X and Y were exponentially distributed with parameter : 3 Proofs Table. Comparison of MSE Sample Size MSE ( ϕdf n ) MSE( ϕd ˆF n ) n = n = n = n = n = n = n = n = 8.. n = 9.. n =..9 To prove Theorem., note that under continuity of F (.8) applies. Also we may assume w.l.o.g. that all data are nonnegative. Hence all appearing integrals will be over the positive real line hereafter. In our first lemma we provide a bound for the function C/C n. This will be needed in proofs to handle negligible terms. Though, by Glivenko-Cantelli, C n C uniformly, bounding the ratio is more delicate in view of possible holes and the non-monotonicity of C and C n. Lemma 3.. Assume F is continuous. For any λ such that αλ one has ( ) C(X i ) P sup X i :X i b C n (X i ) λ λe exp[ G(b)αλ]. Proof. The proof is similar to that of Lemma. in Stute (993), which provides an exponential bound for the sup extended over the left tails X i b rather than the right tails. Together with the above mentioned bound from Stute (993), Lemma 3. immediately implies sup i n C(X i ) C n (X i ) = O P() as n. (3.) 9
10 Assertion (3.) will be of some importance in forthcoming proofs, as it will allow us to replace the random C n appearing in denominators with the deterministic C. We are now ready to expand ϕd ˆF n. By (.) and (.3), ϕd ˆF n = ϕ(x) C n (x) exp n ln ϕ(x) C n (x) γ n(x) n(dx). [ ] nc n (y) + n(dy) F n(dx) Set γ(x) = F (x) = exp From Taylor s expansion we obtain x C(y). where γ n (x) = γ(x) + e n(x) [B n (x) + D n (x) + D n (x)], B n (x) = n D n (x) = D n (x) = ln [ ] nc n (y) + n(dy) + Fn(dy), C(y) [C n (y) + n C(y)] [C n (y) + n ]C(y) n(dy) Fn(dy) C n (y) +, n and n (x) is between the exponents of γ n (x) and γ(x). Particularly, we have n (x). Setting S n = S n = S n3 = ϕ(x) C n (x) e n(x) B n (x) n(dx) ϕ(x)γ(x) C n (x) n(dx) ϕ(x) C n (x) e n(x) D n (x) n(dx)
11 and we thus get S n4 = ϕ(x) C n (x) e n(x) D n (x) n(dx), ϕd ˆF n = S n + S n + S n3 + S n4. (3.) In the next lemma we study the functions D n and D n more closely. To motivate the following note that for each fixed x such that F (x ) < D n (x ) with probability one. Actually, by standard Glivenko-Cantelli arguments, D n (x) with probability one uniformly on x x. Similarly, when we consider the standardized processes x n / D n (x), it is easy to show their distributional convergence in the Skorokhod space D[, x ]. Things, however, change if we study D n on the whole support of. Since = C(y) we cannot expect uniform convergence on the whole support of. The situation is similar for the cumulative hazard function Λ, where uniform convergence of the Nelson-Aalen estimator Λ n may be obtained only on compact subsets of the support of F. When one evaluates these processes at x = X i it is known though, that the uniform deviation between Λ n and Λ does not got to zero but remains at least bounded. See Theorem. in Zhou (99). Similar things turn out to be true for D n and D n, as Lemma 3. will show. Our proofs are different though since compared with Zhou (99) we shall apply a truncation technique which in proofs guarantees that the sup of D n and D n are bounded on large but not too large sets. Lemma 3.. We have and sup D n (X i ) = O P () (3.3) i n sup D n (X i ) = O P (). (3.4) i n
12 Proof. Assume w.l.o.g. that has unbounded support on the right. Otherwise, replace by b F = sup{x : (x) < }. For a given ε >, one may find some small c = c ε and a sequence a n such that (a n ) = c n and P(X n:n a n ) ε. Actually, this follows from P(X n:n a n ) = [ c n ]n exp( c). It therefore suffices to bound D n and D n on (, a n ]. By Lemma 3., we may replace C n in the denominator by C. Hence with large probability and up to a constant factor, (3.4) is bounded from above by a n C n (y) C(y) C (y) n(dy) + n a n n(dy) C (y). Using (ii) of (A), it is easily seen that the expectation of the second term is bounded as n. The expectation of the first term is less than or equal to a n E C n (y) C(y) + n C (y) a n n C / (y) + n C (y) = O(). This proves (3.4). As for D n, we already mentioned that D n converges to zero with probability one uniformly on each interval [, x ] with F (x ) <. Also for each fixed x, we have ED n (x) = and VarD n (x) n x C (y). Moreover, by the construction of a n and the definition of and C, we have VarD n (a n ) = O(). Conclude that D n (a n ) = O P (). Apply standard tightness arguments for the (non-standardized) process D n, see Billingsley (968), p. 8, to get that D n is uniformly bounded on [, a n ]. This, however, implies (3.3) and completes the proof of Lemma 3.. Our next Lemma implies that S n is negligible. Lemma 3.3. Under the assumptions of Theorem., we have S n = o P (n / ).
13 Proof. We first bound B n (x). Since, for x, we have we obtain n x x x n(dy) C n(y) n x ( x) ln( x) x, n(dy) [nc n (y) + ] B n(x). (3.5) Recall that nc n on the support of n so that the above integrals are all well defined. Conclude from (3.5) that S n n ϕ(x) C n (x) e n(x) n(dy) C n(y) n(dx). Now, as in the proof of Lemma 3., for a given ε >, we may choose some small c = c ε and a sequence a n such that (a n ) = c n and P(X n:n a n ) ε. Hence on this event, n has all its mass on [, a n ], and integration with respect to n may thus be restricted to x a n. Furthermore, by Lemmas 3. and 3., the processes C/C n and D n + D n remain stochastically bounded when restricted to the support of n, as n. Together with B n (x) it therefore suffices to show that a n n / n(dy) C (y) n(dx) = o P (). The expectation of the left-hand side is less than or equal to a n n / For any fixed x with F (x ) <, the integral x C (y) (dx). C (y) (dx) is finite, since F (y) is bounded away from zero there and by assumption (dy) df G (y) = α G <. The same holds true for the integral a n x x 3 C (y) (dx).
14 It remains to study a n x x x C (y) (dx). This integral is bounded from above, however, by G (x ) Now apply Cauchy-Schwarz to get a n x ϕ(x) F (dx) F (x) x a n x ϕ(x) F (dx) F (x). ϕ (x)f (dx) / a n F (dx) [ F (x)] The second integral is less than or equal to [ F (a n )]. Since c n = (a n ) = α a n G(z)F (dz) α ( F (a n )), the second square root is O(n / ). On the other hand,the first integral can be made arbitrarily small by choosing x large enough. This concludes the proof of Lemma 3.3. Next we study S n. For this, the following Lemma will be crucial. Lemma 3.4. Under the assumptions of Theorem., we have ϕ γ[c Cn ] I n df C n = o P (n / ). C n /. Proof. As in the proof of Lemma 3.3 we have X i a n for all i =,..., n with large probability. Similarly, consider a sequence b n > such that (b n ) = c, c sufficiently small, n such that X i b n for i =,,..., n with large probability. In other words, up to an event of small probability, we may restrict integration in I n to the interval [b n, a n ]. Also, in view of Lemma 3., the C n in the denominator may be replaced by C (times a constant), with large probability. Hence on 4
15 a set with large probability we have n I n = n ϕ(x i ) γ(x i ) {bn X i a n } C (X i= i )C n (X i ) { n ϕ(x i ) γ(x i ) {...} n C 3 (X i ) n i= + n 3 n i= j i [ n ( {Yj X i X j } C(X i ) ) + n ( C(X i)) j i ( {...} C(X i ) )} ϕ(x i ) γ(x i ) {...} C (X i )C n (X i ). (3.6) The first sum has an expectation not exceeding n a n b n (dx). C (x) Fix a small positive x and, as in the previous proof, a large x. Assume w.l.o.g. that b n x x a n. The integral ] x x ϕ γ C d is finite. So, the middle part contributes an error O( ), which is more than n a n we want. The upper part,..., is dealt with as in the proof of Lemma x 3.3, yielding a bound o(n / ) as x gets large. The same holds true for the lower part. Finally, the second sum in (3.6) may be studied along the same lines as for the first, upon starting with the inequality nc n (X i ). The proof is complete. Corollary 3.. Under the assumptions of Theorem., we have S n = ϕγ C d n + ϕγ[c C n ] C d n + o P (n / ). We shall come back to S n later but first proceed to S n3. Recalling D n, we have S n3 = ϕ(x)e n(x) C n (x) Fn(dy) C(y) n(dx). To get an expansion for S n3, the next lemma will be crucial. 5
16 Lemma 3.5. Under the assumptions of Theorem., we have [ ] ϕ(x)γ(x) II n e n(x)+ C(y) F n(dy) C n (x) C(y) n(dx) = o P (n / ). Proof. By Cauchy-Schwarz, on a set with large probability, nii n n / a n a n n / C n (x) C n (x) [ n(dy) C(y) e F n(x)+ (dy) C(y) ] n(dx) n(dy). (3.7) By Lemma 3., we may again replace C n by C. The expectation of the resulting first integral is then less than or equal to a n n / C (y) (dx), which was already shown to be o() in the proof of Lemma 3.3. It therefore remains to show that also the second factor in (3.7) is o P (). Putting we may write z n (x) = n (x) + C(y), [ e z n (x) ] = z n (x)e z n(x), where z n (x) is between zero and z n (x). Recalling that n (x) is between the first integral in B n (x) and and B n (x), we may infer that C(y) z n (x) is uniformly bounded from above in probability, on the support of Fn, as n. Hence it suffices to bound the term a n n / which in turn is less than or equal to a n n / n / zn(x) n(dx), [B n (x) + D n (x) + D n (x)] n(dx) a n [B n(x) + (D n (x) + D n (x)) ]Fn(dx). 6
17 By (3.5), a n Bn(x) n(dx) n a n n(dy) C n(y) n(dx). To bound the right-hand side, replace C n with C first. Then the squared term may be viewed as a V-statistic, of which the leading term is a U- statistic. Its expectation is n n Thus we have to show that a n n ϕ(x) C (y) C (y). F (dx) = o(n / ). This follows as in the proof of Lemma 3.3. Only the powers of F (x) are different. Finally, the error terms involving D n and D n may be dealt with similarly. The proof is complete. Lemma 3.6. Under the assumptions of Theorem., we have ϕγ[cn C](x) CC n (x) Fn(dy) C(y) n(dx) = o P (n / ). Proof. As before. Lemmas 3.5 and 3.6 immediately imply the following Corollary, which brings us to the desired representation of S n3. Corollary 3.. Under the assumptions of Theorem., we have the representation S n3 = ϕ(x)γ(x) Fn(dy) C(y) n(dx) + o P (n / ). We now proceed to S n4. As a first step to get the desired representation, we shall need the following lemma. Lemma 3.7. Under the assumptions of Theorem., S n4 admits the expansion S n4 = ϕ(x)e n(x) C n (x) + o P (n / ). C n (y) + C(y) n C (y) n(dy)fn(dx) (3.8) 7
18 Proof. Recalling the definition of D n, the difference between S n4 and the leading term in (3.8) becomes ϕ(x)e n(x) C n (x) [C n (y) + n C(y)] C (y)[c n (y) + n ] n(dy) n(dx). Taking Lemma 3. into account, the last expression is bounded in absolute values from above, with large probability, by ϕ(x) e n(x) [C n (y) + n C(y)] C 3 (y) n(dy)fn(dx). In the proof of Lemma 3.5 we have argued that z n (x) = n (x) + C(y) is uniformly bounded from above on the support of n, as n, with probability one. Consequently we may replace exp( n (x)) by γ(x). Also we may wish to extend the integral only up to a n. The expectation of the resulting term then does not exceed a n ϕ(x) C(y) + 4 n n F (dx). C 3 (y) These terms already appeared in the proof of Lemma 3.3 and were shown to be o(n / ) there. The proof therefore is complete. In the following we shall omit the summand in (3.8), since its contribution n is also negligible. The next Lemma will enable us to replace exp( n (x)) by γ(x). Lemma 3.8. Under the assumptions of Theorem., we have III n ϕ(x)γ(x) C n (x) = o P (n / ). [ e F n(x)+ (dy) C(y) ] C n C df C n n(dx) Proof. Cauchy-Schwarz leads to a bound for niiin similar to (3.7) for niin. The second factor is exactly the same as before. The first factor is dealt with in the following Lemma. 8
19 Lemma 3.9. Under the assumptions of Theorem., we get C n (y) C(y) C (y) n(dy) Fn(dx) = o P (n / ). Proof. As in previous proofs it is enough to extend the x-integral up to a n. It is then easily checked that the expectation of the resulting term is less than or equal to (when n > 3) a n where ϕ(x) G(y z)( F (y z)) + 4 (n 3)α n (dz)f (dx), C (y)c (z) y z = min(y, z) and y z = max(y, z). Neglecting 4n for a moment, the rest equals a n G(y)( F (z)) ϕ(x) (dz)f (dx) (n 3)α C (y)c (z) = = (n 3) n 3 n 3 a n a n a n y ϕ(x) ϕ(x) ϕ(x) ( F (y)) C(z) z y F (dz)f (dy)f (dx) C(z) F (dy)f (dz)f (dx) ( F (y)) F (dz)f (dx). C(z)( F (z)) The last integral already appeared in the proof of Lemma 3.3 and was shown to be o(n / ). The error term 4 yields similar bounds so that the n proof is complete. In the final step C n in the denominator of (3.8) may be replaced with C. Details are omitted. Hence we come up with the following representation of S n4. Corollary 3.3. Under the assumptions of Theorem., we have S n4 = ϕ(x)γ(x) + o P (n / ). C n (y) C(y) C (y) n(dy)fn(dx) 9
20 We are now ready to give the Proof of Theorem.. Corollaries 3., 3. and 3.3 yield representations of the relevant terms S n, S n3 and S n4 as V-statistics. As is known, see Serfling (98), a V-statistic equals a U-statistic, up to an error o P (n / ). If, in S n, S n3 and S n4, we replace Fn by, we come up with degenerate U-statistics which are all of the order o P (n / ). Collecting the leading terms and an application of Fubini s theorem yield the proof of Theorem. with ˆF n instead of F n. As to F n, apply the inequality i i i a j b j a j b j, a j, b j j= j= j= to the weights of F n and ˆF n. It follows that ϕd ˆF n ϕdf n n 3 n i= ϕ(x i:n ) C n (X i:n ) = n ϕ(x) C n (x) i j= C n(x j:n ) C n(y) n(dy) n(dx). Use Lemma 3. again to replace C n with C and apply the SLLN (for U- statistics) to get that the last term is O P (n ). This completes the proof of Theorem.. Proof of Lemma.. The proof proceeds by induction on i. For i = and j =, the LHS of (.6) equals /nc U n (U :n ), and so does the RHS. The assertion then follows from (.4). For i = and j =, the LHS of (.6) equals, by (.8) and (.5), nc U n (U :n ) nc U n (U :n ) nc U n (U :n ) = nc U n (U :n ) = nc n (X :n ). For i = and j >, the proof follows the same pattern, by repeated use of (.5). Hence the assertion of Lemma. holds true for i =. Assuming that (.6) holds true for all indices less than or equal to i, we obtain, by (.7), F n (X i:n ) = F U n (U idi :n). This may be used as a starting point to show (.6) for i +. Again make repeated use of (.4) and (.5). Acknowledgement. The research of Jane-Ling Wang was supported by an NSF grant NSF
21 References Billingsley, P. (968) Convergence of probability meassures. Wiley, New York. Chao, M.-T. and Lo, S.-H. (988) Some representations of the nonparametric maximum likelihood estimators with truncated data. Ann. Statist.,, He, S. and Yang, G. L. (998a) The strong law under random truncation. Ann. Statist., 6, 99-. He, S. and Yang, G. L. (998b) Estimation of the truncation probability in the random truncation model. Ann. Statist., 6, -7. Keiding, N. and Gill, R. D. (99) Random truncation models and Markov processes. Ann. Statist., 8, Lynden-Bell, D. (97) A method of allowing for known observational selection in small samples applied to 3CR quasars. Monthly Notices Roy. Astronom. Soc., 55, Sellero, C. S., Manteiga, W. G. and Van Keilegom, I. (5) Uniform representation of product-limit integrals with applications. Scand. J. Statist., 3, Serfling, R. J. (98) Approximation theorems of mathematical statistics. Wiley, New York. Stute, W. (993) Almost sure representations of the product-limit estimator for truncated data. Ann. Statist.,, Stute, W. (994) U-Statistic processes: a martingale approach. Ann. Probab.,, Stute, W. (995) The central limit theorem under random censorship. Ann. Statist., 3, Stute, W. (4) Kaplan-Meier integrals. Handbook of Statistics, 3, Elsevier, Amsterdam. Stute, W. and Wang, J. L. (993) The strong law under random censorship. Ann. Statist.,, Wang, M. C. (989) A semiparametric model for randomly truncated data. J. Amer. Statist. Assoc., 84, Woodroofe, M. (985) Estimating a distribution function with truncated data. Ann. Statist., 3,
22 Zhou, M. (99) Some properties of the Kaplan-Meier estimator for independent nonidentically distributed random variables. Ann. Statist., 9,
Estimation of the Bivariate and Marginal Distributions with Censored Data
Estimation of the Bivariate and Marginal Distributions with Censored Data Michael Akritas and Ingrid Van Keilegom Penn State University and Eindhoven University of Technology May 22, 2 Abstract Two new
More informationGOODNESS-OF-FIT TEST FOR RANDOMLY CENSORED DATA BASED ON MAXIMUM CORRELATION. Ewa Strzalkowska-Kominiak and Aurea Grané (1)
Working Paper 4-2 Statistics and Econometrics Series (4) July 24 Departamento de Estadística Universidad Carlos III de Madrid Calle Madrid, 26 2893 Getafe (Spain) Fax (34) 9 624-98-49 GOODNESS-OF-FIT TEST
More informationLecture 21: Convergence of transformations and generating a random variable
Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous
More informationEmpirical likelihood ratio with arbitrarily censored/truncated data by EM algorithm
Empirical likelihood ratio with arbitrarily censored/truncated data by EM algorithm Mai Zhou 1 University of Kentucky, Lexington, KY 40506 USA Summary. Empirical likelihood ratio method (Thomas and Grunkmier
More informationThe main results about probability measures are the following two facts:
Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a
More informationSTAT 6385 Survey of Nonparametric Statistics. Order Statistics, EDF and Censoring
STAT 6385 Survey of Nonparametric Statistics Order Statistics, EDF and Censoring Quantile Function A quantile (or a percentile) of a distribution is that value of X such that a specific percentage of the
More informationChapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability
Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with
More informationAdditive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535
Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June
More informationA COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky
A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky Empirical likelihood with right censored data were studied by Thomas and Grunkmier (1975), Li (1995),
More information11 Survival Analysis and Empirical Likelihood
11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with
More information1 Glivenko-Cantelli type theorems
STA79 Lecture Spring Semester Glivenko-Cantelli type theorems Given i.i.d. observations X,..., X n with unknown distribution function F (t, consider the empirical (sample CDF ˆF n (t = I [Xi t]. n Then
More informationSmooth nonparametric estimation of a quantile function under right censoring using beta kernels
Smooth nonparametric estimation of a quantile function under right censoring using beta kernels Chanseok Park 1 Department of Mathematical Sciences, Clemson University, Clemson, SC 29634 Short Title: Smooth
More informationSTAT 7032 Probability Spring Wlodek Bryc
STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,
More informationWeak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria
Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then
More informationAsymptotic statistics using the Functional Delta Method
Quantiles, Order Statistics and L-Statsitics TU Kaiserslautern 15. Februar 2015 Motivation Functional The delta method introduced in chapter 3 is an useful technique to turn the weak convergence of random
More informationProduct-limit estimators of the survival function with left or right censored data
Product-limit estimators of the survival function with left or right censored data 1 CREST-ENSAI Campus de Ker-Lann Rue Blaise Pascal - BP 37203 35172 Bruz cedex, France (e-mail: patilea@ensai.fr) 2 Institut
More informationAsymptotics of minimax stochastic programs
Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming
More informationThe International Journal of Biostatistics
The International Journal of Biostatistics Volume 1, Issue 1 2005 Article 3 Score Statistics for Current Status Data: Comparisons with Likelihood Ratio and Wald Statistics Moulinath Banerjee Jon A. Wellner
More informationEMPIRICAL ENVELOPE MLE AND LR TESTS. Mai Zhou University of Kentucky
EMPIRICAL ENVELOPE MLE AND LR TESTS Mai Zhou University of Kentucky Summary We study in this paper some nonparametric inference problems where the nonparametric maximum likelihood estimator (NPMLE) are
More informationRegression model with left truncated data
Regression model with left truncated data School of Mathematics Sang Hui-yan 00001107 October 13, 2003 Abstract For linear regression model with truncated data, Bhattacharya, Chernoff, and Yang(1983)(BCY)
More informationis a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.
Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable
More informationChapter 4: Asymptotic Properties of the MLE
Chapter 4: Asymptotic Properties of the MLE Daniel O. Scharfstein 09/19/13 1 / 1 Maximum Likelihood Maximum likelihood is the most powerful tool for estimation. In this part of the course, we will consider
More information4 Sums of Independent Random Variables
4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables
More informationAsymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University
More informationSize and Shape of Confidence Regions from Extended Empirical Likelihood Tests
Biometrika (2014),,, pp. 1 13 C 2014 Biometrika Trust Printed in Great Britain Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests BY M. ZHOU Department of Statistics, University
More informationLaplace s Equation. Chapter Mean Value Formulas
Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic
More informationStatistical inference on Lévy processes
Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline
More informationSTAT Sample Problem: General Asymptotic Results
STAT331 1-Sample Problem: General Asymptotic Results In this unit we will consider the 1-sample problem and prove the consistency and asymptotic normality of the Nelson-Aalen estimator of the cumulative
More informationAsymptotic Properties of Kaplan-Meier Estimator. for Censored Dependent Data. Zongwu Cai. Department of Mathematics
To appear in Statist. Probab. Letters, 997 Asymptotic Properties of Kaplan-Meier Estimator for Censored Dependent Data by Zongwu Cai Department of Mathematics Southwest Missouri State University Springeld,
More informationarxiv: v2 [math.st] 20 Nov 2016
A Lynden-Bell integral estimator for the tail inde of right-truncated data with a random threshold Nawel Haouas Abdelhakim Necir Djamel Meraghni Brahim Brahimi Laboratory of Applied Mathematics Mohamed
More informationTheorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1
Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be
More informationExtreme Value Analysis and Spatial Extremes
Extreme Value Analysis and Department of Statistics Purdue University 11/07/2013 Outline Motivation 1 Motivation 2 Extreme Value Theorem and 3 Bayesian Hierarchical Models Copula Models Max-stable Models
More informationMarshall-Olkin Bivariate Exponential Distribution: Generalisations and Applications
CHAPTER 6 Marshall-Olkin Bivariate Exponential Distribution: Generalisations and Applications 6.1 Introduction Exponential distributions have been introduced as a simple model for statistical analysis
More informationIf g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get
18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in
More informationBayesian estimation of the discrepancy with misspecified parametric models
Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012
More informationBahadur representations for bootstrap quantiles 1
Bahadur representations for bootstrap quantiles 1 Yijun Zuo Department of Statistics and Probability, Michigan State University East Lansing, MI 48824, USA zuo@msu.edu 1 Research partially supported by
More informationInconsistency of the MLE for the joint distribution. of interval censored survival times. and continuous marks
Inconsistency of the MLE for the joint distribution of interval censored survival times and continuous marks By M.H. Maathuis and J.A. Wellner Department of Statistics, University of Washington, Seattle,
More information1. Introduction In many biomedical studies, the random survival time of interest is never observed and is only known to lie before an inspection time
ASYMPTOTIC PROPERTIES OF THE GMLE WITH CASE 2 INTERVAL-CENSORED DATA By Qiqing Yu a;1 Anton Schick a, Linxiong Li b;2 and George Y. C. Wong c;3 a Dept. of Mathematical Sciences, Binghamton University,
More informationFunctional Analysis Exercise Class
Functional Analysis Exercise Class Week 9 November 13 November Deadline to hand in the homeworks: your exercise class on week 16 November 20 November Exercises (1) Show that if T B(X, Y ) and S B(Y, Z)
More informationFULL LIKELIHOOD INFERENCES IN THE COX MODEL
October 20, 2007 FULL LIKELIHOOD INFERENCES IN THE COX MODEL BY JIAN-JIAN REN 1 AND MAI ZHOU 2 University of Central Florida and University of Kentucky Abstract We use the empirical likelihood approach
More informationEfficiency of Profile/Partial Likelihood in the Cox Model
Efficiency of Profile/Partial Likelihood in the Cox Model Yuichi Hirose School of Mathematics, Statistics and Operations Research, Victoria University of Wellington, New Zealand Summary. This paper shows
More informationReal Analysis Math 131AH Rudin, Chapter #1. Dominique Abdi
Real Analysis Math 3AH Rudin, Chapter # Dominique Abdi.. If r is rational (r 0) and x is irrational, prove that r + x and rx are irrational. Solution. Assume the contrary, that r+x and rx are rational.
More information3 Integration and Expectation
3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ
More informationPractical conditions on Markov chains for weak convergence of tail empirical processes
Practical conditions on Markov chains for weak convergence of tail empirical processes Olivier Wintenberger University of Copenhagen and Paris VI Joint work with Rafa l Kulik and Philippe Soulier Toronto,
More informationDensity estimators for the convolution of discrete and continuous random variables
Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract
More informationMAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9
MAT 570 REAL ANALYSIS LECTURE NOTES PROFESSOR: JOHN QUIGG SEMESTER: FALL 204 Contents. Sets 2 2. Functions 5 3. Countability 7 4. Axiom of choice 8 5. Equivalence relations 9 6. Real numbers 9 7. Extended
More informationLarge Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n
Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the
More informationTESTS FOR LOCATION WITH K SAMPLES UNDER THE KOZIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississip
TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississippi University, MS38677 K-sample location test, Koziol-Green
More informationQuantile-quantile plots and the method of peaksover-threshold
Problems in SF2980 2009-11-09 12 6 4 2 0 2 4 6 0.15 0.10 0.05 0.00 0.05 0.10 0.15 Figure 2: qqplot of log-returns (x-axis) against quantiles of a standard t-distribution with 4 degrees of freedom (y-axis).
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationSome basic elements of Probability Theory
Chapter I Some basic elements of Probability Theory 1 Terminology (and elementary observations Probability theory and the material covered in a basic Real Variables course have much in common. However
More informationKrzysztof Burdzy University of Washington. = X(Y (t)), t 0}
VARIATION OF ITERATED BROWNIAN MOTION Krzysztof Burdzy University of Washington 1. Introduction and main results. Suppose that X 1, X 2 and Y are independent standard Brownian motions starting from 0 and
More informationA Conditional Approach to Modeling Multivariate Extremes
A Approach to ing Multivariate Extremes By Heffernan & Tawn Department of Statistics Purdue University s April 30, 2014 Outline s s Multivariate Extremes s A central aim of multivariate extremes is trying
More informationLecture 4 Lebesgue spaces and inequalities
Lecture 4: Lebesgue spaces and inequalities 1 of 10 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 4 Lebesgue spaces and inequalities Lebesgue spaces We have seen how
More informationApproximate Self Consistency for Middle-Censored Data
Approximate Self Consistency for Middle-Censored Data by S. Rao Jammalamadaka Department of Statistics & Applied Probability, University of California, Santa Barbara, CA 93106, USA. and Srikanth K. Iyer
More informationThe best expert versus the smartest algorithm
Theoretical Computer Science 34 004 361 380 www.elsevier.com/locate/tcs The best expert versus the smartest algorithm Peter Chen a, Guoli Ding b; a Department of Computer Science, Louisiana State University,
More informationLecture 6 Basic Probability
Lecture 6: Basic Probability 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 6 Basic Probability Probability spaces A mathematical setup behind a probabilistic
More informationGoodness-of-fit tests for the cure rate in a mixture cure model
Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,
More informationCharacterization of Upper Comonotonicity via Tail Convex Order
Characterization of Upper Comonotonicity via Tail Convex Order Hee Seok Nam a,, Qihe Tang a, Fan Yang b a Department of Statistics and Actuarial Science, University of Iowa, 241 Schaeffer Hall, Iowa City,
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationConvergence rates in weighted L 1 spaces of kernel density estimators for linear processes
Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research
More informationExercises in Extreme value theory
Exercises in Extreme value theory 2016 spring semester 1. Show that L(t) = logt is a slowly varying function but t ǫ is not if ǫ 0. 2. If the random variable X has distribution F with finite variance,
More informationAFT Models and Empirical Likelihood
AFT Models and Empirical Likelihood Mai Zhou Department of Statistics, University of Kentucky Collaborators: Gang Li (UCLA); A. Bathke; M. Kim (Kentucky) Accelerated Failure Time (AFT) models: Y = log(t
More informationBrownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539
Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationGeneral Glivenko-Cantelli theorems
The ISI s Journal for the Rapid Dissemination of Statistics Research (wileyonlinelibrary.com) DOI: 10.100X/sta.0000......................................................................................................
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationProduct measure and Fubini s theorem
Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω
More informationCURRENT STATUS LINEAR REGRESSION. By Piet Groeneboom and Kim Hendrickx Delft University of Technology and Hasselt University
CURRENT STATUS LINEAR REGRESSION By Piet Groeneboom and Kim Hendrickx Delft University of Technology and Hasselt University We construct n-consistent and asymptotically normal estimates for the finite
More informationLIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974
LIMITS FOR QUEUES AS THE WAITING ROOM GROWS by Daniel P. Heyman Ward Whitt Bell Communications Research AT&T Bell Laboratories Red Bank, NJ 07701 Murray Hill, NJ 07974 May 11, 1988 ABSTRACT We study the
More information5 Introduction to the Theory of Order Statistics and Rank Statistics
5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order
More informationTHE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION
THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION JAINUL VAGHASIA Contents. Introduction. Notations 3. Background in Probability Theory 3.. Expectation and Variance 3.. Convergence
More information2 Functions of random variables
2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as
More information1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =
Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values
More informationExtreme Value Theory and Applications
Extreme Value Theory and Deauville - 04/10/2013 Extreme Value Theory and Introduction Asymptotic behavior of the Sum Extreme (from Latin exter, exterus, being on the outside) : Exceeding the ordinary,
More informationSelf-normalized Cramér-Type Large Deviations for Independent Random Variables
Self-normalized Cramér-Type Large Deviations for Independent Random Variables Qi-Man Shao National University of Singapore and University of Oregon qmshao@darkwing.uoregon.edu 1. Introduction Let X, X
More informationWLLN for arrays of nonnegative random variables
WLLN for arrays of nonnegative random variables Stefan Ankirchner Thomas Kruse Mikhail Urusov November 8, 26 We provide a weak law of large numbers for arrays of nonnegative and pairwise negatively associated
More informationTHE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON
GEORGIAN MATHEMATICAL JOURNAL: Vol. 3, No. 2, 1996, 153-176 THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON M. SHASHIASHVILI Abstract. The Skorokhod oblique reflection problem is studied
More informationRefining the Central Limit Theorem Approximation via Extreme Value Theory
Refining the Central Limit Theorem Approximation via Extreme Value Theory Ulrich K. Müller Economics Department Princeton University February 2018 Abstract We suggest approximating the distribution of
More informationChapter 2: Resampling Maarten Jansen
Chapter 2: Resampling Maarten Jansen Randomization tests Randomized experiment random assignment of sample subjects to groups Example: medical experiment with control group n 1 subjects for true medicine,
More informationChapter 7. Extremal Problems. 7.1 Extrema and Local Extrema
Chapter 7 Extremal Problems No matter in theoretical context or in applications many problems can be formulated as problems of finding the maximum or minimum of a function. Whenever this is the case, advanced
More informationSupplement to Quantile-Based Nonparametric Inference for First-Price Auctions
Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract
More informationLecture 8: Information Theory and Statistics
Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang
More informationA note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis
A note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis Jon A. Wellner and Vladimir Koltchinskii Abstract. Proofs are given of the limiting null distributions of the
More informationRandom Bernstein-Markov factors
Random Bernstein-Markov factors Igor Pritsker and Koushik Ramachandran October 20, 208 Abstract For a polynomial P n of degree n, Bernstein s inequality states that P n n P n for all L p norms on the unit
More informationMath212a1413 The Lebesgue integral.
Math212a1413 The Lebesgue integral. October 28, 2014 Simple functions. In what follows, (X, F, m) is a space with a σ-field of sets, and m a measure on F. The purpose of today s lecture is to develop the
More informationPrinciple of Mathematical Induction
Advanced Calculus I. Math 451, Fall 2016, Prof. Vershynin Principle of Mathematical Induction 1. Prove that 1 + 2 + + n = 1 n(n + 1) for all n N. 2 2. Prove that 1 2 + 2 2 + + n 2 = 1 n(n + 1)(2n + 1)
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationA PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS
Statistica Sinica 20 2010, 365-378 A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS Liang Peng Georgia Institute of Technology Abstract: Estimating tail dependence functions is important for applications
More informationEstimation of Dynamic Regression Models
University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.
More informationASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE. By Michael Nussbaum Weierstrass Institute, Berlin
The Annals of Statistics 1996, Vol. 4, No. 6, 399 430 ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE By Michael Nussbaum Weierstrass Institute, Berlin Signal recovery in Gaussian
More informationResearch Article Exponential Inequalities for Positively Associated Random Variables and Applications
Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 008, Article ID 38536, 11 pages doi:10.1155/008/38536 Research Article Exponential Inequalities for Positively Associated
More informationFinite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product
Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )
More informationMATH 409 Advanced Calculus I Lecture 12: Uniform continuity. Exponential functions.
MATH 409 Advanced Calculus I Lecture 12: Uniform continuity. Exponential functions. Uniform continuity Definition. A function f : E R defined on a set E R is called uniformly continuous on E if for every
More informationProblem 1: Compactness (12 points, 2 points each)
Final exam Selected Solutions APPM 5440 Fall 2014 Applied Analysis Date: Tuesday, Dec. 15 2014, 10:30 AM to 1 PM You may assume all vector spaces are over the real field unless otherwise specified. Your
More informationINDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem
More informationFULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH
FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH Jian-Jian Ren 1 and Mai Zhou 2 University of Central Florida and University of Kentucky Abstract: For the regression parameter
More informationSection 8.2. Asymptotic normality
30 Section 8.2. Asymptotic normality We assume that X n =(X 1,...,X n ), where the X i s are i.i.d. with common density p(x; θ 0 ) P= {p(x; θ) :θ Θ}. We assume that θ 0 is identified in the sense that
More information