36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1
|
|
- Maude Greer
- 5 years ago
- Views:
Transcription
1 36. Multisample U-statistics jointly distributed U-statistics Lehmann 6.1 In this topic, we generalize the idea of U-statistics in two different directions. First, we consider single U-statistics for situations in which there is more than one sample. Next, we consider the joint asymptotic distribution of two (single-sample) U-statistics. We begin by generalizing the idea of U-statistics to the case in which we have more than one rom sample. Suppose that X i1,..., X ini is an iid sample from F i for all 1 i s. In other words, we have s rom samples, each potentially from a different distribution, n i is the size of the ith sample. We may define a statistical functional θ = E φ (X 11,..., X 1a1 ; X 21,..., X 2a2 ; ; X s1,..., X sas ). (132) Notice that the kernel φ in Equation (132) has a 1 +a 2 + +a s arguments; furthermore, we assume that the first a 1 of them may be permuted without changing the value of φ, the next a 2 of them may be permuted without changing the value of φ, etc. In other words, there are s distinct blocks of arguments of φ, φ is symmetric in its arguments within each of these blocks. Not surprisingly, the U-statistic corresponding to the functional (132) is U N = ( 1 1 n1 ) ( ns ) a 1 a s 1 i 1< <i a1 n 1 1 r 1< <r as n s Note that N denotes the total of all the sample sizes: N = n n s. φ ( X 1i1,..., X 1ia1 ; ; X sr1,..., X sras ). (133) As we did in the case of single-sample U-statistics, define for 0 j 1 a i,..., 0 j s a s φ j1 j s (X 11,..., X 1j1 ; ; X s1,..., X sjs ) = E {φ(x 11,..., X 1a1 ; ; X s1,..., X sas ) X 11,..., X 1j1,, X s1,..., X sjs } (134) σ 2 j1 j s = Var φ j1 j s (X 11,..., X 1j1 ; ; X s1,..., X sjs ). (135) By an argument similar to the one used in the proof of Theorem 35.1, but much more tedious notationally, we can show that σ 2 j 1 j s = Cov {φ(x 11,..., X 1a1 ; ; X s1,..., X sas ), φ(x 11,..., X 1j1, X 1,a1+1,..., X 1,a1+(a 1 j 1); ; X s1,..., X sjs, X s,as+1,..., X s,as+(a s j s)) }. (136) Notice that some of the j i may equal 0. This was not true in the single-sample case, since φ 0 would have merely been the constant θ, so σ 2 0 = 0. In the special case when s = 2, Equations (134), (135) (136) become φ ij (X 1,..., X i ; Y 1,..., Y j ) = E {φ(x 1,..., X a1 ; Y 1,..., Y a2 ) X 1,..., X i, Y 1,..., Y j }, σ 2 ij = Var φ ij (X 1,..., X i ; Y 1..., Y j ), σ 2 ij = Cov {φ(x 1,..., X a1 ; Y 1,..., Y a2 ), φ(x 1,..., X i, X a1+1,..., X a1+(a 1 i); Y 1,..., Y j, Y a2+1,..., Y a2+(a 2 j)) }, 104
2 respectively, for 0 i a 1 0 j a 2. Although we will not derive it here as we did for the single-sample case, there is an analagous asymptotic normality result for multisample U-statistics, as follows. Theorem 36.1 Suppose that for i = 1,..., s, X i1,..., X ini is a rom sample from the distribution F i that these s samples are independent of each other. Suppose further that there exist constants ρ 1,..., ρ s in the interval (0, 1) such that n i /N ρ i for all i that σ 2 a 1 a s <. Then N(UN θ) L N(0, σ 2 ), where σ 2 = a2 1 ρ 1 σ a2 s ρ s σ Although the notation required for the multisample U-statistic theory is nightmarish, life becomes considerably simpler in the case s = 2 a 1 = a 2 = 1, in which case we obtain U N = 1 n 1 n 2 n 1 n 2 φ(x 1i ; X 2j ). j=1 Equivalently, we may assume that X 1,..., X m are an iid sample from F Y 1,..., Y n are an iid sample from G, which gives U N = 1 mn m j=1 In the case of the U-statistic of Equation (137), Theorem 36.1 states that ( ) N(UN θ) L N 0, σ2 10 ρ + σ2 01, 1 ρ n φ(x i ; Y j ). (137) where ρ = lim m/n, σ 2 10 = Cov {φ(x 1 ; Y 1 ), φ(x 1 ; Y 2 )}, σ 2 01 = Cov {φ(x 1 ; Y 1 ), φ(x 2 ; Y 1 )}. Example 36.1 For X 1,... X m iid from F Y 1,..., Y n iid from G, consider the Wilcoxon ranksum statistic W, defined to be the sum of the ranks of the Y i among the combined sample. We showed in Equation (79) that W = 1 2 n(n + 1) + m j=1 n I{X i < Y j }. Therefore, if we let φ(x; y) = I{x < y}, then the corresponding two-sample U-statistic U N is related to W by W = 1 2 n(n + 1) + mnu N. Therefore, we may use Theorem 36.1 to obtain the asymptotic normality of U N, therefore of W, just as we did in Topic 26. However, unlike in Topic 26, we make no assumption here that F G are merely shifted versions of one another. Thus, we may now obtain in principle the asymptotic distribution of the rank-sum statistic for any two distributions F G that we wish, so long as they have finite second moments. The other direction in which we will generalize the development of U-statistics is consideration of the joint distribution of two single-sample U-statistics. Suppose that there are two kernel functions, φ(x 1,..., x a ) ϕ(x 1,..., x b ), we define the two corresponding U-statistics U (1) n = 1 ( n a) 1 i 1< <i a n φ(x i1,..., X ia ) 105
3 U (2) n = 1 ( n b) 1 j 1< <j b n φ(x j1,..., X ja ) for a single rom sample X 1,..., X n from F. Define θ 1 = E U n (1) θ 2 = E U n (2). Furthermore, define γ ij to be the covariance between φ i (X 1,..., X i ) ϕ j (X 1,..., X j ), where φ i ϕ j are defined as in Equation (126). It may be proved that γ ij = Cov { φ(x 1,..., X a ), ϕ(x 1,..., X j, X a+1,..., X a+(b j) ) }. (138) Note in particular that γ ij depends only on the value of min{i, j}. The following theorem, stated without proof, gives the joint asymptotic distribution of U (1) n U (2) n. Theorem 36.2 Suppose X 1,..., X n is a rom sample from F that φ ϕ are two kernel functions satisfying Var φ(x 1,..., X a ) < Var ϕ(x 1,..., X b ) <. Define τ1 2 = Var φ 1 (X 1 ) τ2 2 = Var ϕ 1 (X 1 ), let γ ij be defined as in Equation (138). Then { (U (1)) n n U n (2) ( θ1 θ 2 ) } {( ) L 0 N, 0 ( a 2 τ 2 1 abγ 11 abγ 11 b 2 τ 2 2 )}. Problems Problem 36.1 Suppose X 1,..., X m Y 1,..., Y n are independent iid samples from distributions Unif(0, θ) Unif(µ, µ + θ), respectively. Assume m/n ρ as m, n 0 < µ < θ. (a) Find the asymptotic distribution of the U-statistic of Equation (137), where φ(x; y) = I{x < y}. In so doing, find a function g(x) such that E (U N ) = g(µ). (b) Find the asymptotic distribution of g(y X). (c) Find the range of values of µ for which the Wilcoxon estimate of g(µ) is asymptotically more efficient than g(y X). (The ARE in this case is the ratio of asymptotic variances. See the discussion of (4.3.12) (4.3.13) on p. 240 for details.) Problem 36.2 Solve each part of Problem 36.1, but this time under the assumptions that the independent rom samples X 1,..., X m Y 1,..., Y n satisfy P (X 1 t) = P (Y 1 θ t) = t 2 for t [0, 1] 0 < θ < 1. As in Problem 36.1, assume m/n ρ (0, 1). Problem 36.3 Suppose X 1,..., X m Y 1,..., Y n are independent iid samples from distributions N(0, 1) N(µ, 1), respectively. Assume m/(m + n) 1/2 as m, n. Let U N be the U-statistic of Equation (137), where φ(x; y) = I{x < y}. Suppose that θ(µ) σ 2 (µ) are such that N[UN θ(µ)] L N[0, σ 2 (µ)]. Calculate θ(µ) σ 2 (µ) for µ {.2,.5, 1, 1.5, 2}. Hint: This problem requires a bit of numerical integration. There are a couple of ways you might do this. Mathematica will do it quite easily. There is a function called integrate in Splus one called quad in MATLAB for integrating a function. If you cannot get any of these to work for you, let me know. 106
4 Problem 36.4 Suppose X 1, X 2,... are iid with finite variance. Define S 2 n = 1 n 1 n (x i x) 2 let G n be Gini s mean difference, the U-statistic defined in Problem Note that S 2 n is also a U-statistic, corresponding to the kernel function φ(x 1, x 2 ) = (x 1 x 2 ) 2. (a) If X i are distributed as Unif(0, θ), give the joint asymptotic distribution of G n S n by first finding the joint asymptotic distribution of the U-statistics G n Sn. 2 Note that the covariance matrix need not be positive definite; in this problem, the covariance matrix is singular. (b) The singular asymptotic covariance matrix in this problem implies that as n, the joint distribution of G n S n becomes concentrated on a line. Does this appear to be the case? For 1000 samples of size n from Uniform(0, 1), plot scatterplots of G n against S n. Take n {5, 25, 100}. 107
5 37. Bootstrapping Lehmann 6.5 In this topic, we introduce the bootstrap, even though the asymptotics content of what follows is not very high. Essentially, the only large-sample theory we rely on here is the weak law of large numbers. Nonetheless, bootstrapping is included here because of its natural relationship with the concepts of statistical functionals plug-in estimators seen in recent topics, also because it is an increasingly popular often misunderstood method. Consider a statistical functional λ n (F ) that depends on n. For instance, λ n (F ) may be some property, such as bias or variance, of an estimator ˆθ n of θ = θ(f ) based on a rom sample of size n from some distribution F. As an example, let θ(f ) = F 1 ( 1 2) be the median of F. Take ˆθn to be the mth order statistic from a rom sample of size n = 2m 1 from F. Consider the bias λ B n (F ) = E F ˆθn θ(f ) the variance λ V n (F ) = E F ˆθ2 n (E F ˆθn ) 2. Theoretical properties of λ B n λ V n are very difficult to obtain. Even asymptotics aren t very helpful, since n(ˆθ n θ) L N{0, 1/(4f 2 (θ))} tells us only that the bias goes to zero the limiting variance may be very hard to estimate because it involves the unknown quantity f(θ), which is hard to estimate. Consider the plug-in estimators λ B n ( ˆF n ) λ V n ( ˆF n ). (Recall that ˆF n denotes the empirical distribution function, which puts a mass of 1 n on each of the n sample points.) In our median example, λ B n ( ˆF n ) = E ˆFn ˆθ n ˆθ n λ V n ( ˆF n ) = E ˆFn (ˆθ n) 2 (E ˆFn ˆθ n ) 2, where ˆθ n is the sample median from a rom sample X 1,..., X n from ˆF n. To see how difficult it is to calculate λ B n ( ˆF n ) λ V n ( ˆF n ), consider the simplest nontrivial case, n = 3: Conditional on the order statistics (X (1), X (2), X (3) ), there are 27 equally likely possibilities for the value of (X 1, X 2, X 3 ), the sample of size 3 from ˆF n, namely (X (1), X (1), X (1) ), (X (1), X (1), X (2) ),..., (X (3), X (3), X (3) ). Of these 27 possibilities, exactly = 7 have the value X (1) occurring 2 or 3 times. Therefore, we obtain This implies that P (ˆθ n = X (1) ) = 7 27, P (ˆθ n = X (2) ) = 13 27, P (ˆθ n = X (3) ) = E ˆFn ˆθ n = 1 27 (7X (1) + 13X (2) + 7X (3) ) E ˆFn (ˆθ n) 2 = 1 27 (7X2 (1) + 13X2 (2) + 7X2 (3) ). Therefore, since ˆθ n = X (2), we obtain λ B n ( ˆF n ) = 1 27 (7X (1) 14X (2) + 7X (3) ) λ V n ( ˆF n ) = (10X2 (1) + 13X2 (2) + 10X2 (3) 13X (1)X (2) 13X (2) X (3) 7X (1) X (3) ). 108
6 To obtain the sampling distribution of these estimators, of course, we would have to consider the joint distribution of (X (1), X (2), X (3) ). Naturally, the calculations become even more difficult as n increases. Alternatively, we could use resampling in order to approximate λ B n ( ˆF n ) λ V n ( ˆF n ). This is the bootstrapping idea, it works like this: For some large number B, simulate B rom samples from ˆF n, namely X 11,..., X 1n,. X B1,..., X Bn, approximate a quantity like E ˆFn ˆθ n by the sample mean 1 B B ˆθ in, where ˆθ in is the sample median of the ith bootstrap sample X i1,..., X in. Notice that the weak law of large numbers asserts that 1 B B ˆθ in P E ˆFn ˆθ n. To recap, then, we wish to estimate some parameter λ n (F ) for an unknown distribution F based on a rom sample from F. We estimate λ n (F ) by λ n ( ˆF n ), but it is not easy to evaluate λ n ( ˆF n ) so we approximate λ n ( ˆF n ) by resampling B times from ˆF n obtain a bootstrap estimator λ B,n. Thus, there are two relevant issues: 1. How good is the approximation of λ n ( ˆF n ) by λ B,n? (Note that λ n( ˆF n ) is NOT an unknown parameter; it is known but hard to evaluate.) 2. How precise is the estimation of λ n (F ) by λ n ( ˆF n )? Question 1 is usually easy to answer using asymptotics; these asymptotics involve letting B usually rely on the weak law or the central limit theorem. For example, if we have an expectation functional λ n (F ) = E F h(x 1,..., X n ), then as B. λ B,n = 1 B B h(xi1,..., Xin) P λ n ( ˆF n ) Question 2, on the other h, is often tricky; asymptotic results involve letting n are hled case-by-case. We will not discuss these asymptotics here. On a related note, however, there is an argument in Lehmann s book (on pages ) about why a plug-in estimator may be better than an asymptotic estimator. That is, if it is possible to show λ n (F ) λ as n, then as an estimator of λ n (F ), λ n ( ˆF n ) may be preferable to λ. We conclude this topic by considering the so-called parametric bootstrap. If we assume that the unknown distribution function F comes from a family of distribution functions indexed by a parameter µ, then λ n (F ) is really λ n (F µ ). Then, instead of the plug-in estimator λ n ( ˆF n ), we might consider the estimator λ n (Fˆµ ), where ˆµ is an estimator of µ. Everything proceeds as in the nonparametric version of bootstrapping. Since it may not be easy to evaulate λ n (Fˆµ ) explicitly, we first find ˆµ then take B rom samples of size n, X11,..., X1n through XB1,..., X Bn, from Fˆµ. These samples are used to approximate λ n (Fˆµ ). 109
7 Example 37.1 Suppose X 1,..., X n is a rom sample from Poisson(µ). Take ˆµ = X. Suppose λ n (F µ ) = Var Fµ ˆµ. In this case, we happen to know that λ n (F µ ) = µ/n, but let s ignore this knowledge apply a parametric bootstrap. For some large B, say 500, generate B samples from Poisson(ˆµ) use the sample variance of ˆµ as an approximation to λ n (Fˆµ ). In R, with µ = 1 n = 20 we obtain > x <- rpois(20,1) # Generate the sample from F > muhat <- mean(x) > muhat [1] 0.85 > muhatstar <- rep(0,500) # Allocate the vector for muhatstar > for(i in 1:500) muhatstar[i] <- mean(rpois(20,muhat)) > var(muhatstar) [1] Note that the estimate is close to the known true value Obviously, this example is simplistic because we already know that λ n (F ) = µ/n, which makes ˆµ/n a more natural estimator. However, it is not always so simple to obtain a closed-form expression for λ n (F ). Incidentally, we could also use a nonparametric bootstrap approach in this example: > for (i in 1:500) muhatstar2[i] <- mean(sample(x,replace=t)) > var(muhatstar2) [1] Of course, is an approximation to λ n ( ˆF n ) rather than λ n (Fˆµ ). Furthermore, we can obtain a result arbitrarily close to λ n ( ˆF n ) by increasing the value of B: > muhatstar2_rep(0,100000) > for (i in 1:100000) muhatstar2[i] <- mean(sample(x,replace=t)) > var(muhatstar2) [1] In fact, it is in principle possible to obtain an approximate variance for our estimates of λ n ( ˆF n ) λ n (Fˆµ ),, using the central limit theorem, construct approximate confidence intervals for these quantities. This would allow us to specify the quantities to any desired level of accuracy. Problems Problem 37.1 (a) Devise a nonparametric bootstrap scheme for setting confidence intervals for β in the linear regression model Y i = α + βx i + ɛ i. There is more than one possible answer. (b) Using B = 1000, implement your scheme on the following dataset to obtain a 95% confidence interval. Compare your answer with the stard 95% confidence interval. Y x (In the dataset, Y is the number of manatee deaths due to collisions with powerboats in Florida x is the number of powerboat registrations in thouss for even years from ) Problem 37.2 Consider the following dataset that lists the latitude mean August temperature for 7 US cities. The residuals are listed for use in part (b). 110
8 City Miami Phoenix Memphis Baltimore Pittsburgh Boston Portl, OR Latitude Temperature Residual Minitab gives the following output for a simple linear regression: Predictor Coef SE Coef T P Constant latitude S = R-Sq = 61.5% R-Sq(adj) = 53.8% Note that this gives an asymptotic estimate of the variance of the slope parameter as = In each case below, use the described method to simulate B = 500 bootstrap samples (x b1, y b1 ),..., (x b7, y b7 ) for b = 1,..., B. For each b, refit the model to obtain ˆβ b. Report the sample variance of ˆβ 1,..., ˆβ B compare with the asymptotic estimate of (a) Parametric bootstrap. Take x bi = x i for all b i. Let ybi = ˆβ 0 + ˆβ 1 x i + ɛ i, where ɛ i N(0, ˆσ 2 ). Obtain ˆβ 0, ˆβ 1, ˆσ 2 from the above output. (b) Nonparametric bootstrap I. Take x bi = x i for all b i. Let ybi = ˆβ 0 + ˆβ 1 x i + rbi, where rb1,..., r b7 is an iid sample from the empirical distribution of the residuals from the original model (you may want to refit the original model to find these residuals). (c) Nonparametric bootstrap II. Let (x b1, y b1 ),..., (x b7, y b7 ) be an iid sample from the empirical distribution of (x 1, y 1 ),..., (x 7, y 7 ). Note: In Splus, you can obtain the slope coefficient of the linear regression of the vector y on the vector x using lm(y~x)$coef[2]. Problem 37.3 The same resampling idea that is exploited in the bootstrap can be used to approximate the value of difficult integrals by a technique sometimes called Monte Carlo integration. Suppose we wish to compute θ = e x2 cos 3 (x) dx. Using the NIntegrate function in Mathematica, we find that θ = (a) Define g(t) = 2e t2 cos 3 (t). Let U 1,..., U n be an iid uniform(0,1) sample. Let Prove that ˆθ 1 P θ. ˆθ 1 = 1 n n g(u i ). (b) Define h(t) = 2 2t. Prove that if we take V i = 1 U i for each i, then V i is a rom variable with density h(t). Prove that with we have ˆθ 2 P θ. ˆθ 2 = 1 n n g(v i ) h(v i ), 111
9 (c) For n = 1000, simulate ˆθ 1 ˆθ 2. Give estimates of the variance for each estimator by reporting ˆσ 2 /n for each, where ˆσ 2 is the sample variance of the g(u i ) or the g(v i )/h(v i ) as the case may be. (d) Plot, on the same set of axes, g(t), h(t), the stard uniform density for t [0, 1]. From this plot, explain why the variance of ˆθ 2 is smaller than the variance of ˆθ 1. [Incidentally, the technique of drawing rom variables from a density h whose shape is close to the function g of interest is a variance-reduction technique known as importance sampling.] Note: This was sort of a silly example, since it was really easy to get an exact value for θ using numerical methods. However, with certain high-dimensional integrals, the curse of dimensionality makes exact numerical methods extremely time-consuming computationally; thus, Monte Carlo integration does have a practical use in such cases. 112
Chapter 10. U-statistics Statistical Functionals and V-Statistics
Chapter 10 U-statistics When one is willing to assume the existence of a simple random sample X 1,..., X n, U- statistics generalize common notions of unbiased estimation such as the sample mean and the
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationAdvanced Statistics II: Non Parametric Tests
Advanced Statistics II: Non Parametric Tests Aurélien Garivier ParisTech February 27, 2011 Outline Fitting a distribution Rank Tests for the comparison of two samples Two unrelated samples: Mann-Whitney
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Output Analysis for Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Output Analysis
More informationChapter 2: Resampling Maarten Jansen
Chapter 2: Resampling Maarten Jansen Randomization tests Randomized experiment random assignment of sample subjects to groups Example: medical experiment with control group n 1 subjects for true medicine,
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order
More informationBootstrap tests. Patrick Breheny. October 11. Bootstrap vs. permutation tests Testing for equality of location
Bootstrap tests Patrick Breheny October 11 Patrick Breheny STA 621: Nonparametric Statistics 1/14 Introduction Conditioning on the observed data to obtain permutation tests is certainly an important idea
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2
MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 1 Bootstrapped Bias and CIs Given a multiple regression model with mean and
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationThe bootstrap. Patrick Breheny. December 6. The empirical distribution function The bootstrap
Patrick Breheny December 6 Patrick Breheny BST 764: Applied Statistical Modeling 1/21 The empirical distribution function Suppose X F, where F (x) = Pr(X x) is a distribution function, and we wish to estimate
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend
More informationLecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf
Lecture 13: 2011 Bootstrap ) R n x n, θ P)) = τ n ˆθn θ P) Example: ˆθn = X n, τ n = n, θ = EX = µ P) ˆθ = min X n, τ n = n, θ P) = sup{x : F x) 0} ) Define: J n P), the distribution of τ n ˆθ n θ P) under
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationProblem Selected Scores
Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected
More informationStatement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.
MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss
More informationConfidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods
Chapter 4 Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods 4.1 Introduction It is now explicable that ridge regression estimator (here we take ordinary ridge estimator (ORE)
More informationThe Nonparametric Bootstrap
The Nonparametric Bootstrap The nonparametric bootstrap may involve inferences about a parameter, but we use a nonparametric procedure in approximating the parametric distribution using the ECDF. We use
More informationBootstrap & Confidence/Prediction intervals
Bootstrap & Confidence/Prediction intervals Olivier Roustant Mines Saint-Étienne 2017/11 Olivier Roustant (EMSE) Bootstrap & Confidence/Prediction intervals 2017/11 1 / 9 Framework Consider a model with
More informationUniversity of California San Diego and Stanford University and
First International Workshop on Functional and Operatorial Statistics. Toulouse, June 19-21, 2008 K-sample Subsampling Dimitris N. olitis andjoseph.romano University of California San Diego and Stanford
More informationSTAT 830 Non-parametric Inference Basics
STAT 830 Non-parametric Inference Basics Richard Lockhart Simon Fraser University STAT 801=830 Fall 2012 Richard Lockhart (Simon Fraser University)STAT 830 Non-parametric Inference Basics STAT 801=830
More informationBootstrap. Director of Center for Astrostatistics. G. Jogesh Babu. Penn State University babu.
Bootstrap G. Jogesh Babu Penn State University http://www.stat.psu.edu/ babu Director of Center for Astrostatistics http://astrostatistics.psu.edu Outline 1 Motivation 2 Simple statistical problem 3 Resampling
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Bootstrap for Regression Week 9, Lecture 1
MA 575 Linear Models: Cedric E. Ginestet, Boston University Bootstrap for Regression Week 9, Lecture 1 1 The General Bootstrap This is a computer-intensive resampling algorithm for estimating the empirical
More informationProperties of the least squares estimates
Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares
More informationSTAT 730 Chapter 4: Estimation
STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum
More information401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.
401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis
More informationChapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics
Chapter 6 Order Statistics and Quantiles 61 Extreme Order Statistics Suppose we have a finite sample X 1,, X n Conditional on this sample, we define the values X 1),, X n) to be a permutation of X 1,,
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationBootstrap, Jackknife and other resampling methods
Bootstrap, Jackknife and other resampling methods Part III: Parametric Bootstrap Rozenn Dahyot Room 128, Department of Statistics Trinity College Dublin, Ireland dahyot@mee.tcd.ie 2005 R. Dahyot (TCD)
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationRandom vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.
Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just
More informationLecture 1: Random number generation, permutation test, and the bootstrap. August 25, 2016
Lecture 1: Random number generation, permutation test, and the bootstrap August 25, 2016 Statistical simulation 1/21 Statistical simulation (Monte Carlo) is an important part of statistical method research.
More informationOne-Sample Numerical Data
One-Sample Numerical Data quantiles, boxplot, histogram, bootstrap confidence intervals, goodness-of-fit tests University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationCasuality and Programme Evaluation
Casuality and Programme Evaluation Lecture V: Difference-in-Differences II Dr Martin Karlsson University of Duisburg-Essen Summer Semester 2017 M Karlsson (University of Duisburg-Essen) Casuality and Programme
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationFinal Examination Statistics 200C. T. Ferguson June 11, 2009
Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean
More informationOn robust and efficient estimation of the center of. Symmetry.
On robust and efficient estimation of the center of symmetry Howard D. Bondell Department of Statistics, North Carolina State University Raleigh, NC 27695-8203, U.S.A (email: bondell@stat.ncsu.edu) Abstract
More informationSTAT 730 Chapter 5: Hypothesis Testing
STAT 730 Chapter 5: Hypothesis Testing Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 28 Likelihood ratio test def n: Data X depend on θ. The
More information5601 Notes: The Sandwich Estimator
560 Notes: The Sandwich Estimator Charles J. Geyer December 6, 2003 Contents Maximum Likelihood Estimation 2. Likelihood for One Observation................... 2.2 Likelihood for Many IID Observations...............
More informationSTAT440/840: Statistical Computing
First Prev Next Last STAT440/840: Statistical Computing Paul Marriott pmarriott@math.uwaterloo.ca MC 6096 February 2, 2005 Page 1 of 41 First Prev Next Last Page 2 of 41 Chapter 3: Data resampling: the
More informationECE531 Lecture 10b: Maximum Likelihood Estimation
ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So
More information1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).
Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,
More information0.1 The plug-in principle for finding estimators
The Bootstrap.1 The plug-in principle for finding estimators Under a parametric model P = {P θ ;θ Θ} (or a non-parametric P = {P F ;F F}), any real-valued characteristic τ of a particular member P θ (or
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research
More informationMLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22
MLE and GMM Li Zhao, SJTU Spring, 2017 Li Zhao MLE and GMM 1 / 22 Outline 1 MLE 2 GMM 3 Binary Choice Models Li Zhao MLE and GMM 2 / 22 Maximum Likelihood Estimation - Introduction For a linear model y
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationInference based on robust estimators Part 2
Inference based on robust estimators Part 2 Matias Salibian-Barrera 1 Department of Statistics University of British Columbia ECARES - Dec 2007 Matias Salibian-Barrera (UBC) Robust inference (2) ECARES
More informationContents 1. Contents
Contents 1 Contents 1 One-Sample Methods 3 1.1 Parametric Methods.................... 4 1.1.1 One-sample Z-test (see Chapter 0.3.1)...... 4 1.1.2 One-sample t-test................. 6 1.1.3 Large sample
More informationNon-Parametric Bootstrap Mean. Squared Error Estimation For M- Quantile Estimators Of Small Area. Averages, Quantiles And Poverty
Working Paper M11/02 Methodology Non-Parametric Bootstrap Mean Squared Error Estimation For M- Quantile Estimators Of Small Area Averages, Quantiles And Poverty Indicators Stefano Marchetti, Nikos Tzavidis,
More informationRecall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n
Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the
More informationInference For High Dimensional M-estimates. Fixed Design Results
: Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and
More informationCox s proportional hazards/regression model - model assessment
Cox s proportional hazards/regression model - model assessment Rasmus Waagepetersen September 27, 2017 Topics: Plots based on estimated cumulative hazards Cox-Snell residuals: overall check of fit Martingale
More informationLinear Methods for Prediction
Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we
More informationBrief Review on Estimation Theory
Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on
More informationBetter Bootstrap Confidence Intervals
by Bradley Efron University of Washington, Department of Statistics April 12, 2012 An example Suppose we wish to make inference on some parameter θ T (F ) (e.g. θ = E F X ), based on data We might suppose
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationStatistics & Data Sciences: First Year Prelim Exam May 2018
Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book
More informationSparse Nonparametric Density Estimation in High Dimensions Using the Rodeo
Outline in High Dimensions Using the Rodeo Han Liu 1,2 John Lafferty 2,3 Larry Wasserman 1,2 1 Statistics Department, 2 Machine Learning Department, 3 Computer Science Department, Carnegie Mellon University
More informationModel-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego
Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships
More informationExercises. (a) Prove that m(t) =
Exercises 1. Lack of memory. Verify that the exponential distribution has the lack of memory property, that is, if T is exponentially distributed with parameter λ > then so is T t given that T > t for
More informationStatistical Inference
Statistical Inference Liu Yang Florida State University October 27, 2016 Liu Yang, Libo Wang (Florida State University) Statistical Inference October 27, 2016 1 / 27 Outline The Bayesian Lasso Trevor Park
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationCramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics. Wen-Xin Zhou
Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics Wen-Xin Zhou Department of Mathematics and Statistics University of Melbourne Joint work with Prof. Qi-Man
More informationLawrence D. Brown* and Daniel McCarthy*
Comments on the paper, An adaptive resampling test for detecting the presence of significant predictors by I. W. McKeague and M. Qian Lawrence D. Brown* and Daniel McCarthy* ABSTRACT: This commentary deals
More informationThe assumptions are needed to give us... valid standard errors valid confidence intervals valid hypothesis tests and p-values
Statistical Consulting Topics The Bootstrap... The bootstrap is a computer-based method for assigning measures of accuracy to statistical estimates. (Efron and Tibshrani, 1998.) What do we do when our
More informationThe Bootstrap: Theory and Applications. Biing-Shen Kuo National Chengchi University
The Bootstrap: Theory and Applications Biing-Shen Kuo National Chengchi University Motivation: Poor Asymptotic Approximation Most of statistical inference relies on asymptotic theory. Motivation: Poor
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationRegression #5: Confidence Intervals and Hypothesis Testing (Part 1)
Regression #5: Confidence Intervals and Hypothesis Testing (Part 1) Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #5 1 / 24 Introduction What is a confidence interval? To fix ideas, suppose
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More information5601 Notes: The Subsampling Bootstrap
5601 Notes: The Subsampling Bootstrap Charles J. Geyer April 13, 2006 1 Web Pages This handout accompanies the web pages http://www.stat.umn.edu/geyer/5601/examp/subboot.html http://www.stat.umn.edu/geyer/5601/examp/subtoot.html
More informationSimple linear regression
Simple linear regression Thomas Lumley BIOST 578C Linear model Linear regression is usually presented in terms of a model Y = α + βx + ɛ ɛ N(0, σ 2 ) because the theoretical analysis is pretty for this
More informationEcon 2120: Section 2
Econ 2120: Section 2 Part I - Linear Predictor Loose Ends Ashesh Rambachan Fall 2018 Outline Big Picture Matrix Version of the Linear Predictor and Least Squares Fit Linear Predictor Least Squares Omitted
More informationNonparametric hypothesis tests and permutation tests
Nonparametric hypothesis tests and permutation tests 1.7 & 2.3. Probability Generating Functions 3.8.3. Wilcoxon Signed Rank Test 3.8.2. Mann-Whitney Test Prof. Tesler Math 283 Fall 2018 Prof. Tesler Wilcoxon
More informationLARGE SAMPLE BEHAVIOR OF SOME WELL-KNOWN ROBUST ESTIMATORS UNDER LONG-RANGE DEPENDENCE
LARGE SAMPLE BEHAVIOR OF SOME WELL-KNOWN ROBUST ESTIMATORS UNDER LONG-RANGE DEPENDENCE C. LÉVY-LEDUC, H. BOISTARD, E. MOULINES, M. S. TAQQU, AND V. A. REISEN Abstract. The paper concerns robust location
More informationSimulating Uniform- and Triangular- Based Double Power Method Distributions
Journal of Statistical and Econometric Methods, vol.6, no.1, 2017, 1-44 ISSN: 1792-6602 (print), 1792-6939 (online) Scienpress Ltd, 2017 Simulating Uniform- and Triangular- Based Double Power Method Distributions
More informationStatistics 3657 : Moment Approximations
Statistics 3657 : Moment Approximations Preliminaries Suppose that we have a r.v. and that we wish to calculate the expectation of g) for some function g. Of course we could calculate it as Eg)) by the
More informationWhat to do today (Nov 22, 2018)?
What to do today (Nov 22, 2018)? Part 1. Introduction and Review (Chp 1-5) Part 2. Basic Statistical Inference (Chp 6-9) Part 3. Important Topics in Statistics (Chp 10-13) Part 4. Further Topics (Selected
More informationMonte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan
Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More information3 Joint Distributions 71
2.2.3 The Normal Distribution 54 2.2.4 The Beta Density 58 2.3 Functions of a Random Variable 58 2.4 Concluding Remarks 64 2.5 Problems 64 3 Joint Distributions 71 3.1 Introduction 71 3.2 Discrete Random
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationWeek 9 The Central Limit Theorem and Estimation Concepts
Week 9 and Estimation Concepts Week 9 and Estimation Concepts Week 9 Objectives 1 The Law of Large Numbers and the concept of consistency of averages are introduced. The condition of existence of the population
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationHT Introduction. P(X i = x i ) = e λ λ x i
MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework
More informationPeter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8
Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall
More informationBootstrap Confidence Intervals
Bootstrap Confidence Intervals Patrick Breheny September 18 Patrick Breheny STA 621: Nonparametric Statistics 1/22 Introduction Bootstrap confidence intervals So far, we have discussed the idea behind
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationL11 Hypothesis testing
Hypothesis Simple Composite Hypothesis Simple Composite Monte-Carlo and Empirical Methods for Statistical Inference, FMS091/MASM11 L11 Hypothesis testing Hypothesis testing Statistical hypothesises Testing
More informationOutline of GLMs. Definitions
Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density
More informationQuantitative Empirical Methods Exam
Quantitative Empirical Methods Exam Yale Department of Political Science, August 2016 You have seven hours to complete the exam. This exam consists of three parts. Back up your assertions with mathematics
More informationSome Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2
STA 248 H1S MIDTERM TEST February 26, 2008 SURNAME: SOLUTIONS GIVEN NAME: STUDENT NUMBER: INSTRUCTIONS: Time: 1 hour and 50 minutes Aids allowed: calculator Tables of the standard normal, t and chi-square
More informationLinear models. Linear models are computationally convenient and remain widely used in. applied econometric research
Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y
More informationInference For High Dimensional M-estimates: Fixed Design Results
Inference For High Dimensional M-estimates: Fixed Design Results Lihua Lei, Peter Bickel and Noureddine El Karoui Department of Statistics, UC Berkeley Berkeley-Stanford Econometrics Jamboree, 2017 1/49
More informationInference Tutorial 2
Inference Tutorial 2 This sheet covers the basics of linear modelling in R, as well as bootstrapping, and the frequentist notion of a confidence interval. When working in R, always create a file containing
More informationPOLI 8501 Introduction to Maximum Likelihood Estimation
POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,
More information