36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1

Size: px
Start display at page:

Download "36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1"

Transcription

1 36. Multisample U-statistics jointly distributed U-statistics Lehmann 6.1 In this topic, we generalize the idea of U-statistics in two different directions. First, we consider single U-statistics for situations in which there is more than one sample. Next, we consider the joint asymptotic distribution of two (single-sample) U-statistics. We begin by generalizing the idea of U-statistics to the case in which we have more than one rom sample. Suppose that X i1,..., X ini is an iid sample from F i for all 1 i s. In other words, we have s rom samples, each potentially from a different distribution, n i is the size of the ith sample. We may define a statistical functional θ = E φ (X 11,..., X 1a1 ; X 21,..., X 2a2 ; ; X s1,..., X sas ). (132) Notice that the kernel φ in Equation (132) has a 1 +a 2 + +a s arguments; furthermore, we assume that the first a 1 of them may be permuted without changing the value of φ, the next a 2 of them may be permuted without changing the value of φ, etc. In other words, there are s distinct blocks of arguments of φ, φ is symmetric in its arguments within each of these blocks. Not surprisingly, the U-statistic corresponding to the functional (132) is U N = ( 1 1 n1 ) ( ns ) a 1 a s 1 i 1< <i a1 n 1 1 r 1< <r as n s Note that N denotes the total of all the sample sizes: N = n n s. φ ( X 1i1,..., X 1ia1 ; ; X sr1,..., X sras ). (133) As we did in the case of single-sample U-statistics, define for 0 j 1 a i,..., 0 j s a s φ j1 j s (X 11,..., X 1j1 ; ; X s1,..., X sjs ) = E {φ(x 11,..., X 1a1 ; ; X s1,..., X sas ) X 11,..., X 1j1,, X s1,..., X sjs } (134) σ 2 j1 j s = Var φ j1 j s (X 11,..., X 1j1 ; ; X s1,..., X sjs ). (135) By an argument similar to the one used in the proof of Theorem 35.1, but much more tedious notationally, we can show that σ 2 j 1 j s = Cov {φ(x 11,..., X 1a1 ; ; X s1,..., X sas ), φ(x 11,..., X 1j1, X 1,a1+1,..., X 1,a1+(a 1 j 1); ; X s1,..., X sjs, X s,as+1,..., X s,as+(a s j s)) }. (136) Notice that some of the j i may equal 0. This was not true in the single-sample case, since φ 0 would have merely been the constant θ, so σ 2 0 = 0. In the special case when s = 2, Equations (134), (135) (136) become φ ij (X 1,..., X i ; Y 1,..., Y j ) = E {φ(x 1,..., X a1 ; Y 1,..., Y a2 ) X 1,..., X i, Y 1,..., Y j }, σ 2 ij = Var φ ij (X 1,..., X i ; Y 1..., Y j ), σ 2 ij = Cov {φ(x 1,..., X a1 ; Y 1,..., Y a2 ), φ(x 1,..., X i, X a1+1,..., X a1+(a 1 i); Y 1,..., Y j, Y a2+1,..., Y a2+(a 2 j)) }, 104

2 respectively, for 0 i a 1 0 j a 2. Although we will not derive it here as we did for the single-sample case, there is an analagous asymptotic normality result for multisample U-statistics, as follows. Theorem 36.1 Suppose that for i = 1,..., s, X i1,..., X ini is a rom sample from the distribution F i that these s samples are independent of each other. Suppose further that there exist constants ρ 1,..., ρ s in the interval (0, 1) such that n i /N ρ i for all i that σ 2 a 1 a s <. Then N(UN θ) L N(0, σ 2 ), where σ 2 = a2 1 ρ 1 σ a2 s ρ s σ Although the notation required for the multisample U-statistic theory is nightmarish, life becomes considerably simpler in the case s = 2 a 1 = a 2 = 1, in which case we obtain U N = 1 n 1 n 2 n 1 n 2 φ(x 1i ; X 2j ). j=1 Equivalently, we may assume that X 1,..., X m are an iid sample from F Y 1,..., Y n are an iid sample from G, which gives U N = 1 mn m j=1 In the case of the U-statistic of Equation (137), Theorem 36.1 states that ( ) N(UN θ) L N 0, σ2 10 ρ + σ2 01, 1 ρ n φ(x i ; Y j ). (137) where ρ = lim m/n, σ 2 10 = Cov {φ(x 1 ; Y 1 ), φ(x 1 ; Y 2 )}, σ 2 01 = Cov {φ(x 1 ; Y 1 ), φ(x 2 ; Y 1 )}. Example 36.1 For X 1,... X m iid from F Y 1,..., Y n iid from G, consider the Wilcoxon ranksum statistic W, defined to be the sum of the ranks of the Y i among the combined sample. We showed in Equation (79) that W = 1 2 n(n + 1) + m j=1 n I{X i < Y j }. Therefore, if we let φ(x; y) = I{x < y}, then the corresponding two-sample U-statistic U N is related to W by W = 1 2 n(n + 1) + mnu N. Therefore, we may use Theorem 36.1 to obtain the asymptotic normality of U N, therefore of W, just as we did in Topic 26. However, unlike in Topic 26, we make no assumption here that F G are merely shifted versions of one another. Thus, we may now obtain in principle the asymptotic distribution of the rank-sum statistic for any two distributions F G that we wish, so long as they have finite second moments. The other direction in which we will generalize the development of U-statistics is consideration of the joint distribution of two single-sample U-statistics. Suppose that there are two kernel functions, φ(x 1,..., x a ) ϕ(x 1,..., x b ), we define the two corresponding U-statistics U (1) n = 1 ( n a) 1 i 1< <i a n φ(x i1,..., X ia ) 105

3 U (2) n = 1 ( n b) 1 j 1< <j b n φ(x j1,..., X ja ) for a single rom sample X 1,..., X n from F. Define θ 1 = E U n (1) θ 2 = E U n (2). Furthermore, define γ ij to be the covariance between φ i (X 1,..., X i ) ϕ j (X 1,..., X j ), where φ i ϕ j are defined as in Equation (126). It may be proved that γ ij = Cov { φ(x 1,..., X a ), ϕ(x 1,..., X j, X a+1,..., X a+(b j) ) }. (138) Note in particular that γ ij depends only on the value of min{i, j}. The following theorem, stated without proof, gives the joint asymptotic distribution of U (1) n U (2) n. Theorem 36.2 Suppose X 1,..., X n is a rom sample from F that φ ϕ are two kernel functions satisfying Var φ(x 1,..., X a ) < Var ϕ(x 1,..., X b ) <. Define τ1 2 = Var φ 1 (X 1 ) τ2 2 = Var ϕ 1 (X 1 ), let γ ij be defined as in Equation (138). Then { (U (1)) n n U n (2) ( θ1 θ 2 ) } {( ) L 0 N, 0 ( a 2 τ 2 1 abγ 11 abγ 11 b 2 τ 2 2 )}. Problems Problem 36.1 Suppose X 1,..., X m Y 1,..., Y n are independent iid samples from distributions Unif(0, θ) Unif(µ, µ + θ), respectively. Assume m/n ρ as m, n 0 < µ < θ. (a) Find the asymptotic distribution of the U-statistic of Equation (137), where φ(x; y) = I{x < y}. In so doing, find a function g(x) such that E (U N ) = g(µ). (b) Find the asymptotic distribution of g(y X). (c) Find the range of values of µ for which the Wilcoxon estimate of g(µ) is asymptotically more efficient than g(y X). (The ARE in this case is the ratio of asymptotic variances. See the discussion of (4.3.12) (4.3.13) on p. 240 for details.) Problem 36.2 Solve each part of Problem 36.1, but this time under the assumptions that the independent rom samples X 1,..., X m Y 1,..., Y n satisfy P (X 1 t) = P (Y 1 θ t) = t 2 for t [0, 1] 0 < θ < 1. As in Problem 36.1, assume m/n ρ (0, 1). Problem 36.3 Suppose X 1,..., X m Y 1,..., Y n are independent iid samples from distributions N(0, 1) N(µ, 1), respectively. Assume m/(m + n) 1/2 as m, n. Let U N be the U-statistic of Equation (137), where φ(x; y) = I{x < y}. Suppose that θ(µ) σ 2 (µ) are such that N[UN θ(µ)] L N[0, σ 2 (µ)]. Calculate θ(µ) σ 2 (µ) for µ {.2,.5, 1, 1.5, 2}. Hint: This problem requires a bit of numerical integration. There are a couple of ways you might do this. Mathematica will do it quite easily. There is a function called integrate in Splus one called quad in MATLAB for integrating a function. If you cannot get any of these to work for you, let me know. 106

4 Problem 36.4 Suppose X 1, X 2,... are iid with finite variance. Define S 2 n = 1 n 1 n (x i x) 2 let G n be Gini s mean difference, the U-statistic defined in Problem Note that S 2 n is also a U-statistic, corresponding to the kernel function φ(x 1, x 2 ) = (x 1 x 2 ) 2. (a) If X i are distributed as Unif(0, θ), give the joint asymptotic distribution of G n S n by first finding the joint asymptotic distribution of the U-statistics G n Sn. 2 Note that the covariance matrix need not be positive definite; in this problem, the covariance matrix is singular. (b) The singular asymptotic covariance matrix in this problem implies that as n, the joint distribution of G n S n becomes concentrated on a line. Does this appear to be the case? For 1000 samples of size n from Uniform(0, 1), plot scatterplots of G n against S n. Take n {5, 25, 100}. 107

5 37. Bootstrapping Lehmann 6.5 In this topic, we introduce the bootstrap, even though the asymptotics content of what follows is not very high. Essentially, the only large-sample theory we rely on here is the weak law of large numbers. Nonetheless, bootstrapping is included here because of its natural relationship with the concepts of statistical functionals plug-in estimators seen in recent topics, also because it is an increasingly popular often misunderstood method. Consider a statistical functional λ n (F ) that depends on n. For instance, λ n (F ) may be some property, such as bias or variance, of an estimator ˆθ n of θ = θ(f ) based on a rom sample of size n from some distribution F. As an example, let θ(f ) = F 1 ( 1 2) be the median of F. Take ˆθn to be the mth order statistic from a rom sample of size n = 2m 1 from F. Consider the bias λ B n (F ) = E F ˆθn θ(f ) the variance λ V n (F ) = E F ˆθ2 n (E F ˆθn ) 2. Theoretical properties of λ B n λ V n are very difficult to obtain. Even asymptotics aren t very helpful, since n(ˆθ n θ) L N{0, 1/(4f 2 (θ))} tells us only that the bias goes to zero the limiting variance may be very hard to estimate because it involves the unknown quantity f(θ), which is hard to estimate. Consider the plug-in estimators λ B n ( ˆF n ) λ V n ( ˆF n ). (Recall that ˆF n denotes the empirical distribution function, which puts a mass of 1 n on each of the n sample points.) In our median example, λ B n ( ˆF n ) = E ˆFn ˆθ n ˆθ n λ V n ( ˆF n ) = E ˆFn (ˆθ n) 2 (E ˆFn ˆθ n ) 2, where ˆθ n is the sample median from a rom sample X 1,..., X n from ˆF n. To see how difficult it is to calculate λ B n ( ˆF n ) λ V n ( ˆF n ), consider the simplest nontrivial case, n = 3: Conditional on the order statistics (X (1), X (2), X (3) ), there are 27 equally likely possibilities for the value of (X 1, X 2, X 3 ), the sample of size 3 from ˆF n, namely (X (1), X (1), X (1) ), (X (1), X (1), X (2) ),..., (X (3), X (3), X (3) ). Of these 27 possibilities, exactly = 7 have the value X (1) occurring 2 or 3 times. Therefore, we obtain This implies that P (ˆθ n = X (1) ) = 7 27, P (ˆθ n = X (2) ) = 13 27, P (ˆθ n = X (3) ) = E ˆFn ˆθ n = 1 27 (7X (1) + 13X (2) + 7X (3) ) E ˆFn (ˆθ n) 2 = 1 27 (7X2 (1) + 13X2 (2) + 7X2 (3) ). Therefore, since ˆθ n = X (2), we obtain λ B n ( ˆF n ) = 1 27 (7X (1) 14X (2) + 7X (3) ) λ V n ( ˆF n ) = (10X2 (1) + 13X2 (2) + 10X2 (3) 13X (1)X (2) 13X (2) X (3) 7X (1) X (3) ). 108

6 To obtain the sampling distribution of these estimators, of course, we would have to consider the joint distribution of (X (1), X (2), X (3) ). Naturally, the calculations become even more difficult as n increases. Alternatively, we could use resampling in order to approximate λ B n ( ˆF n ) λ V n ( ˆF n ). This is the bootstrapping idea, it works like this: For some large number B, simulate B rom samples from ˆF n, namely X 11,..., X 1n,. X B1,..., X Bn, approximate a quantity like E ˆFn ˆθ n by the sample mean 1 B B ˆθ in, where ˆθ in is the sample median of the ith bootstrap sample X i1,..., X in. Notice that the weak law of large numbers asserts that 1 B B ˆθ in P E ˆFn ˆθ n. To recap, then, we wish to estimate some parameter λ n (F ) for an unknown distribution F based on a rom sample from F. We estimate λ n (F ) by λ n ( ˆF n ), but it is not easy to evaluate λ n ( ˆF n ) so we approximate λ n ( ˆF n ) by resampling B times from ˆF n obtain a bootstrap estimator λ B,n. Thus, there are two relevant issues: 1. How good is the approximation of λ n ( ˆF n ) by λ B,n? (Note that λ n( ˆF n ) is NOT an unknown parameter; it is known but hard to evaluate.) 2. How precise is the estimation of λ n (F ) by λ n ( ˆF n )? Question 1 is usually easy to answer using asymptotics; these asymptotics involve letting B usually rely on the weak law or the central limit theorem. For example, if we have an expectation functional λ n (F ) = E F h(x 1,..., X n ), then as B. λ B,n = 1 B B h(xi1,..., Xin) P λ n ( ˆF n ) Question 2, on the other h, is often tricky; asymptotic results involve letting n are hled case-by-case. We will not discuss these asymptotics here. On a related note, however, there is an argument in Lehmann s book (on pages ) about why a plug-in estimator may be better than an asymptotic estimator. That is, if it is possible to show λ n (F ) λ as n, then as an estimator of λ n (F ), λ n ( ˆF n ) may be preferable to λ. We conclude this topic by considering the so-called parametric bootstrap. If we assume that the unknown distribution function F comes from a family of distribution functions indexed by a parameter µ, then λ n (F ) is really λ n (F µ ). Then, instead of the plug-in estimator λ n ( ˆF n ), we might consider the estimator λ n (Fˆµ ), where ˆµ is an estimator of µ. Everything proceeds as in the nonparametric version of bootstrapping. Since it may not be easy to evaulate λ n (Fˆµ ) explicitly, we first find ˆµ then take B rom samples of size n, X11,..., X1n through XB1,..., X Bn, from Fˆµ. These samples are used to approximate λ n (Fˆµ ). 109

7 Example 37.1 Suppose X 1,..., X n is a rom sample from Poisson(µ). Take ˆµ = X. Suppose λ n (F µ ) = Var Fµ ˆµ. In this case, we happen to know that λ n (F µ ) = µ/n, but let s ignore this knowledge apply a parametric bootstrap. For some large B, say 500, generate B samples from Poisson(ˆµ) use the sample variance of ˆµ as an approximation to λ n (Fˆµ ). In R, with µ = 1 n = 20 we obtain > x <- rpois(20,1) # Generate the sample from F > muhat <- mean(x) > muhat [1] 0.85 > muhatstar <- rep(0,500) # Allocate the vector for muhatstar > for(i in 1:500) muhatstar[i] <- mean(rpois(20,muhat)) > var(muhatstar) [1] Note that the estimate is close to the known true value Obviously, this example is simplistic because we already know that λ n (F ) = µ/n, which makes ˆµ/n a more natural estimator. However, it is not always so simple to obtain a closed-form expression for λ n (F ). Incidentally, we could also use a nonparametric bootstrap approach in this example: > for (i in 1:500) muhatstar2[i] <- mean(sample(x,replace=t)) > var(muhatstar2) [1] Of course, is an approximation to λ n ( ˆF n ) rather than λ n (Fˆµ ). Furthermore, we can obtain a result arbitrarily close to λ n ( ˆF n ) by increasing the value of B: > muhatstar2_rep(0,100000) > for (i in 1:100000) muhatstar2[i] <- mean(sample(x,replace=t)) > var(muhatstar2) [1] In fact, it is in principle possible to obtain an approximate variance for our estimates of λ n ( ˆF n ) λ n (Fˆµ ),, using the central limit theorem, construct approximate confidence intervals for these quantities. This would allow us to specify the quantities to any desired level of accuracy. Problems Problem 37.1 (a) Devise a nonparametric bootstrap scheme for setting confidence intervals for β in the linear regression model Y i = α + βx i + ɛ i. There is more than one possible answer. (b) Using B = 1000, implement your scheme on the following dataset to obtain a 95% confidence interval. Compare your answer with the stard 95% confidence interval. Y x (In the dataset, Y is the number of manatee deaths due to collisions with powerboats in Florida x is the number of powerboat registrations in thouss for even years from ) Problem 37.2 Consider the following dataset that lists the latitude mean August temperature for 7 US cities. The residuals are listed for use in part (b). 110

8 City Miami Phoenix Memphis Baltimore Pittsburgh Boston Portl, OR Latitude Temperature Residual Minitab gives the following output for a simple linear regression: Predictor Coef SE Coef T P Constant latitude S = R-Sq = 61.5% R-Sq(adj) = 53.8% Note that this gives an asymptotic estimate of the variance of the slope parameter as = In each case below, use the described method to simulate B = 500 bootstrap samples (x b1, y b1 ),..., (x b7, y b7 ) for b = 1,..., B. For each b, refit the model to obtain ˆβ b. Report the sample variance of ˆβ 1,..., ˆβ B compare with the asymptotic estimate of (a) Parametric bootstrap. Take x bi = x i for all b i. Let ybi = ˆβ 0 + ˆβ 1 x i + ɛ i, where ɛ i N(0, ˆσ 2 ). Obtain ˆβ 0, ˆβ 1, ˆσ 2 from the above output. (b) Nonparametric bootstrap I. Take x bi = x i for all b i. Let ybi = ˆβ 0 + ˆβ 1 x i + rbi, where rb1,..., r b7 is an iid sample from the empirical distribution of the residuals from the original model (you may want to refit the original model to find these residuals). (c) Nonparametric bootstrap II. Let (x b1, y b1 ),..., (x b7, y b7 ) be an iid sample from the empirical distribution of (x 1, y 1 ),..., (x 7, y 7 ). Note: In Splus, you can obtain the slope coefficient of the linear regression of the vector y on the vector x using lm(y~x)$coef[2]. Problem 37.3 The same resampling idea that is exploited in the bootstrap can be used to approximate the value of difficult integrals by a technique sometimes called Monte Carlo integration. Suppose we wish to compute θ = e x2 cos 3 (x) dx. Using the NIntegrate function in Mathematica, we find that θ = (a) Define g(t) = 2e t2 cos 3 (t). Let U 1,..., U n be an iid uniform(0,1) sample. Let Prove that ˆθ 1 P θ. ˆθ 1 = 1 n n g(u i ). (b) Define h(t) = 2 2t. Prove that if we take V i = 1 U i for each i, then V i is a rom variable with density h(t). Prove that with we have ˆθ 2 P θ. ˆθ 2 = 1 n n g(v i ) h(v i ), 111

9 (c) For n = 1000, simulate ˆθ 1 ˆθ 2. Give estimates of the variance for each estimator by reporting ˆσ 2 /n for each, where ˆσ 2 is the sample variance of the g(u i ) or the g(v i )/h(v i ) as the case may be. (d) Plot, on the same set of axes, g(t), h(t), the stard uniform density for t [0, 1]. From this plot, explain why the variance of ˆθ 2 is smaller than the variance of ˆθ 1. [Incidentally, the technique of drawing rom variables from a density h whose shape is close to the function g of interest is a variance-reduction technique known as importance sampling.] Note: This was sort of a silly example, since it was really easy to get an exact value for θ using numerical methods. However, with certain high-dimensional integrals, the curse of dimensionality makes exact numerical methods extremely time-consuming computationally; thus, Monte Carlo integration does have a practical use in such cases. 112

Chapter 10. U-statistics Statistical Functionals and V-Statistics

Chapter 10. U-statistics Statistical Functionals and V-Statistics Chapter 10 U-statistics When one is willing to assume the existence of a simple random sample X 1,..., X n, U- statistics generalize common notions of unbiased estimation such as the sample mean and the

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Advanced Statistics II: Non Parametric Tests

Advanced Statistics II: Non Parametric Tests Advanced Statistics II: Non Parametric Tests Aurélien Garivier ParisTech February 27, 2011 Outline Fitting a distribution Rank Tests for the comparison of two samples Two unrelated samples: Mann-Whitney

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Output Analysis for Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Output Analysis

More information

Chapter 2: Resampling Maarten Jansen

Chapter 2: Resampling Maarten Jansen Chapter 2: Resampling Maarten Jansen Randomization tests Randomized experiment random assignment of sample subjects to groups Example: medical experiment with control group n 1 subjects for true medicine,

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

Bootstrap tests. Patrick Breheny. October 11. Bootstrap vs. permutation tests Testing for equality of location

Bootstrap tests. Patrick Breheny. October 11. Bootstrap vs. permutation tests Testing for equality of location Bootstrap tests Patrick Breheny October 11 Patrick Breheny STA 621: Nonparametric Statistics 1/14 Introduction Conditioning on the observed data to obtain permutation tests is certainly an important idea

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 1 Bootstrapped Bias and CIs Given a multiple regression model with mean and

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

The bootstrap. Patrick Breheny. December 6. The empirical distribution function The bootstrap

The bootstrap. Patrick Breheny. December 6. The empirical distribution function The bootstrap Patrick Breheny December 6 Patrick Breheny BST 764: Applied Statistical Modeling 1/21 The empirical distribution function Suppose X F, where F (x) = Pr(X x) is a distribution function, and we wish to estimate

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend

More information

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf Lecture 13: 2011 Bootstrap ) R n x n, θ P)) = τ n ˆθn θ P) Example: ˆθn = X n, τ n = n, θ = EX = µ P) ˆθ = min X n, τ n = n, θ P) = sup{x : F x) 0} ) Define: J n P), the distribution of τ n ˆθ n θ P) under

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Problem Selected Scores

Problem Selected Scores Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected

More information

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:. MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss

More information

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods Chapter 4 Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods 4.1 Introduction It is now explicable that ridge regression estimator (here we take ordinary ridge estimator (ORE)

More information

The Nonparametric Bootstrap

The Nonparametric Bootstrap The Nonparametric Bootstrap The nonparametric bootstrap may involve inferences about a parameter, but we use a nonparametric procedure in approximating the parametric distribution using the ECDF. We use

More information

Bootstrap & Confidence/Prediction intervals

Bootstrap & Confidence/Prediction intervals Bootstrap & Confidence/Prediction intervals Olivier Roustant Mines Saint-Étienne 2017/11 Olivier Roustant (EMSE) Bootstrap & Confidence/Prediction intervals 2017/11 1 / 9 Framework Consider a model with

More information

University of California San Diego and Stanford University and

University of California San Diego and Stanford University and First International Workshop on Functional and Operatorial Statistics. Toulouse, June 19-21, 2008 K-sample Subsampling Dimitris N. olitis andjoseph.romano University of California San Diego and Stanford

More information

STAT 830 Non-parametric Inference Basics

STAT 830 Non-parametric Inference Basics STAT 830 Non-parametric Inference Basics Richard Lockhart Simon Fraser University STAT 801=830 Fall 2012 Richard Lockhart (Simon Fraser University)STAT 830 Non-parametric Inference Basics STAT 801=830

More information

Bootstrap. Director of Center for Astrostatistics. G. Jogesh Babu. Penn State University babu.

Bootstrap. Director of Center for Astrostatistics. G. Jogesh Babu. Penn State University  babu. Bootstrap G. Jogesh Babu Penn State University http://www.stat.psu.edu/ babu Director of Center for Astrostatistics http://astrostatistics.psu.edu Outline 1 Motivation 2 Simple statistical problem 3 Resampling

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Bootstrap for Regression Week 9, Lecture 1

MA 575 Linear Models: Cedric E. Ginestet, Boston University Bootstrap for Regression Week 9, Lecture 1 MA 575 Linear Models: Cedric E. Ginestet, Boston University Bootstrap for Regression Week 9, Lecture 1 1 The General Bootstrap This is a computer-intensive resampling algorithm for estimating the empirical

More information

Properties of the least squares estimates

Properties of the least squares estimates Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares

More information

STAT 730 Chapter 4: Estimation

STAT 730 Chapter 4: Estimation STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Chapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics

Chapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics Chapter 6 Order Statistics and Quantiles 61 Extreme Order Statistics Suppose we have a finite sample X 1,, X n Conditional on this sample, we define the values X 1),, X n) to be a permutation of X 1,,

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Bootstrap, Jackknife and other resampling methods

Bootstrap, Jackknife and other resampling methods Bootstrap, Jackknife and other resampling methods Part III: Parametric Bootstrap Rozenn Dahyot Room 128, Department of Statistics Trinity College Dublin, Ireland dahyot@mee.tcd.ie 2005 R. Dahyot (TCD)

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables. Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just

More information

Lecture 1: Random number generation, permutation test, and the bootstrap. August 25, 2016

Lecture 1: Random number generation, permutation test, and the bootstrap. August 25, 2016 Lecture 1: Random number generation, permutation test, and the bootstrap August 25, 2016 Statistical simulation 1/21 Statistical simulation (Monte Carlo) is an important part of statistical method research.

More information

One-Sample Numerical Data

One-Sample Numerical Data One-Sample Numerical Data quantiles, boxplot, histogram, bootstrap confidence intervals, goodness-of-fit tests University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Casuality and Programme Evaluation

Casuality and Programme Evaluation Casuality and Programme Evaluation Lecture V: Difference-in-Differences II Dr Martin Karlsson University of Duisburg-Essen Summer Semester 2017 M Karlsson (University of Duisburg-Essen) Casuality and Programme

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

Final Examination Statistics 200C. T. Ferguson June 11, 2009

Final Examination Statistics 200C. T. Ferguson June 11, 2009 Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean

More information

On robust and efficient estimation of the center of. Symmetry.

On robust and efficient estimation of the center of. Symmetry. On robust and efficient estimation of the center of symmetry Howard D. Bondell Department of Statistics, North Carolina State University Raleigh, NC 27695-8203, U.S.A (email: bondell@stat.ncsu.edu) Abstract

More information

STAT 730 Chapter 5: Hypothesis Testing

STAT 730 Chapter 5: Hypothesis Testing STAT 730 Chapter 5: Hypothesis Testing Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 28 Likelihood ratio test def n: Data X depend on θ. The

More information

5601 Notes: The Sandwich Estimator

5601 Notes: The Sandwich Estimator 560 Notes: The Sandwich Estimator Charles J. Geyer December 6, 2003 Contents Maximum Likelihood Estimation 2. Likelihood for One Observation................... 2.2 Likelihood for Many IID Observations...............

More information

STAT440/840: Statistical Computing

STAT440/840: Statistical Computing First Prev Next Last STAT440/840: Statistical Computing Paul Marriott pmarriott@math.uwaterloo.ca MC 6096 February 2, 2005 Page 1 of 41 First Prev Next Last Page 2 of 41 Chapter 3: Data resampling: the

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,

More information

0.1 The plug-in principle for finding estimators

0.1 The plug-in principle for finding estimators The Bootstrap.1 The plug-in principle for finding estimators Under a parametric model P = {P θ ;θ Θ} (or a non-parametric P = {P F ;F F}), any real-valued characteristic τ of a particular member P θ (or

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research

More information

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22 MLE and GMM Li Zhao, SJTU Spring, 2017 Li Zhao MLE and GMM 1 / 22 Outline 1 MLE 2 GMM 3 Binary Choice Models Li Zhao MLE and GMM 2 / 22 Maximum Likelihood Estimation - Introduction For a linear model y

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Inference based on robust estimators Part 2

Inference based on robust estimators Part 2 Inference based on robust estimators Part 2 Matias Salibian-Barrera 1 Department of Statistics University of British Columbia ECARES - Dec 2007 Matias Salibian-Barrera (UBC) Robust inference (2) ECARES

More information

Contents 1. Contents

Contents 1. Contents Contents 1 Contents 1 One-Sample Methods 3 1.1 Parametric Methods.................... 4 1.1.1 One-sample Z-test (see Chapter 0.3.1)...... 4 1.1.2 One-sample t-test................. 6 1.1.3 Large sample

More information

Non-Parametric Bootstrap Mean. Squared Error Estimation For M- Quantile Estimators Of Small Area. Averages, Quantiles And Poverty

Non-Parametric Bootstrap Mean. Squared Error Estimation For M- Quantile Estimators Of Small Area. Averages, Quantiles And Poverty Working Paper M11/02 Methodology Non-Parametric Bootstrap Mean Squared Error Estimation For M- Quantile Estimators Of Small Area Averages, Quantiles And Poverty Indicators Stefano Marchetti, Nikos Tzavidis,

More information

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

Cox s proportional hazards/regression model - model assessment

Cox s proportional hazards/regression model - model assessment Cox s proportional hazards/regression model - model assessment Rasmus Waagepetersen September 27, 2017 Topics: Plots based on estimated cumulative hazards Cox-Snell residuals: overall check of fit Martingale

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

Better Bootstrap Confidence Intervals

Better Bootstrap Confidence Intervals by Bradley Efron University of Washington, Department of Statistics April 12, 2012 An example Suppose we wish to make inference on some parameter θ T (F ) (e.g. θ = E F X ), based on data We might suppose

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo Outline in High Dimensions Using the Rodeo Han Liu 1,2 John Lafferty 2,3 Larry Wasserman 1,2 1 Statistics Department, 2 Machine Learning Department, 3 Computer Science Department, Carnegie Mellon University

More information

Model-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego

Model-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships

More information

Exercises. (a) Prove that m(t) =

Exercises. (a) Prove that m(t) = Exercises 1. Lack of memory. Verify that the exponential distribution has the lack of memory property, that is, if T is exponentially distributed with parameter λ > then so is T t given that T > t for

More information

Statistical Inference

Statistical Inference Statistical Inference Liu Yang Florida State University October 27, 2016 Liu Yang, Libo Wang (Florida State University) Statistical Inference October 27, 2016 1 / 27 Outline The Bayesian Lasso Trevor Park

More information

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)

More information

Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics. Wen-Xin Zhou

Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics. Wen-Xin Zhou Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics Wen-Xin Zhou Department of Mathematics and Statistics University of Melbourne Joint work with Prof. Qi-Man

More information

Lawrence D. Brown* and Daniel McCarthy*

Lawrence D. Brown* and Daniel McCarthy* Comments on the paper, An adaptive resampling test for detecting the presence of significant predictors by I. W. McKeague and M. Qian Lawrence D. Brown* and Daniel McCarthy* ABSTRACT: This commentary deals

More information

The assumptions are needed to give us... valid standard errors valid confidence intervals valid hypothesis tests and p-values

The assumptions are needed to give us... valid standard errors valid confidence intervals valid hypothesis tests and p-values Statistical Consulting Topics The Bootstrap... The bootstrap is a computer-based method for assigning measures of accuracy to statistical estimates. (Efron and Tibshrani, 1998.) What do we do when our

More information

The Bootstrap: Theory and Applications. Biing-Shen Kuo National Chengchi University

The Bootstrap: Theory and Applications. Biing-Shen Kuo National Chengchi University The Bootstrap: Theory and Applications Biing-Shen Kuo National Chengchi University Motivation: Poor Asymptotic Approximation Most of statistical inference relies on asymptotic theory. Motivation: Poor

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1)

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1) Regression #5: Confidence Intervals and Hypothesis Testing (Part 1) Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #5 1 / 24 Introduction What is a confidence interval? To fix ideas, suppose

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

5601 Notes: The Subsampling Bootstrap

5601 Notes: The Subsampling Bootstrap 5601 Notes: The Subsampling Bootstrap Charles J. Geyer April 13, 2006 1 Web Pages This handout accompanies the web pages http://www.stat.umn.edu/geyer/5601/examp/subboot.html http://www.stat.umn.edu/geyer/5601/examp/subtoot.html

More information

Simple linear regression

Simple linear regression Simple linear regression Thomas Lumley BIOST 578C Linear model Linear regression is usually presented in terms of a model Y = α + βx + ɛ ɛ N(0, σ 2 ) because the theoretical analysis is pretty for this

More information

Econ 2120: Section 2

Econ 2120: Section 2 Econ 2120: Section 2 Part I - Linear Predictor Loose Ends Ashesh Rambachan Fall 2018 Outline Big Picture Matrix Version of the Linear Predictor and Least Squares Fit Linear Predictor Least Squares Omitted

More information

Nonparametric hypothesis tests and permutation tests

Nonparametric hypothesis tests and permutation tests Nonparametric hypothesis tests and permutation tests 1.7 & 2.3. Probability Generating Functions 3.8.3. Wilcoxon Signed Rank Test 3.8.2. Mann-Whitney Test Prof. Tesler Math 283 Fall 2018 Prof. Tesler Wilcoxon

More information

LARGE SAMPLE BEHAVIOR OF SOME WELL-KNOWN ROBUST ESTIMATORS UNDER LONG-RANGE DEPENDENCE

LARGE SAMPLE BEHAVIOR OF SOME WELL-KNOWN ROBUST ESTIMATORS UNDER LONG-RANGE DEPENDENCE LARGE SAMPLE BEHAVIOR OF SOME WELL-KNOWN ROBUST ESTIMATORS UNDER LONG-RANGE DEPENDENCE C. LÉVY-LEDUC, H. BOISTARD, E. MOULINES, M. S. TAQQU, AND V. A. REISEN Abstract. The paper concerns robust location

More information

Simulating Uniform- and Triangular- Based Double Power Method Distributions

Simulating Uniform- and Triangular- Based Double Power Method Distributions Journal of Statistical and Econometric Methods, vol.6, no.1, 2017, 1-44 ISSN: 1792-6602 (print), 1792-6939 (online) Scienpress Ltd, 2017 Simulating Uniform- and Triangular- Based Double Power Method Distributions

More information

Statistics 3657 : Moment Approximations

Statistics 3657 : Moment Approximations Statistics 3657 : Moment Approximations Preliminaries Suppose that we have a r.v. and that we wish to calculate the expectation of g) for some function g. Of course we could calculate it as Eg)) by the

More information

What to do today (Nov 22, 2018)?

What to do today (Nov 22, 2018)? What to do today (Nov 22, 2018)? Part 1. Introduction and Review (Chp 1-5) Part 2. Basic Statistical Inference (Chp 6-9) Part 3. Important Topics in Statistics (Chp 10-13) Part 4. Further Topics (Selected

More information

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

3 Joint Distributions 71

3 Joint Distributions 71 2.2.3 The Normal Distribution 54 2.2.4 The Beta Density 58 2.3 Functions of a Random Variable 58 2.4 Concluding Remarks 64 2.5 Problems 64 3 Joint Distributions 71 3.1 Introduction 71 3.2 Discrete Random

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Week 9 The Central Limit Theorem and Estimation Concepts

Week 9 The Central Limit Theorem and Estimation Concepts Week 9 and Estimation Concepts Week 9 and Estimation Concepts Week 9 Objectives 1 The Law of Large Numbers and the concept of consistency of averages are introduced. The condition of existence of the population

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

HT Introduction. P(X i = x i ) = e λ λ x i

HT Introduction. P(X i = x i ) = e λ λ x i MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Bootstrap Confidence Intervals

Bootstrap Confidence Intervals Bootstrap Confidence Intervals Patrick Breheny September 18 Patrick Breheny STA 621: Nonparametric Statistics 1/22 Introduction Bootstrap confidence intervals So far, we have discussed the idea behind

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

L11 Hypothesis testing

L11 Hypothesis testing Hypothesis Simple Composite Hypothesis Simple Composite Monte-Carlo and Empirical Methods for Statistical Inference, FMS091/MASM11 L11 Hypothesis testing Hypothesis testing Statistical hypothesises Testing

More information

Outline of GLMs. Definitions

Outline of GLMs. Definitions Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density

More information

Quantitative Empirical Methods Exam

Quantitative Empirical Methods Exam Quantitative Empirical Methods Exam Yale Department of Political Science, August 2016 You have seven hours to complete the exam. This exam consists of three parts. Back up your assertions with mathematics

More information

Some Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2

Some Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2 STA 248 H1S MIDTERM TEST February 26, 2008 SURNAME: SOLUTIONS GIVEN NAME: STUDENT NUMBER: INSTRUCTIONS: Time: 1 hour and 50 minutes Aids allowed: calculator Tables of the standard normal, t and chi-square

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information

Inference For High Dimensional M-estimates: Fixed Design Results

Inference For High Dimensional M-estimates: Fixed Design Results Inference For High Dimensional M-estimates: Fixed Design Results Lihua Lei, Peter Bickel and Noureddine El Karoui Department of Statistics, UC Berkeley Berkeley-Stanford Econometrics Jamboree, 2017 1/49

More information

Inference Tutorial 2

Inference Tutorial 2 Inference Tutorial 2 This sheet covers the basics of linear modelling in R, as well as bootstrapping, and the frequentist notion of a confidence interval. When working in R, always create a file containing

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information