Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n

Size: px
Start display at page:

Download "Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n"

Transcription

1 Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible tests. To this end, let l(θ) denote the loglikelihood and ˆθ n the consistent root of the likelihood equation. Intuitively, the farther θ 0 is from ˆθ n, the stronger the evidence against the null hypothesis. But how far is far enough? Note that if θ 0 is close to ˆθ n, then l(θ 0 ) should also be close to l(ˆθ n ) and l (θ 0 ) should be close to l (ˆθ n ) = 0. If we base a test upon the value of (ˆθ n θ 0 ), we obtain a Wald test. If we base a test upon the value of l(ˆθ n ) l(θ 0 ), we obtain a likelihood ratio test. If we base a test upon the value of l (θ 0 ), we obtain a (Rao) score test. Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : ( ) d 1 n(ˆθn θ 0 ) N 0, ; (9.1) I(θ 0 ) 1 n l (θ 0 ) d N{0, I(θ 0 )}; (9.) 1 n l (θ 0 ) P I(θ 0 ). (9.3) 139

2 Equation (9.) is proved using the central limit theorem; Equation (9.3) is proved using the weak law of large numbers; and Equation (9.1) is the result of a Taylor expansion together with Equations (9.) and (9.3). The three equations above will help us to justify the definitions of Wald, score, and likelihood ratio tests to follow. Equation (9.1) suggests that if we define W n = ni(θ 0 )(ˆθ n θ 0 ), (9.4) then W n, called a Wald statistic, should converge in distribution to standard normal under H 0. Note that this fact remains true if we define W n = nî(ˆθ n θ 0 ), (9.5) where Î is consistent for θ 0; for example, Î could be I(ˆθ n ). Definition 9.1 A Wald test is any test that rejects H 0 : θ = θ 0 in favor of H 1 : θ θ 0 when W n u α/ for W n defined as in Equation (9.4) or Equation (9.5). As usual, u α/ denotes the 1 α quantile of the standard normal distribution. Equation (9.) suggests that if we define R n = 1 ni(θ0 ) l (θ 0 ), (9.6) then R n, called a Rao score statistic or simply a score statistic, converges in distribution to standard normal under H 0. We could also replace I(θ 0 ) by a consistent estimator Î as in Equation (9.5), but usually this is not done: One of the main benefits of the score statistic is that it is not necessary to compute ˆθ n, and using I(ˆθ n ) instead of I(θ 0 ) would defeat this purpose. Definition 9. A score test, sometimes called a Rao score test, is any test that rejects H 0 : θ = θ 0 in favor of H 1 : θ θ 0 when R n u α/ for R n defined as in Equation (9.6). The third type of test, the likelihood ratio test, requires a bit of development. It is a test based on the statistic If we Taylor expand l (ˆθ n ) = 0 around the point θ 0, we obtain n = l(ˆθ n ) l(θ 0 ). (9.7) l (θ 0 ) = (ˆθ n θ 0 ) { l (θ 0 ) + o P (1)}. (9.8) 140

3 We now use a second Taylor expansion, this time of l(ˆθ n ), to obtain n = (ˆθ n θ 0 )l (θ 0 ) + 1 (ˆθ n θ 0 ) [l (θ 0 ) + o P (1)]. (9.9) If we now substitute Equation (9.8) into Equation (9.9), we obtain { n = n(ˆθ n θ 0 ) 1 n l (θ 0 ) + 1 ( )} 1 n l (θ 0 ) + o P. (9.10) n By Equations (9.1) and (9.3) and Slutsky s theorem, this implies that d n χ 1 under the null hypothesis. Noting that the 1 α quantile of a χ 1 distribution is u α/, we make the following definition. Definition 9.3 A likelihood ratio test is any test that rejects H 0 : θ = θ 0 in favor of H 0 : θ θ 0 when n u α/ or, equivalently, when n u α/. Since it may be shown that n W n P 0 and W n R n P 0, the three tests defined above Wald tests, score tests, and likelihood ratio tests are asymptotically equivalent in the sense that under H 0, they reach the same decision with probability approaching 1 as n. However, they may be quite different for a fixed sample size n, and furthermore they have some relative advantages and disadvantages with respect to one another. For example, It is straightforward to create one-sided Wald and score tests (i.e., tests of H 0 : θ = θ 0 against H 1 : θ > θ 0 or H 1 : θ < θ 0 ), but this is more difficult with a likelihood ratio test. The score test does not require ˆθ n whereas the other two tests do. The Wald test is most easily interpretable and yields immediate confidence intervals. The score test and likelihood ratio test are invariant under reparameterization, whereas the Wald test is not. Example 9.4 Suppose that X 1,..., X n are independent with density f θ (x) = θe xθ I{x > 0}. Then l(θ) = n(log θ θx n ), which yields ( ) 1 l (θ) = n θ X n and l (θ) = n θ. From the form of l (θ), we see that I(θ) = θ, and setting l (θ) = 0 yields ˆθ n = 1/X n. From these facts, we obtain as the Wald, score, and likelihood ratio 141

4 statistics ( ) n 1 W n = θ 0, θ 0 X ( n ) 1 R n = θ 0 n X n = W n, and θ 0 θ 0 X n n = n { X n (X n θ 0 ) log(θ 0 X n ) }. Thus, we reject H 0 : θ = θ 0 in favor of H 1 : θ θ 0 whenever W n u α/, R n u α/, or n u α/, depending on which test we re using. Generalizing the three tests to the multiparameter setting is straightforward. Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0, where θ R k. Then W n R n n def = n(ˆθ θ 0 ) T I(θ 0 )(ˆθ θ 0 ) def = 1 n l(θ0 ) T I 1 (θ 0 ) l(θ 0 ) def = l(ˆθ) l(θ 0 ) d χ k, (9.11) d χ k, and (9.1) d 1 χ k. (9.13) Therefore, if c k α denotes the 1 α quantile of the χ k distribution, then the multivariate Wald test, score test, and likelihood ratio test reject H 0 when W n c k α, R n c k α, and n c k α, respectively. As in the one-parameter case, the Wald test may also be defined with I(θ 0 ) replaced by a consistent estimator Î. Exercises for Section 9.1 Exercise 9.1 Let X 1,..., X n be a simple random sample from a Pareto distribution with density f(x) = θc θ x (θ+1) I{x > c} for a known constant c > 0 and parameter θ > 0. Derive the Wald, Rao, and likelihood ratio tests of θ = θ 0 against a two-sided alternative. Exercise 9. Suppose that X is multinomial(n, p), where p R k. In order to satisfy the regularity condition that the parameter space be an open set, define θ = (p 1,..., p k 1 ). Suppose that we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. (a) Prove that the Wald and score tests are the same as the usual Pearson chisquare test. (b) Derive the likelihood ratio statistic n. 14

5 9. Contiguity and Local Alternatives Suppose that we wish to test the hypotheses H 0 : θ = θ 0 and H 1 : θ > θ 0. (9.14) The test is to be based on a statistic T n, where as always n denotes the sample size, and we shall decide to reject H 0 in (9.14) if T n C n (9.15) for some constant C n. The test in question may be one of the three types of tests introduced in Section 9.1, or it may be an entirely different test. We may define some basic asymptotic concepts regarding tests of this type. Definition 9.5 If P θ0 (T n C n ) α for test (9.15), then test (9.15) is said to have asymptotic level α. Definition 9.6 If two different tests of the same hypotheses reach the same conclusion with probability approaching 1 under the null hypothesis as n, the tests are said to be asymptotically equivalent. The power of test (9.15) under the alternative θ is defined to be β n (θ) = P θ (T n C n ). We expect that the power should approach 1. Definition 9.7 A test (or, more precisely, a sequence of tests) is said to be consistent against the alternative θ if β n (θ) 1. Unfortunately, the concepts we have defined so far are of limited usefulness. If we wish to compare two different tests of the same hypotheses, then if the tests are both sensible they should be asymptotically equivalent and consistent. Thus, consistency is nice but it doesn t tell us much; asymptotic equivalence is nice but it doesn t allow us to compare tests. We make things more interesting by considering, instead of a fixed alternative θ, a sequence of alternatives θ 1, θ,.... Let us make some assumptions about the asymptotic distribution of the test statistic T n in (9.15). First, define µ(θ) and nτ (θ) to be the mean and variance, respectively, of T n when θ is the true value of the parameter. In other words, let µ(θ) = E θ (T n ) and τ (θ) = 1 n Var θ(t n ). We assume that if the null hypothesis is true, which means θ = θ 0, then n{tn µ(θ 0 )} d N(0, 1) (9.16) τ (θ 0 ) 143

6 as n. Furthermore, we assume that if the alternatives are true, which means that θ = θ n for each n, then n{tn µ(θ n )} τ (θ n ) d N(0, 1) (9.17) as n. The limit in (9.17) is trickier than the one in (9.16) because it assumes that the underlying parameter is changing along with n. Consider Example 9.8. Example 9.8 Suppose we have a sequence of independently and identically distributed random variables X 1, X,... with common distribution F θ. In a one-sample t- test, the test statistic T n may be taken to be the sample mean X n. In this case, the limit in (9.16) follows immediately from the central limit theorem. Yet to verify (9.17), it is necessary to consider a triangular array of random variables: X 11 F θ1 X 1, X F θ X 31, X 3, X 33 F θ3... We may often check that the Lyapunov or Lindeberg condition is satisfied, then use the results of Section 4. to establish (9.17). In fact, the existence of, say, a finite third absolute moment, γ(θ) = E θ X 1 3, is sufficient because 1 ( nτ (θ n )) 3 E X ni 3 = γ(θ n) nτ 3 (θ n ) and the Lyapunov condition holds as long as γ(θ n )/τ 3 (θ n ) tends to some finite limit. We generally assume that θ n θ 0, so as long as γ(θ) and τ(θ) are continuous, γ(θ n )/τ 3 (θ n ) tends to γ(θ 0 )/τ 3 (θ 0 ). If (9.16) and (9.17) are true, we may calculate the power of the test against the sequence of alternatives θ 1, θ,... in a straightforward way using normal distribution calculations. As mentioned in Example 9.8, we typically assume that this sequence converges to θ 0. Such a sequence is often referred to as a contiguous sequence of alternatives, since contiguous means next to ; the idea of contiguity is that we choose not a single alternative hypothesis but a sequence of alternatives next to the null hypothesis. First, we should determine a value for C n so that test (9.15) has asymptotic level α. Define u α to be the 1 α quantile of the standard normal distribution. By limit (9.16), { P θ0 T n µ(θ 0 ) τ(θ } 0)u α α; n 144

7 therefore, we define a new test, namely reject H 0 in (9.14) if T n µ(θ 0 ) + τ(θ 0)u α n (9.18) and conclude that test (9.18) has asymptotic level α as desired. We now calculate the power of test (9.18) against the alternative θ n : { β n (θ n ) = P θn T n µ(θ 0 ) + τ(θ } 0)u α n = P θn { n{tn µ(θ n )} τ(θ n ) τ(θ n) τ(θ 0 ) + n(θn θ 0 ) µ(θ n) µ(θ 0 ) τ(θ 0 ) θ n θ 0 u α }.(9.19) Thus, β n (θ n ) tends to an interesting limit (i.e., a limit between α and 1) if τ(θ n ) τ(θ 0 ); n(θn θ 0 ) tends to a nonzero, finite limit; and µ(θ) is differentiable at θ 0. This fact is summarized in the following theorem: Theorem 9.9 Let θ n > θ 0 for all n. Suppose that limits (9.17) and (9.16) hold, τ(θ) is continuous at θ 0, µ(θ) is differentiable at θ 0, and n(θ n θ 0 ) for some finite > 0. If µ (θ 0 ) or τ(θ 0 ) depends on n, then suppose that µ (θ 0 )/τ(θ 0 ) tends to a nonzero, finite limit. Then if β n (θ n ) denotes the power of test (9.18) against the alternative θ n, ( ) µ (θ 0 ) β n (θ n ) lim Φ u α. n τ(θ 0 ) The proof of Theorem 9.9 merely uses Equation (9.19) and Slutsky s theorem, since the hypotheses of the theorem imply that τ(θ n )/τ(θ 0 ) 1 and {µ(θ n ) µ(θ 0 )}/(θ n θ 0 ) µ (θ 0 ). Example 9.10 Let X Binomial(n, p n ), where p n = p 0 + / n and T n = X/n. To test H 0 : p = p 0 against H 1 : p > p 0, note that n(tn p 0 ) d N(0, 1) p0 (1 p 0 ) under H 0. Thus, test (9.18) says to reject H 0 whenever T n p 0 +u α p0 (1 p 0 )/n. This test has asymptotic level α. Since τ(p) = p(1 p) is continuous and µ(p) = p is differentiable, Theorem 9.9 applies in this case as long as we can verify the limit (9.17). Let X n1,..., X nn be independent Bernoulli(p n ) random variables. Then if X n1 p n,..., X nn p n can be shown to satisfy the Lyapunov condition, we have n(tn p n ) d N(0, 1) τ(p n ) 145

8 and so Theorem 9.9 applies. The Lyapunov condition follows since X ni p n 1 implies 1 {Var (nt n )} E X ni p n 4 n {np n (1 p n )} 0. Thus, we conclude by Theorem 9.9 that ( ) β n (p n ) Φ p0 (1 p 0 ) u α. To apply this result, suppose that we wish to test whether a coin is fair by flipping it 100 times. We reject H 0 : p = 1/ in favor of H 1 : p > 1/ if the number of successes divided by 100 is at least as large as 1/ + u.05 /0, or The power of this test against the alternative p = 0.6 is approximately ( ) 100( ) Φ = Φ( 1.645) = Compare this asymptotic approximation with the exact power: The probability of at least 59 successes out of 100 for a binomial(100, 0.6) random variable is As seen in Example 9.10, the asymptotic result of Theorem 9.9 may be applied to a fixed sample size n and a fixed alternative θ as follows: ( ) n{µ(θ) µ(θ0 )} β n (θ) Φ u α (9.0) τ(θ 0 ) There is an alternative formulation that yields a slightly different approximation. Starting from { } n{tn µ(θ)} τ(θ 0 ) n{µ(θ) β n (θ) = P θ u α τ(θ) τ(θ) µ(θ0 )}, τ(θ) we obtain ( ) n{µ(θ) µ(θ0 )} τ(θ 0 ) β n (θ) Φ u α. (9.1) τ(θ) τ(θ) Applying approximation (9.1) to the binomial case of Example 9.10, we obtain instead of 0.63 for the approximate power. 146

9 We may invert approximations (9.0) and (9.1) to obtain approximate sample sizes required to achieve desired power β against alternative θ. From (9.0) we obtain and from (9.1) we obtain n (u α u β )τ(θ 0 ) µ(θ) µ(θ 0 ) (9.) u α τ(θ 0 ) u β τ(θ) n. (9.3) µ(θ) µ(θ 0 ) We may compare tests by considering the relative sample sizes necessary to achieve the same power at the same level against the same alternative. Definition 9.11 Given tests 1 and of the same hypotheses with asymptotic level α and a sequence of alternatives {θ k }, suppose that β (1) m k (θ k ) β and β () n k (θ k ) β as k for some sequences {m k } and {n k } of sample sizes. Then the asymptotic relative efficiency (ARE) of test 1 with respect to test is assuming this limit exists. n k e 1, = lim, k m k In Examples 9.1 and 9.13, we consider two different tests for the same hypotheses. Then, in Example 9.15, we compute their asymptotic relative efficiency. Example 9.1 Suppose we have paired data (X 1, Y 1 ),..., (X n, Y n ). Let Z i = Y i X i for all i. Assume that the Z i are independent and identically distributed with distribution function P (Z i z) = F (z θ) for some θ, where f(z) = F (z) exists and is symmetric about 0. Let W 1,..., W n be a permutation of Z 1,..., Z n such that W 1 W W n. We wish to test H 0 : θ = 0 against H 1 : θ > 0. First, consider the Wilcoxon signed rank test. Define R n = ii{w i > 0}. Then under H 0, the I{W i > 0} are independent Bernoulli(1/) random variables. Thus, i n(n + 1) i n(n + 1)(n + 1) E R n = = and Var R n = =

10 Furthermore, one may prove (see Exercise 9.3) that R n E R n Var Rn d N(0, 1). Thus, a test with asymptotic level α rejects H 0 when R n n(n + 1) 4 + u ατ(0) n, where τ(0) = n (n + 1)(n + 1)/4. Furthermore, we may apply Theorem 9.9. Now we must find E R n under the alternative θ n = / n. As earlier, we let µ(θ) denote E R n under the assumption that θ is the true value of the parameter. Note that in this case, µ(θ) depends on n even though this dependence is not explicit in the notation. First, we note that since W i W j for i j, W i +W j > 0 if and only if W j > 0. Therefore, j I{W i + W j > 0} = ji{w j > 0} and so we may rewrite R n in the form R n = j=1 Therefore, we obtain µ(θ n ) = j=1 j I{W i + W j > 0} = j=1 j I{Z i + Z j > 0}. j P θn (Z i + Z j > 0) = np θn (Z 1 > 0) + Since P θn (Z 1 > 0) = P θn (Z 1 θ n > θ n ) = 1 F ( θ n ) and ( ) n P θn (Z 1 + Z > 0). P θn (Z 1 + Z > 0) = P θn {(Z 1 θ n ) + (Z θ n ) > θ n } we conclude that = E P θn {Z 1 θ n > θ n (Z θ n ) Z } = µ (θ) = nf(θ) + Thus, because f( z) = f(z) by assumption, µ (0) = nf(0) + n(n 1) {1 F ( θ n z)}f(z) dz, ( ) n f( θ z)f(z) dz. 148 f (z) dz.

11 Letting we obtain µ (0) lim n τ(0) = lim n K = f (z) dz, 4{f(0) + (n 1)K} (n + 1)(n + 1) = K 1. Therefore, Theorem 9.9 shows that β n (θ n ) Φ( K 1 u α ). (9.4) Example 9.13 As in Example 9.1, suppose we have paired data (X 1, Y 1 ),..., (X n, Y n ) and Z i = Y i X i for all i. The Z i are independent and identically distributed with distribution function P (Z i z) = F (z θ) for some θ, where f(z) = F (z) exists and is symmetric about 0. Suppose that the variance of Z i is σ. Since the t-test (unknown variance) and z-test (known variance) have the same asymptotic properties, let s consider the z-test for simplicity. Then τ(θ) = σ for all θ. The relevant statistic is merely Z n, and the central limit theorem implies nzn /σ d N(0, 1) under the null hypothesis. Therefore, the z-test in this case rejects H 0 : θ = 0 in favor of H 1 : θ > 0 whenever Z n > u α σ/ n. The central limit theorem shows that n(zn θ n ) d N(0, 1) σ under the alternatives θ n = / n. Therefore, by Theorem 9.9, we obtain since µ (θ) = 1 for all θ. β n (θ n ) Φ( /σ u α ) (9.5) Before finding the asymptotic relative efficiency (ARE) of the Wilcoxon signed rank test and the t-test, we prove a lemma that enables this calculation. Suppose that for two tests, called test 1 and test, we use sample sizes m and n, respectively. We want m and n to tend to infinity together, an idea we make explicit by setting m = m k and n = n k for k = 1,,.... Suppose that we wish to apply both tests to the same sequence of alternative hypotheses θ 1, θ,.... As usual, we make the assumption that (θ k θ 0 ) times the square root of the sample size tends to a finite, nonzero limit as k. Thus, we assume mk (θ k θ 0 ) 1 and n k (θ k θ 0 ). 149

12 Then if Theorem 9.9 may be applied to both tests, define c 1 = lim µ 1(θ 0 )/τ 1 (θ 0 ) and c = lim µ (θ 0 )/τ (θ 0 ). The theorem says that β mk (θ k ) Φ { 1 c 1 u α } and β nk (θ k ) Φ { c u α }. (9.6) To find the ARE, then, Definition 9.11 specifies that we assume that the two limits in (9.6) are the same, which implies 1 c 1 = c, or n k c 1. m k Thus, the ARE of test 1 with respect to test equals (c 1 /c ). This result is summed up in the following lemma, which defines a new term, efficacy. c Lemma 9.14 to be For a test to which Theorem 9.9 applies, define the efficacy of the test c = lim n µ (θ 0 )/τ(θ 0 ). (9.7) Suppose that Theorem 9.9 applies to each of two tests, called test 1 and test. Then the ARE of test 1 with respect to test equals (c 1 /c ). Example 9.15 Using the results of Examples 9.1 and 9.13, we conclude that the efficacies of the Wilcoxon signed rank test and the t-test are 1 f (z) dz and 1 σ, respectively. Thus, Lemma 9.14 implies that the ARE of the signed rank test to the t-test equals ( 1σ f (z) dz). In the case of normally distributed data, we may verify without too much difficulty that the integral above equals (σ π) 1, so the ARE is 3/π Notice how close this is to one, suggesting that for normal data, we lose very little efficiency by using a signed rank test instead of a t-test. In fact, it may be shown that this asymptotic relative efficiency has a lower bound of However, there is no upper bound on the ARE in this case, which means that examples exist for which the t-test is arbitrarily inefficient compared to the signed rank test. Exercises for Section

13 Exercise 9.3 Assume that θ n 0 for some sequence θ 1, θ,.... In Example 9.1, Prove that R n E R n Var Rn d N(0, 1) under the assumption that for each n, the distribution of R n is determined by the parameter θ n. Note that this proof allows us to apply Theorem 9.9, since it covers both the contiguous alternatives and, by taking θ n = θ 0 for all n, the null hypothesis. Exercise 9.4 For the hypotheses considered in Examples 9.1 and 9.13, the sign test is based on the statistic N + = #{i : Z i > 0}. Since n(n + /n 1) d N(0, 1) under the null hypothesis, the sign test (with continuity correction) rejects H 0 when N + 1 u α n + n. (a) Find the efficacy of the sign test. Make sure to indicate how you go about verifying Equation (9.17). (b) Find the ARE of the sign test with respect to the signed rank test and the t-test. Evalaute each of these for the case of normal data. Exercise 9.5 Suppose X 1,..., X n are a simple random sample from a uniform (0, θ) distribution. We wish to test H 0 : θ = θ 0 against H 1 : θ > θ 0 at α =.05. Define Q 1 and Q 3 to be the first and third quartiles of the sample. Consider test A, which rejects when and test B, which rejects when Q 3 Q 1 θ 0 A n, X θ 0 B n. Based on the asymptotic distribution of X and the joint asymptotic distribution of (Q 1, Q 3 ), find the values of A n and B n that correspond with the test in (9.18). Then find the asymptotic relative efficiency of test A relative to test B. Exercise 9.6 Let P θ be a family of probability distributions indexed by a real parameter θ. If X P θ, define µ(θ) = E (X) and σ (θ) = Var (X). Now let θ 1, θ,... be a sequence of parameter values such that θ n θ 0 as n. Suppose that 151

14 E θn X +δ < M for all n for some positive δ and M. Also suppose that for each n, X n1,..., X nn are independent with distribution P θn and define X n = n X ni/n. Prove that if σ (θ 0 ) < and σ (θ) is continuous at the point θ 0, then n[xn µ(θ n )] d N(0, σ (θ 0 )) as n. Hint: a + b +δ +δ ( a +δ + b +δ). Exercise 9.7 Suppose X 1, X,... are independent exponential random variables with mean θ. Consider the test of H 0 : θ = 1 vs H 1 : θ > 1 in which we reject H 0 when where α =.05. X n 1 + u α n, (a) Derive an asymptotic approximation to the power of the test for a fixed sample size n and alternative θ. Tell where you use the result of Problem 9.6. (b) Because the sum of independent exponential random variables is a gamma random variable, it is possible to compute the power exactly in this case. Create a table in which you compare the exact power of the test against the alternative θ = 1. to the asymptotic approximation in part (a) for n {5, 10, 15, 0}. Exercise 9.8 Let X 1,..., X n be independent from Poisson (λ). Create a table in which you list the exact power along with approximations (9.0) and (9.1) for the test that rejects H 0 : λ = 1 in favor of H 1 : λ > 1 when n(xn λ 0 ) λ0 u α, where n = 0 and α =.05, against each of the alternatives 1.1, 1.5, and. Exercise 9.9 Let X 1,..., X n be an independent sample from an exponential distribution with mean λ, and Y 1,..., Y n be an independent sample from an exponential distribution with mean µ. Assume that X i and Y i are independent. We are interested in testing the hypothesis H 0 : λ = µ versus H 1 : λ > µ. Consider the statistic T n = (I i 1/)/ n, 15

15 where I i is the indicator variable I i = I(X i > Y i ). (a) Derive the asymptotic distribution of T n under the null hypothesis. (b) Use the Lindeberg Theorem to show that, under the local alternative hypothesis (λ n, µ n ) = (λ + n 1/ δ, λ), where δ > 0, n (I i ρ n ) L N(0, 1), where ρ n = λ n = λ + n 1/ δ nρn (1 ρ n ) λ n + µ n λ + n 1/ δ. (c) Using the conclusion of part (b), derive the asymptotic distribution of T n under the local alternative specified in (b). Exercise 9.10 Suppose X 1,..., X m is a simple random sample and Y 1,..., Y n is another simple random sample independent of the X i, with P (X i t) = t for t [0, 1] and P (Y i t) = (t θ) for t [θ, θ + 1]. Assume m/(m + n) ρ as m, n and 0 < θ < 1. Find the asymptotic distribution of m + n[g(y X) g(θ)]. 9.3 The Wilcoxon Rank-Sum Test Suppose that X 1,..., X m and Y 1,..., Y n are two independent simple random samples, with P (X i t) = P (Y j t + θ) = F (t) (9.8) for some continuous distribution function F (t) with f(t) = F (t). Thus, the distribution of the Y j is shifted by θ from the distribution of the X i. We wish to test H 0 : θ = 0 against H 1 : θ > 0. To do the asymptotics here, we will assume that n and m are actually both elements of separate sequences of sample sizes, indexed by a third variable, say k. Thus, m = m k and n = n k both go to as k, and we suppress the subscript k on m and n for convenience of notation. Suppose that we combine the X i and Y j into a single sample of size m + n. Define the Wilcoxon rank-sum statistic to be S k = Rank of Y j among combined sample. j=1 Letting Y (1),..., Y (n) denote the order statistics for the sample of Y j as usual, we may rewrite S k in the following way: S k = Rank of Y (j) among combined sample j=1 153

16 = = = ( j + #{i : Xi < Y (j) } ) j=1 n(n + 1) n(n + 1) + + j=1 j=1 m I{X i < Y (j) } m I{X i < Y j }. (9.9) Let N = m + n, and suppose that m/n ρ as k for some constant ρ (0, 1). First, we will establish the asymptotic behavior of S k under the null hypothesis. Define µ(θ) = E θ S k and τ(θ) = N Var S k. To evaluate µ(θ 0 ) and τ(θ 0 ), let Z i = n j=1 I{X i < Y j }. Then the Z i are identically distributed but not independent, and we have E θ0 Z i = n/ and Furthermore, E θ0 Z i Z j = which implies Therefore, and r=1 s=1 Var θ0 Z i = n 4 + n(n 1) Cov θ 0 (I{X i < Y 1 }, I{X i < Y }) = n n(n 1) n(n + ) =. 1 P θ0 (X i < Y r and X j < Y s ) = Cov θ0 (Z i, Z j ) = µ(θ 0 ) = n(n + 1) n(n 1) 4 + mn n(n 1) 4 n(n 1) 4 + n 3 n 4 = n 1. = n(n + 1) τ (θ 0 ) = Nm Var Z 1 + Nm(m 1) Cov (Z 1, Z ) = Nmn(n + ) Nm(m 1)n = Nmn(N + 1) np θ0 (X i < Y 1 and X j < Y 1 ),

17 Let θ 1, θ,... be a sequence of alternatives such that N(θ k θ 0 ) for a positive, finite constant. It is possible to show that under H 0 and N{Sk µ(θ 0 )} τ(θ 0 ) d N(0, 1) (9.30) N{Sk µ(θ k )} τ(θ 0 ) d N(0, 1) (9.31) under the alternatives {θ k } (see Exercise 9.11). By expression (9.30), the test based on S k with asymptotic level α rejects H 0 : θ = 0 in favor of H a : θ > 0 whenever S k µ(θ 0 ) + u α τ(θ 0 )/ N, or S k n(n + 1) mn(n + 1) + u α. 1 The Wilcoxon rank-sum test is sometimes referred to as the Mann-Whitney test. This alternative name helps to distinguish this test from the similarly named Wilcoxon signed rank test. To find the limiting power of the rank-sum test, we may use Theorem 9.9 to conclude that ( ) µ (θ 0 ) β k (θ k ) lim Φ u α. (9.3) k τ(θ 0 ) According to expression (9.3), we must evaluate µ (θ 0 ). To this end, note that Therefore, This gives P θ (X 1 < Y 1 ) = E θ {P θ (X 1 < Y 1 Y 1 )} = E θ F (Y 1 ) = = d dθ P θ(x 1 < Y 1 ) = µ (0) = mn F (y)f(y θ) dy F (y + θ)f(y) dy. f(y + θ)f(y) dy. f (y) dy. 155

18 Thus, the efficacy of the Wilcoxon rank-sum test is µ (θ 0 ) lim k τ(θ 0 ) = lim k mn 1 f (y) dy mnn(n + 1) = 1ρ(1 ρ) The asymptotic power of the test follows immediately from (9.3). f (y) dy. Exercises for Section 9.3 Exercise 9.11 (a) Prove expression (9.30) under the null hypothesis H 0 : θ = θ 0 (where θ 0 = 0). (b) Prove expression (9.31) under the sequence of alternatives θ 1, θ,... satisfying N(θk θ 0 ). Hint: In both (a) and (b), start by verifying either the Lindeberg condition or the Lyapunov condition. Exercise 9.1 Suppose Var X i = Var Y i = σ < and we wish to test the hypotheses H 0 : θ = 0 vs. H 1 : θ > 0 using the two-sample Z-statistic Y X σ 1 m + 1 n Note that this Z-statistic is s/σ times the usual T-statistic, where s is the pooled sample standard deviation, so the asymptotic properties of the T-statistic are the same as those of the Z-statistic. (a) Find the efficacy of the Z test. Justify your use of Theorem 9.9. (b) Find the ARE of the Z test with respect to the rank-sum test for normally distributed data. (c) Find the ARE of the Z test with respect to the rank-sum test if the data come from a double exponential distribution with f(t) = 1 λ e t/λ. (d) Prove by example that the ARE of the Z-test with respect to the rank-sum test can be arbitrarily close to zero.. 156

Maximum Likelihood Estimation

Maximum Likelihood Estimation Chapter 8 Maximum Likelihood Estimation 8. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Chapter 7 Maximum Likelihood Estimation 7. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

under the null hypothesis, the sign test (with continuity correction) rejects H 0 when α n + n 2 2.

under the null hypothesis, the sign test (with continuity correction) rejects H 0 when α n + n 2 2. Assignment 13 Exercise 8.4 For the hypotheses consiere in Examples 8.12 an 8.13, the sign test is base on the statistic N + = #{i : Z i > 0}. Since 2 n(n + /n 1) N(0, 1) 2 uner the null hypothesis, the

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007 Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Session 3 The proportional odds model and the Mann-Whitney test

Session 3 The proportional odds model and the Mann-Whitney test Session 3 The proportional odds model and the Mann-Whitney test 3.1 A unified approach to inference 3.2 Analysis via dichotomisation 3.3 Proportional odds 3.4 Relationship with the Mann-Whitney test Session

More information

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

HYPOTHESIS TESTING: FREQUENTIST APPROACH. HYPOTHESIS TESTING: FREQUENTIST APPROACH. These notes summarize the lectures on (the frequentist approach to) hypothesis testing. You should be familiar with the standard hypothesis testing from previous

More information

Loglikelihood and Confidence Intervals

Loglikelihood and Confidence Intervals Stat 504, Lecture 2 1 Loglikelihood and Confidence Intervals The loglikelihood function is defined to be the natural logarithm of the likelihood function, l(θ ; x) = log L(θ ; x). For a variety of reasons,

More information

Suggested solutions to written exam Jan 17, 2012

Suggested solutions to written exam Jan 17, 2012 LINKÖPINGS UNIVERSITET Institutionen för datavetenskap Statistik, ANd 73A36 THEORY OF STATISTICS, 6 CDTS Master s program in Statistics and Data Mining Fall semester Written exam Suggested solutions to

More information

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1 Chapter 4 HOMEWORK ASSIGNMENTS These homeworks may be modified as the semester progresses. It is your responsibility to keep up to date with the correctly assigned homeworks. There may be some errors in

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30 MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)

More information

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

Topic 15: Simple Hypotheses

Topic 15: Simple Hypotheses Topic 15: November 10, 2009 In the simplest set-up for a statistical hypothesis, we consider two values θ 0, θ 1 in the parameter space. We write the test as H 0 : θ = θ 0 versus H 1 : θ = θ 1. H 0 is

More information

Chapter 4. Theory of Tests. 4.1 Introduction

Chapter 4. Theory of Tests. 4.1 Introduction Chapter 4 Theory of Tests 4.1 Introduction Parametric model: (X, B X, P θ ), P θ P = {P θ θ Θ} where Θ = H 0 +H 1 X = K +A : K: critical region = rejection region / A: acceptance region A decision rule

More information

Chapter 3. Point Estimation. 3.1 Introduction

Chapter 3. Point Estimation. 3.1 Introduction Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.

More information

Lecture 1: Introduction

Lecture 1: Introduction Principles of Statistics Part II - Michaelmas 208 Lecturer: Quentin Berthet Lecture : Introduction This course is concerned with presenting some of the mathematical principles of statistical theory. One

More information

Final Examination Statistics 200C. T. Ferguson June 11, 2009

Final Examination Statistics 200C. T. Ferguson June 11, 2009 Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean

More information

STAT 514 Solutions to Assignment #6

STAT 514 Solutions to Assignment #6 STAT 514 Solutions to Assignment #6 Question 1: Suppose that X 1,..., X n are a simple random sample from a Weibull distribution with density function f θ x) = θcx c 1 exp{ θx c }I{x > 0} for some fixed

More information

Exercises and Answers to Chapter 1

Exercises and Answers to Chapter 1 Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean

More information

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain 0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher

More information

Asymptotic Statistics-VI. Changliang Zou

Asymptotic Statistics-VI. Changliang Zou Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous

More information

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j. Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That

More information

Greene, Econometric Analysis (6th ed, 2008)

Greene, Econometric Analysis (6th ed, 2008) EC771: Econometrics, Spring 2010 Greene, Econometric Analysis (6th ed, 2008) Chapter 17: Maximum Likelihood Estimation The preferred estimator in a wide variety of econometric settings is that derived

More information

4.5.1 The use of 2 log Λ when θ is scalar

4.5.1 The use of 2 log Λ when θ is scalar 4.5. ASYMPTOTIC FORM OF THE G.L.R.T. 97 4.5.1 The use of 2 log Λ when θ is scalar Suppose we wish to test the hypothesis NH : θ = θ where θ is a given value against the alternative AH : θ θ on the basis

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015 STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis

More information

f(y θ) = g(t (y) θ)h(y)

f(y θ) = g(t (y) θ)h(y) EXAM3, FINAL REVIEW (and a review for some of the QUAL problems): No notes will be allowed, but you may bring a calculator. Memorize the pmf or pdf f, E(Y ) and V(Y ) for the following RVs: 1) beta(δ,

More information

Stat 411 Lecture Notes 03 Likelihood and Maximum Likelihood Estimation

Stat 411 Lecture Notes 03 Likelihood and Maximum Likelihood Estimation Stat 411 Lecture Notes 03 Likelihood and Maximum Likelihood Estimation Ryan Martin www.math.uic.edu/~rgmartin Version: August 19, 2013 1 Introduction Previously we have discussed various properties of

More information

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test. Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the

More information

Chapter 7. Hypothesis Testing

Chapter 7. Hypothesis Testing Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function

More information

Theoretical Statistics. Lecture 23.

Theoretical Statistics. Lecture 23. Theoretical Statistics. Lecture 23. Peter Bartlett 1. Recall: QMD and local asymptotic normality. [vdv7] 2. Convergence of experiments, maximum likelihood. 3. Relative efficiency of tests. [vdv14] 1 Local

More information

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple

More information

Stat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota

Stat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota Stat 5102 Lecture Slides Deck 3 Charles J. Geyer School of Statistics University of Minnesota 1 Likelihood Inference We have learned one very general method of estimation: method of moments. the Now we

More information

STA 260: Statistics and Probability II

STA 260: Statistics and Probability II Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition

More information

Theory of Statistics.

Theory of Statistics. Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Testing Statistical Hypotheses

Testing Statistical Hypotheses E.L. Lehmann Joseph P. Romano Testing Statistical Hypotheses Third Edition 4y Springer Preface vii I Small-Sample Theory 1 1 The General Decision Problem 3 1.1 Statistical Inference and Statistical Decisions

More information

14.30 Introduction to Statistical Methods in Economics Spring 2009

14.30 Introduction to Statistical Methods in Economics Spring 2009 MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Chapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics

Chapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics Chapter 6 Order Statistics and Quantiles 61 Extreme Order Statistics Suppose we have a finite sample X 1,, X n Conditional on this sample, we define the values X 1),, X n) to be a permutation of X 1,,

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

Lecture 32: Asymptotic confidence sets and likelihoods

Lecture 32: Asymptotic confidence sets and likelihoods Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

Statistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm

Statistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm Statistics GIDP Ph.D. Qualifying Exam Theory Jan, 06, 9:00am-:00pm Instructions: Provide answers on the supplied pads of paper; write on only one side of each sheet. Complete exactly 5 of the 6 problems.

More information

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

Maximum Likelihood Tests and Quasi-Maximum-Likelihood Maximum Likelihood Tests and Quasi-Maximum-Likelihood Wendelin Schnedler Department of Economics University of Heidelberg 10. Dezember 2007 Wendelin Schnedler (AWI) Maximum Likelihood Tests and Quasi-Maximum-Likelihood10.

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Statistics and econometrics

Statistics and econometrics 1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample

More information

1. Fisher Information

1. Fisher Information 1. Fisher Information Let f(x θ) be a density function with the property that log f(x θ) is differentiable in θ throughout the open p-dimensional parameter set Θ R p ; then the score statistic (or score

More information

Statistics Ph.D. Qualifying Exam

Statistics Ph.D. Qualifying Exam Department of Statistics Carnegie Mellon University May 7 2008 Statistics Ph.D. Qualifying Exam You are not expected to solve all five problems. Complete solutions to few problems will be preferred to

More information

TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1

TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1 TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1 1.1 The Probability Model...1 1.2 Finite Discrete Models with Equally Likely Outcomes...5 1.2.1 Tree Diagrams...6 1.2.2 The Multiplication Principle...8

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

Statistics. Statistics

Statistics. Statistics The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,

More information

Lecture 17: Likelihood ratio and asymptotic tests

Lecture 17: Likelihood ratio and asymptotic tests Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f

More information

1 Likelihood. 1.1 Likelihood function. Likelihood & Maximum Likelihood Estimators

1 Likelihood. 1.1 Likelihood function. Likelihood & Maximum Likelihood Estimators Likelihood & Maximum Likelihood Estimators February 26, 2018 Debdeep Pati 1 Likelihood Likelihood is surely one of the most important concepts in statistical theory. We have seen the role it plays in sufficiency,

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Testing Statistical Hypotheses

Testing Statistical Hypotheses E.L. Lehmann Joseph P. Romano, 02LEu1 ttd ~Lt~S Testing Statistical Hypotheses Third Edition With 6 Illustrations ~Springer 2 The Probability Background 28 2.1 Probability and Measure 28 2.2 Integration.........

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants 18.650 Statistics for Applications Chapter 5: Parametric hypothesis testing 1/37 Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

Chapters 10. Hypothesis Testing

Chapters 10. Hypothesis Testing Chapters 10. Hypothesis Testing Some examples of hypothesis testing 1. Toss a coin 100 times and get 62 heads. Is this coin a fair coin? 2. Is the new treatment on blood pressure more effective than the

More information

Parametric Inference

Parametric Inference Parametric Inference Moulinath Banerjee University of Michigan April 14, 2004 1 General Discussion The object of statistical inference is to glean information about an underlying population based on a

More information

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018 Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from

More information

Introducing the Normal Distribution

Introducing the Normal Distribution Department of Mathematics Ma 3/13 KC Border Introduction to Probability and Statistics Winter 219 Lecture 1: Introducing the Normal Distribution Relevant textbook passages: Pitman [5]: Sections 1.2, 2.2,

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1 4 Hypothesis testing 4. Simple hypotheses A computer tries to distinguish between two sources of signals. Both sources emit independent signals with normally distributed intensity, the signals of the first

More information

Problems ( ) 1 exp. 2. n! e λ and

Problems ( ) 1 exp. 2. n! e λ and Problems The expressions for the probability mass function of the Poisson(λ) distribution, and the density function of the Normal distribution with mean µ and variance σ 2, may be useful: ( ) 1 exp. 2πσ

More information

STA 732: Inference. Notes 4. General Classical Methods & Efficiency of Tests B&D 2, 4, 5

STA 732: Inference. Notes 4. General Classical Methods & Efficiency of Tests B&D 2, 4, 5 STA 732: Inference Notes 4. General Classical Methods & Efficiency of Tests B&D 2, 4, 5 1 Going beyond ML with ML type methods 1.1 When ML fails Example. Is LSAT correlated with GPA? Suppose we want to

More information

Statistical Theory MT 2006 Problems 4: Solution sketches

Statistical Theory MT 2006 Problems 4: Solution sketches Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine

More information

Statistical Theory MT 2007 Problems 4: Solution sketches

Statistical Theory MT 2007 Problems 4: Solution sketches Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the

More information

The outline for Unit 3

The outline for Unit 3 The outline for Unit 3 Unit 1. Introduction: The regression model. Unit 2. Estimation principles. Unit 3: Hypothesis testing principles. 3.1 Wald test. 3.2 Lagrange Multiplier. 3.3 Likelihood Ratio Test.

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Hypothesis Testing: The Generalized Likelihood Ratio Test

Hypothesis Testing: The Generalized Likelihood Ratio Test Hypothesis Testing: The Generalized Likelihood Ratio Test Consider testing the hypotheses H 0 : θ Θ 0 H 1 : θ Θ \ Θ 0 Definition: The Generalized Likelihood Ratio (GLR Let L(θ be a likelihood for a random

More information

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

Probability and Statistics

Probability and Statistics Probability and Statistics 1 Contents some stochastic processes Stationary Stochastic Processes 2 4. Some Stochastic Processes 4.1 Bernoulli process 4.2 Binomial process 4.3 Sine wave process 4.4 Random-telegraph

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

For iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions.

For iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions. Large Sample Theory Study approximate behaviour of ˆθ by studying the function U. Notice U is sum of independent random variables. Theorem: If Y 1, Y 2,... are iid with mean µ then Yi n µ Called law of

More information

Nonparametric hypothesis tests and permutation tests

Nonparametric hypothesis tests and permutation tests Nonparametric hypothesis tests and permutation tests 1.7 & 2.3. Probability Generating Functions 3.8.3. Wilcoxon Signed Rank Test 3.8.2. Mann-Whitney Test Prof. Tesler Math 283 Fall 2018 Prof. Tesler Wilcoxon

More information

Solution: First note that the power function of the test is given as follows,

Solution: First note that the power function of the test is given as follows, Problem 4.5.8: Assume the life of a tire given by X is distributed N(θ, 5000 ) Past experience indicates that θ = 30000. The manufacturere claims the tires made by a new process have mean θ > 30000. Is

More information

MATH4427 Notebook 2 Fall Semester 2017/2018

MATH4427 Notebook 2 Fall Semester 2017/2018 MATH4427 Notebook 2 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

Relative efficiency. Patrick Breheny. October 9. Theoretical framework Application to the two-group problem

Relative efficiency. Patrick Breheny. October 9. Theoretical framework Application to the two-group problem Relative efficiency Patrick Breheny October 9 Patrick Breheny STA 621: Nonparametric Statistics 1/15 Relative efficiency Suppose test 1 requires n 1 observations to obtain a certain power β, and that test

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information