Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n
|
|
- Donna Thornton
- 5 years ago
- Views:
Transcription
1 Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible tests. To this end, let l(θ) denote the loglikelihood and ˆθ n the consistent root of the likelihood equation. Intuitively, the farther θ 0 is from ˆθ n, the stronger the evidence against the null hypothesis. But how far is far enough? Note that if θ 0 is close to ˆθ n, then l(θ 0 ) should also be close to l(ˆθ n ) and l (θ 0 ) should be close to l (ˆθ n ) = 0. If we base a test upon the value of (ˆθ n θ 0 ), we obtain a Wald test. If we base a test upon the value of l(ˆθ n ) l(θ 0 ), we obtain a likelihood ratio test. If we base a test upon the value of l (θ 0 ), we obtain a (Rao) score test. Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : ( ) d 1 n(ˆθn θ 0 ) N 0, ; (9.1) I(θ 0 ) 1 n l (θ 0 ) d N{0, I(θ 0 )}; (9.) 1 n l (θ 0 ) P I(θ 0 ). (9.3) 139
2 Equation (9.) is proved using the central limit theorem; Equation (9.3) is proved using the weak law of large numbers; and Equation (9.1) is the result of a Taylor expansion together with Equations (9.) and (9.3). The three equations above will help us to justify the definitions of Wald, score, and likelihood ratio tests to follow. Equation (9.1) suggests that if we define W n = ni(θ 0 )(ˆθ n θ 0 ), (9.4) then W n, called a Wald statistic, should converge in distribution to standard normal under H 0. Note that this fact remains true if we define W n = nî(ˆθ n θ 0 ), (9.5) where Î is consistent for θ 0; for example, Î could be I(ˆθ n ). Definition 9.1 A Wald test is any test that rejects H 0 : θ = θ 0 in favor of H 1 : θ θ 0 when W n u α/ for W n defined as in Equation (9.4) or Equation (9.5). As usual, u α/ denotes the 1 α quantile of the standard normal distribution. Equation (9.) suggests that if we define R n = 1 ni(θ0 ) l (θ 0 ), (9.6) then R n, called a Rao score statistic or simply a score statistic, converges in distribution to standard normal under H 0. We could also replace I(θ 0 ) by a consistent estimator Î as in Equation (9.5), but usually this is not done: One of the main benefits of the score statistic is that it is not necessary to compute ˆθ n, and using I(ˆθ n ) instead of I(θ 0 ) would defeat this purpose. Definition 9. A score test, sometimes called a Rao score test, is any test that rejects H 0 : θ = θ 0 in favor of H 1 : θ θ 0 when R n u α/ for R n defined as in Equation (9.6). The third type of test, the likelihood ratio test, requires a bit of development. It is a test based on the statistic If we Taylor expand l (ˆθ n ) = 0 around the point θ 0, we obtain n = l(ˆθ n ) l(θ 0 ). (9.7) l (θ 0 ) = (ˆθ n θ 0 ) { l (θ 0 ) + o P (1)}. (9.8) 140
3 We now use a second Taylor expansion, this time of l(ˆθ n ), to obtain n = (ˆθ n θ 0 )l (θ 0 ) + 1 (ˆθ n θ 0 ) [l (θ 0 ) + o P (1)]. (9.9) If we now substitute Equation (9.8) into Equation (9.9), we obtain { n = n(ˆθ n θ 0 ) 1 n l (θ 0 ) + 1 ( )} 1 n l (θ 0 ) + o P. (9.10) n By Equations (9.1) and (9.3) and Slutsky s theorem, this implies that d n χ 1 under the null hypothesis. Noting that the 1 α quantile of a χ 1 distribution is u α/, we make the following definition. Definition 9.3 A likelihood ratio test is any test that rejects H 0 : θ = θ 0 in favor of H 0 : θ θ 0 when n u α/ or, equivalently, when n u α/. Since it may be shown that n W n P 0 and W n R n P 0, the three tests defined above Wald tests, score tests, and likelihood ratio tests are asymptotically equivalent in the sense that under H 0, they reach the same decision with probability approaching 1 as n. However, they may be quite different for a fixed sample size n, and furthermore they have some relative advantages and disadvantages with respect to one another. For example, It is straightforward to create one-sided Wald and score tests (i.e., tests of H 0 : θ = θ 0 against H 1 : θ > θ 0 or H 1 : θ < θ 0 ), but this is more difficult with a likelihood ratio test. The score test does not require ˆθ n whereas the other two tests do. The Wald test is most easily interpretable and yields immediate confidence intervals. The score test and likelihood ratio test are invariant under reparameterization, whereas the Wald test is not. Example 9.4 Suppose that X 1,..., X n are independent with density f θ (x) = θe xθ I{x > 0}. Then l(θ) = n(log θ θx n ), which yields ( ) 1 l (θ) = n θ X n and l (θ) = n θ. From the form of l (θ), we see that I(θ) = θ, and setting l (θ) = 0 yields ˆθ n = 1/X n. From these facts, we obtain as the Wald, score, and likelihood ratio 141
4 statistics ( ) n 1 W n = θ 0, θ 0 X ( n ) 1 R n = θ 0 n X n = W n, and θ 0 θ 0 X n n = n { X n (X n θ 0 ) log(θ 0 X n ) }. Thus, we reject H 0 : θ = θ 0 in favor of H 1 : θ θ 0 whenever W n u α/, R n u α/, or n u α/, depending on which test we re using. Generalizing the three tests to the multiparameter setting is straightforward. Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0, where θ R k. Then W n R n n def = n(ˆθ θ 0 ) T I(θ 0 )(ˆθ θ 0 ) def = 1 n l(θ0 ) T I 1 (θ 0 ) l(θ 0 ) def = l(ˆθ) l(θ 0 ) d χ k, (9.11) d χ k, and (9.1) d 1 χ k. (9.13) Therefore, if c k α denotes the 1 α quantile of the χ k distribution, then the multivariate Wald test, score test, and likelihood ratio test reject H 0 when W n c k α, R n c k α, and n c k α, respectively. As in the one-parameter case, the Wald test may also be defined with I(θ 0 ) replaced by a consistent estimator Î. Exercises for Section 9.1 Exercise 9.1 Let X 1,..., X n be a simple random sample from a Pareto distribution with density f(x) = θc θ x (θ+1) I{x > c} for a known constant c > 0 and parameter θ > 0. Derive the Wald, Rao, and likelihood ratio tests of θ = θ 0 against a two-sided alternative. Exercise 9. Suppose that X is multinomial(n, p), where p R k. In order to satisfy the regularity condition that the parameter space be an open set, define θ = (p 1,..., p k 1 ). Suppose that we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. (a) Prove that the Wald and score tests are the same as the usual Pearson chisquare test. (b) Derive the likelihood ratio statistic n. 14
5 9. Contiguity and Local Alternatives Suppose that we wish to test the hypotheses H 0 : θ = θ 0 and H 1 : θ > θ 0. (9.14) The test is to be based on a statistic T n, where as always n denotes the sample size, and we shall decide to reject H 0 in (9.14) if T n C n (9.15) for some constant C n. The test in question may be one of the three types of tests introduced in Section 9.1, or it may be an entirely different test. We may define some basic asymptotic concepts regarding tests of this type. Definition 9.5 If P θ0 (T n C n ) α for test (9.15), then test (9.15) is said to have asymptotic level α. Definition 9.6 If two different tests of the same hypotheses reach the same conclusion with probability approaching 1 under the null hypothesis as n, the tests are said to be asymptotically equivalent. The power of test (9.15) under the alternative θ is defined to be β n (θ) = P θ (T n C n ). We expect that the power should approach 1. Definition 9.7 A test (or, more precisely, a sequence of tests) is said to be consistent against the alternative θ if β n (θ) 1. Unfortunately, the concepts we have defined so far are of limited usefulness. If we wish to compare two different tests of the same hypotheses, then if the tests are both sensible they should be asymptotically equivalent and consistent. Thus, consistency is nice but it doesn t tell us much; asymptotic equivalence is nice but it doesn t allow us to compare tests. We make things more interesting by considering, instead of a fixed alternative θ, a sequence of alternatives θ 1, θ,.... Let us make some assumptions about the asymptotic distribution of the test statistic T n in (9.15). First, define µ(θ) and nτ (θ) to be the mean and variance, respectively, of T n when θ is the true value of the parameter. In other words, let µ(θ) = E θ (T n ) and τ (θ) = 1 n Var θ(t n ). We assume that if the null hypothesis is true, which means θ = θ 0, then n{tn µ(θ 0 )} d N(0, 1) (9.16) τ (θ 0 ) 143
6 as n. Furthermore, we assume that if the alternatives are true, which means that θ = θ n for each n, then n{tn µ(θ n )} τ (θ n ) d N(0, 1) (9.17) as n. The limit in (9.17) is trickier than the one in (9.16) because it assumes that the underlying parameter is changing along with n. Consider Example 9.8. Example 9.8 Suppose we have a sequence of independently and identically distributed random variables X 1, X,... with common distribution F θ. In a one-sample t- test, the test statistic T n may be taken to be the sample mean X n. In this case, the limit in (9.16) follows immediately from the central limit theorem. Yet to verify (9.17), it is necessary to consider a triangular array of random variables: X 11 F θ1 X 1, X F θ X 31, X 3, X 33 F θ3... We may often check that the Lyapunov or Lindeberg condition is satisfied, then use the results of Section 4. to establish (9.17). In fact, the existence of, say, a finite third absolute moment, γ(θ) = E θ X 1 3, is sufficient because 1 ( nτ (θ n )) 3 E X ni 3 = γ(θ n) nτ 3 (θ n ) and the Lyapunov condition holds as long as γ(θ n )/τ 3 (θ n ) tends to some finite limit. We generally assume that θ n θ 0, so as long as γ(θ) and τ(θ) are continuous, γ(θ n )/τ 3 (θ n ) tends to γ(θ 0 )/τ 3 (θ 0 ). If (9.16) and (9.17) are true, we may calculate the power of the test against the sequence of alternatives θ 1, θ,... in a straightforward way using normal distribution calculations. As mentioned in Example 9.8, we typically assume that this sequence converges to θ 0. Such a sequence is often referred to as a contiguous sequence of alternatives, since contiguous means next to ; the idea of contiguity is that we choose not a single alternative hypothesis but a sequence of alternatives next to the null hypothesis. First, we should determine a value for C n so that test (9.15) has asymptotic level α. Define u α to be the 1 α quantile of the standard normal distribution. By limit (9.16), { P θ0 T n µ(θ 0 ) τ(θ } 0)u α α; n 144
7 therefore, we define a new test, namely reject H 0 in (9.14) if T n µ(θ 0 ) + τ(θ 0)u α n (9.18) and conclude that test (9.18) has asymptotic level α as desired. We now calculate the power of test (9.18) against the alternative θ n : { β n (θ n ) = P θn T n µ(θ 0 ) + τ(θ } 0)u α n = P θn { n{tn µ(θ n )} τ(θ n ) τ(θ n) τ(θ 0 ) + n(θn θ 0 ) µ(θ n) µ(θ 0 ) τ(θ 0 ) θ n θ 0 u α }.(9.19) Thus, β n (θ n ) tends to an interesting limit (i.e., a limit between α and 1) if τ(θ n ) τ(θ 0 ); n(θn θ 0 ) tends to a nonzero, finite limit; and µ(θ) is differentiable at θ 0. This fact is summarized in the following theorem: Theorem 9.9 Let θ n > θ 0 for all n. Suppose that limits (9.17) and (9.16) hold, τ(θ) is continuous at θ 0, µ(θ) is differentiable at θ 0, and n(θ n θ 0 ) for some finite > 0. If µ (θ 0 ) or τ(θ 0 ) depends on n, then suppose that µ (θ 0 )/τ(θ 0 ) tends to a nonzero, finite limit. Then if β n (θ n ) denotes the power of test (9.18) against the alternative θ n, ( ) µ (θ 0 ) β n (θ n ) lim Φ u α. n τ(θ 0 ) The proof of Theorem 9.9 merely uses Equation (9.19) and Slutsky s theorem, since the hypotheses of the theorem imply that τ(θ n )/τ(θ 0 ) 1 and {µ(θ n ) µ(θ 0 )}/(θ n θ 0 ) µ (θ 0 ). Example 9.10 Let X Binomial(n, p n ), where p n = p 0 + / n and T n = X/n. To test H 0 : p = p 0 against H 1 : p > p 0, note that n(tn p 0 ) d N(0, 1) p0 (1 p 0 ) under H 0. Thus, test (9.18) says to reject H 0 whenever T n p 0 +u α p0 (1 p 0 )/n. This test has asymptotic level α. Since τ(p) = p(1 p) is continuous and µ(p) = p is differentiable, Theorem 9.9 applies in this case as long as we can verify the limit (9.17). Let X n1,..., X nn be independent Bernoulli(p n ) random variables. Then if X n1 p n,..., X nn p n can be shown to satisfy the Lyapunov condition, we have n(tn p n ) d N(0, 1) τ(p n ) 145
8 and so Theorem 9.9 applies. The Lyapunov condition follows since X ni p n 1 implies 1 {Var (nt n )} E X ni p n 4 n {np n (1 p n )} 0. Thus, we conclude by Theorem 9.9 that ( ) β n (p n ) Φ p0 (1 p 0 ) u α. To apply this result, suppose that we wish to test whether a coin is fair by flipping it 100 times. We reject H 0 : p = 1/ in favor of H 1 : p > 1/ if the number of successes divided by 100 is at least as large as 1/ + u.05 /0, or The power of this test against the alternative p = 0.6 is approximately ( ) 100( ) Φ = Φ( 1.645) = Compare this asymptotic approximation with the exact power: The probability of at least 59 successes out of 100 for a binomial(100, 0.6) random variable is As seen in Example 9.10, the asymptotic result of Theorem 9.9 may be applied to a fixed sample size n and a fixed alternative θ as follows: ( ) n{µ(θ) µ(θ0 )} β n (θ) Φ u α (9.0) τ(θ 0 ) There is an alternative formulation that yields a slightly different approximation. Starting from { } n{tn µ(θ)} τ(θ 0 ) n{µ(θ) β n (θ) = P θ u α τ(θ) τ(θ) µ(θ0 )}, τ(θ) we obtain ( ) n{µ(θ) µ(θ0 )} τ(θ 0 ) β n (θ) Φ u α. (9.1) τ(θ) τ(θ) Applying approximation (9.1) to the binomial case of Example 9.10, we obtain instead of 0.63 for the approximate power. 146
9 We may invert approximations (9.0) and (9.1) to obtain approximate sample sizes required to achieve desired power β against alternative θ. From (9.0) we obtain and from (9.1) we obtain n (u α u β )τ(θ 0 ) µ(θ) µ(θ 0 ) (9.) u α τ(θ 0 ) u β τ(θ) n. (9.3) µ(θ) µ(θ 0 ) We may compare tests by considering the relative sample sizes necessary to achieve the same power at the same level against the same alternative. Definition 9.11 Given tests 1 and of the same hypotheses with asymptotic level α and a sequence of alternatives {θ k }, suppose that β (1) m k (θ k ) β and β () n k (θ k ) β as k for some sequences {m k } and {n k } of sample sizes. Then the asymptotic relative efficiency (ARE) of test 1 with respect to test is assuming this limit exists. n k e 1, = lim, k m k In Examples 9.1 and 9.13, we consider two different tests for the same hypotheses. Then, in Example 9.15, we compute their asymptotic relative efficiency. Example 9.1 Suppose we have paired data (X 1, Y 1 ),..., (X n, Y n ). Let Z i = Y i X i for all i. Assume that the Z i are independent and identically distributed with distribution function P (Z i z) = F (z θ) for some θ, where f(z) = F (z) exists and is symmetric about 0. Let W 1,..., W n be a permutation of Z 1,..., Z n such that W 1 W W n. We wish to test H 0 : θ = 0 against H 1 : θ > 0. First, consider the Wilcoxon signed rank test. Define R n = ii{w i > 0}. Then under H 0, the I{W i > 0} are independent Bernoulli(1/) random variables. Thus, i n(n + 1) i n(n + 1)(n + 1) E R n = = and Var R n = =
10 Furthermore, one may prove (see Exercise 9.3) that R n E R n Var Rn d N(0, 1). Thus, a test with asymptotic level α rejects H 0 when R n n(n + 1) 4 + u ατ(0) n, where τ(0) = n (n + 1)(n + 1)/4. Furthermore, we may apply Theorem 9.9. Now we must find E R n under the alternative θ n = / n. As earlier, we let µ(θ) denote E R n under the assumption that θ is the true value of the parameter. Note that in this case, µ(θ) depends on n even though this dependence is not explicit in the notation. First, we note that since W i W j for i j, W i +W j > 0 if and only if W j > 0. Therefore, j I{W i + W j > 0} = ji{w j > 0} and so we may rewrite R n in the form R n = j=1 Therefore, we obtain µ(θ n ) = j=1 j I{W i + W j > 0} = j=1 j I{Z i + Z j > 0}. j P θn (Z i + Z j > 0) = np θn (Z 1 > 0) + Since P θn (Z 1 > 0) = P θn (Z 1 θ n > θ n ) = 1 F ( θ n ) and ( ) n P θn (Z 1 + Z > 0). P θn (Z 1 + Z > 0) = P θn {(Z 1 θ n ) + (Z θ n ) > θ n } we conclude that = E P θn {Z 1 θ n > θ n (Z θ n ) Z } = µ (θ) = nf(θ) + Thus, because f( z) = f(z) by assumption, µ (0) = nf(0) + n(n 1) {1 F ( θ n z)}f(z) dz, ( ) n f( θ z)f(z) dz. 148 f (z) dz.
11 Letting we obtain µ (0) lim n τ(0) = lim n K = f (z) dz, 4{f(0) + (n 1)K} (n + 1)(n + 1) = K 1. Therefore, Theorem 9.9 shows that β n (θ n ) Φ( K 1 u α ). (9.4) Example 9.13 As in Example 9.1, suppose we have paired data (X 1, Y 1 ),..., (X n, Y n ) and Z i = Y i X i for all i. The Z i are independent and identically distributed with distribution function P (Z i z) = F (z θ) for some θ, where f(z) = F (z) exists and is symmetric about 0. Suppose that the variance of Z i is σ. Since the t-test (unknown variance) and z-test (known variance) have the same asymptotic properties, let s consider the z-test for simplicity. Then τ(θ) = σ for all θ. The relevant statistic is merely Z n, and the central limit theorem implies nzn /σ d N(0, 1) under the null hypothesis. Therefore, the z-test in this case rejects H 0 : θ = 0 in favor of H 1 : θ > 0 whenever Z n > u α σ/ n. The central limit theorem shows that n(zn θ n ) d N(0, 1) σ under the alternatives θ n = / n. Therefore, by Theorem 9.9, we obtain since µ (θ) = 1 for all θ. β n (θ n ) Φ( /σ u α ) (9.5) Before finding the asymptotic relative efficiency (ARE) of the Wilcoxon signed rank test and the t-test, we prove a lemma that enables this calculation. Suppose that for two tests, called test 1 and test, we use sample sizes m and n, respectively. We want m and n to tend to infinity together, an idea we make explicit by setting m = m k and n = n k for k = 1,,.... Suppose that we wish to apply both tests to the same sequence of alternative hypotheses θ 1, θ,.... As usual, we make the assumption that (θ k θ 0 ) times the square root of the sample size tends to a finite, nonzero limit as k. Thus, we assume mk (θ k θ 0 ) 1 and n k (θ k θ 0 ). 149
12 Then if Theorem 9.9 may be applied to both tests, define c 1 = lim µ 1(θ 0 )/τ 1 (θ 0 ) and c = lim µ (θ 0 )/τ (θ 0 ). The theorem says that β mk (θ k ) Φ { 1 c 1 u α } and β nk (θ k ) Φ { c u α }. (9.6) To find the ARE, then, Definition 9.11 specifies that we assume that the two limits in (9.6) are the same, which implies 1 c 1 = c, or n k c 1. m k Thus, the ARE of test 1 with respect to test equals (c 1 /c ). This result is summed up in the following lemma, which defines a new term, efficacy. c Lemma 9.14 to be For a test to which Theorem 9.9 applies, define the efficacy of the test c = lim n µ (θ 0 )/τ(θ 0 ). (9.7) Suppose that Theorem 9.9 applies to each of two tests, called test 1 and test. Then the ARE of test 1 with respect to test equals (c 1 /c ). Example 9.15 Using the results of Examples 9.1 and 9.13, we conclude that the efficacies of the Wilcoxon signed rank test and the t-test are 1 f (z) dz and 1 σ, respectively. Thus, Lemma 9.14 implies that the ARE of the signed rank test to the t-test equals ( 1σ f (z) dz). In the case of normally distributed data, we may verify without too much difficulty that the integral above equals (σ π) 1, so the ARE is 3/π Notice how close this is to one, suggesting that for normal data, we lose very little efficiency by using a signed rank test instead of a t-test. In fact, it may be shown that this asymptotic relative efficiency has a lower bound of However, there is no upper bound on the ARE in this case, which means that examples exist for which the t-test is arbitrarily inefficient compared to the signed rank test. Exercises for Section
13 Exercise 9.3 Assume that θ n 0 for some sequence θ 1, θ,.... In Example 9.1, Prove that R n E R n Var Rn d N(0, 1) under the assumption that for each n, the distribution of R n is determined by the parameter θ n. Note that this proof allows us to apply Theorem 9.9, since it covers both the contiguous alternatives and, by taking θ n = θ 0 for all n, the null hypothesis. Exercise 9.4 For the hypotheses considered in Examples 9.1 and 9.13, the sign test is based on the statistic N + = #{i : Z i > 0}. Since n(n + /n 1) d N(0, 1) under the null hypothesis, the sign test (with continuity correction) rejects H 0 when N + 1 u α n + n. (a) Find the efficacy of the sign test. Make sure to indicate how you go about verifying Equation (9.17). (b) Find the ARE of the sign test with respect to the signed rank test and the t-test. Evalaute each of these for the case of normal data. Exercise 9.5 Suppose X 1,..., X n are a simple random sample from a uniform (0, θ) distribution. We wish to test H 0 : θ = θ 0 against H 1 : θ > θ 0 at α =.05. Define Q 1 and Q 3 to be the first and third quartiles of the sample. Consider test A, which rejects when and test B, which rejects when Q 3 Q 1 θ 0 A n, X θ 0 B n. Based on the asymptotic distribution of X and the joint asymptotic distribution of (Q 1, Q 3 ), find the values of A n and B n that correspond with the test in (9.18). Then find the asymptotic relative efficiency of test A relative to test B. Exercise 9.6 Let P θ be a family of probability distributions indexed by a real parameter θ. If X P θ, define µ(θ) = E (X) and σ (θ) = Var (X). Now let θ 1, θ,... be a sequence of parameter values such that θ n θ 0 as n. Suppose that 151
14 E θn X +δ < M for all n for some positive δ and M. Also suppose that for each n, X n1,..., X nn are independent with distribution P θn and define X n = n X ni/n. Prove that if σ (θ 0 ) < and σ (θ) is continuous at the point θ 0, then n[xn µ(θ n )] d N(0, σ (θ 0 )) as n. Hint: a + b +δ +δ ( a +δ + b +δ). Exercise 9.7 Suppose X 1, X,... are independent exponential random variables with mean θ. Consider the test of H 0 : θ = 1 vs H 1 : θ > 1 in which we reject H 0 when where α =.05. X n 1 + u α n, (a) Derive an asymptotic approximation to the power of the test for a fixed sample size n and alternative θ. Tell where you use the result of Problem 9.6. (b) Because the sum of independent exponential random variables is a gamma random variable, it is possible to compute the power exactly in this case. Create a table in which you compare the exact power of the test against the alternative θ = 1. to the asymptotic approximation in part (a) for n {5, 10, 15, 0}. Exercise 9.8 Let X 1,..., X n be independent from Poisson (λ). Create a table in which you list the exact power along with approximations (9.0) and (9.1) for the test that rejects H 0 : λ = 1 in favor of H 1 : λ > 1 when n(xn λ 0 ) λ0 u α, where n = 0 and α =.05, against each of the alternatives 1.1, 1.5, and. Exercise 9.9 Let X 1,..., X n be an independent sample from an exponential distribution with mean λ, and Y 1,..., Y n be an independent sample from an exponential distribution with mean µ. Assume that X i and Y i are independent. We are interested in testing the hypothesis H 0 : λ = µ versus H 1 : λ > µ. Consider the statistic T n = (I i 1/)/ n, 15
15 where I i is the indicator variable I i = I(X i > Y i ). (a) Derive the asymptotic distribution of T n under the null hypothesis. (b) Use the Lindeberg Theorem to show that, under the local alternative hypothesis (λ n, µ n ) = (λ + n 1/ δ, λ), where δ > 0, n (I i ρ n ) L N(0, 1), where ρ n = λ n = λ + n 1/ δ nρn (1 ρ n ) λ n + µ n λ + n 1/ δ. (c) Using the conclusion of part (b), derive the asymptotic distribution of T n under the local alternative specified in (b). Exercise 9.10 Suppose X 1,..., X m is a simple random sample and Y 1,..., Y n is another simple random sample independent of the X i, with P (X i t) = t for t [0, 1] and P (Y i t) = (t θ) for t [θ, θ + 1]. Assume m/(m + n) ρ as m, n and 0 < θ < 1. Find the asymptotic distribution of m + n[g(y X) g(θ)]. 9.3 The Wilcoxon Rank-Sum Test Suppose that X 1,..., X m and Y 1,..., Y n are two independent simple random samples, with P (X i t) = P (Y j t + θ) = F (t) (9.8) for some continuous distribution function F (t) with f(t) = F (t). Thus, the distribution of the Y j is shifted by θ from the distribution of the X i. We wish to test H 0 : θ = 0 against H 1 : θ > 0. To do the asymptotics here, we will assume that n and m are actually both elements of separate sequences of sample sizes, indexed by a third variable, say k. Thus, m = m k and n = n k both go to as k, and we suppress the subscript k on m and n for convenience of notation. Suppose that we combine the X i and Y j into a single sample of size m + n. Define the Wilcoxon rank-sum statistic to be S k = Rank of Y j among combined sample. j=1 Letting Y (1),..., Y (n) denote the order statistics for the sample of Y j as usual, we may rewrite S k in the following way: S k = Rank of Y (j) among combined sample j=1 153
16 = = = ( j + #{i : Xi < Y (j) } ) j=1 n(n + 1) n(n + 1) + + j=1 j=1 m I{X i < Y (j) } m I{X i < Y j }. (9.9) Let N = m + n, and suppose that m/n ρ as k for some constant ρ (0, 1). First, we will establish the asymptotic behavior of S k under the null hypothesis. Define µ(θ) = E θ S k and τ(θ) = N Var S k. To evaluate µ(θ 0 ) and τ(θ 0 ), let Z i = n j=1 I{X i < Y j }. Then the Z i are identically distributed but not independent, and we have E θ0 Z i = n/ and Furthermore, E θ0 Z i Z j = which implies Therefore, and r=1 s=1 Var θ0 Z i = n 4 + n(n 1) Cov θ 0 (I{X i < Y 1 }, I{X i < Y }) = n n(n 1) n(n + ) =. 1 P θ0 (X i < Y r and X j < Y s ) = Cov θ0 (Z i, Z j ) = µ(θ 0 ) = n(n + 1) n(n 1) 4 + mn n(n 1) 4 n(n 1) 4 + n 3 n 4 = n 1. = n(n + 1) τ (θ 0 ) = Nm Var Z 1 + Nm(m 1) Cov (Z 1, Z ) = Nmn(n + ) Nm(m 1)n = Nmn(N + 1) np θ0 (X i < Y 1 and X j < Y 1 ),
17 Let θ 1, θ,... be a sequence of alternatives such that N(θ k θ 0 ) for a positive, finite constant. It is possible to show that under H 0 and N{Sk µ(θ 0 )} τ(θ 0 ) d N(0, 1) (9.30) N{Sk µ(θ k )} τ(θ 0 ) d N(0, 1) (9.31) under the alternatives {θ k } (see Exercise 9.11). By expression (9.30), the test based on S k with asymptotic level α rejects H 0 : θ = 0 in favor of H a : θ > 0 whenever S k µ(θ 0 ) + u α τ(θ 0 )/ N, or S k n(n + 1) mn(n + 1) + u α. 1 The Wilcoxon rank-sum test is sometimes referred to as the Mann-Whitney test. This alternative name helps to distinguish this test from the similarly named Wilcoxon signed rank test. To find the limiting power of the rank-sum test, we may use Theorem 9.9 to conclude that ( ) µ (θ 0 ) β k (θ k ) lim Φ u α. (9.3) k τ(θ 0 ) According to expression (9.3), we must evaluate µ (θ 0 ). To this end, note that Therefore, This gives P θ (X 1 < Y 1 ) = E θ {P θ (X 1 < Y 1 Y 1 )} = E θ F (Y 1 ) = = d dθ P θ(x 1 < Y 1 ) = µ (0) = mn F (y)f(y θ) dy F (y + θ)f(y) dy. f(y + θ)f(y) dy. f (y) dy. 155
18 Thus, the efficacy of the Wilcoxon rank-sum test is µ (θ 0 ) lim k τ(θ 0 ) = lim k mn 1 f (y) dy mnn(n + 1) = 1ρ(1 ρ) The asymptotic power of the test follows immediately from (9.3). f (y) dy. Exercises for Section 9.3 Exercise 9.11 (a) Prove expression (9.30) under the null hypothesis H 0 : θ = θ 0 (where θ 0 = 0). (b) Prove expression (9.31) under the sequence of alternatives θ 1, θ,... satisfying N(θk θ 0 ). Hint: In both (a) and (b), start by verifying either the Lindeberg condition or the Lyapunov condition. Exercise 9.1 Suppose Var X i = Var Y i = σ < and we wish to test the hypotheses H 0 : θ = 0 vs. H 1 : θ > 0 using the two-sample Z-statistic Y X σ 1 m + 1 n Note that this Z-statistic is s/σ times the usual T-statistic, where s is the pooled sample standard deviation, so the asymptotic properties of the T-statistic are the same as those of the Z-statistic. (a) Find the efficacy of the Z test. Justify your use of Theorem 9.9. (b) Find the ARE of the Z test with respect to the rank-sum test for normally distributed data. (c) Find the ARE of the Z test with respect to the rank-sum test if the data come from a double exponential distribution with f(t) = 1 λ e t/λ. (d) Prove by example that the ARE of the Z-test with respect to the rank-sum test can be arbitrarily close to zero.. 156
Maximum Likelihood Estimation
Chapter 8 Maximum Likelihood Estimation 8. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function
More informationMaximum Likelihood Estimation
Chapter 7 Maximum Likelihood Estimation 7. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More informationDA Freedman Notes on the MLE Fall 2003
DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationunder the null hypothesis, the sign test (with continuity correction) rejects H 0 when α n + n 2 2.
Assignment 13 Exercise 8.4 For the hypotheses consiere in Examples 8.12 an 8.13, the sign test is base on the statistic N + = #{i : Z i > 0}. Since 2 n(n + /n 1) N(0, 1) 2 uner the null hypothesis, the
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationIntroduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007
Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0
More informationHypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes
Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis
More informationSession 3 The proportional odds model and the Mann-Whitney test
Session 3 The proportional odds model and the Mann-Whitney test 3.1 A unified approach to inference 3.2 Analysis via dichotomisation 3.3 Proportional odds 3.4 Relationship with the Mann-Whitney test Session
More informationHYPOTHESIS TESTING: FREQUENTIST APPROACH.
HYPOTHESIS TESTING: FREQUENTIST APPROACH. These notes summarize the lectures on (the frequentist approach to) hypothesis testing. You should be familiar with the standard hypothesis testing from previous
More informationLoglikelihood and Confidence Intervals
Stat 504, Lecture 2 1 Loglikelihood and Confidence Intervals The loglikelihood function is defined to be the natural logarithm of the likelihood function, l(θ ; x) = log L(θ ; x). For a variety of reasons,
More informationSuggested solutions to written exam Jan 17, 2012
LINKÖPINGS UNIVERSITET Institutionen för datavetenskap Statistik, ANd 73A36 THEORY OF STATISTICS, 6 CDTS Master s program in Statistics and Data Mining Fall semester Written exam Suggested solutions to
More informationChapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1
Chapter 4 HOMEWORK ASSIGNMENTS These homeworks may be modified as the semester progresses. It is your responsibility to keep up to date with the correctly assigned homeworks. There may be some errors in
More informationSTAT 461/561- Assignments, Year 2015
STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and
More informationMISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30
MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)
More informationSTAT 135 Lab 5 Bootstrapping and Hypothesis Testing
STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationTopic 15: Simple Hypotheses
Topic 15: November 10, 2009 In the simplest set-up for a statistical hypothesis, we consider two values θ 0, θ 1 in the parameter space. We write the test as H 0 : θ = θ 0 versus H 1 : θ = θ 1. H 0 is
More informationChapter 4. Theory of Tests. 4.1 Introduction
Chapter 4 Theory of Tests 4.1 Introduction Parametric model: (X, B X, P θ ), P θ P = {P θ θ Θ} where Θ = H 0 +H 1 X = K +A : K: critical region = rejection region / A: acceptance region A decision rule
More informationChapter 3. Point Estimation. 3.1 Introduction
Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.
More informationLecture 1: Introduction
Principles of Statistics Part II - Michaelmas 208 Lecturer: Quentin Berthet Lecture : Introduction This course is concerned with presenting some of the mathematical principles of statistical theory. One
More informationFinal Examination Statistics 200C. T. Ferguson June 11, 2009
Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean
More informationSTAT 514 Solutions to Assignment #6
STAT 514 Solutions to Assignment #6 Question 1: Suppose that X 1,..., X n are a simple random sample from a Weibull distribution with density function f θ x) = θcx c 1 exp{ θx c }I{x > 0} for some fixed
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationf(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain
0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationMcGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper
McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationThe purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.
Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That
More informationGreene, Econometric Analysis (6th ed, 2008)
EC771: Econometrics, Spring 2010 Greene, Econometric Analysis (6th ed, 2008) Chapter 17: Maximum Likelihood Estimation The preferred estimator in a wide variety of econometric settings is that derived
More information4.5.1 The use of 2 log Λ when θ is scalar
4.5. ASYMPTOTIC FORM OF THE G.L.R.T. 97 4.5.1 The use of 2 log Λ when θ is scalar Suppose we wish to test the hypothesis NH : θ = θ where θ is a given value against the alternative AH : θ θ on the basis
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationSTAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015
STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis
More informationf(y θ) = g(t (y) θ)h(y)
EXAM3, FINAL REVIEW (and a review for some of the QUAL problems): No notes will be allowed, but you may bring a calculator. Memorize the pmf or pdf f, E(Y ) and V(Y ) for the following RVs: 1) beta(δ,
More informationStat 411 Lecture Notes 03 Likelihood and Maximum Likelihood Estimation
Stat 411 Lecture Notes 03 Likelihood and Maximum Likelihood Estimation Ryan Martin www.math.uic.edu/~rgmartin Version: August 19, 2013 1 Introduction Previously we have discussed various properties of
More informationLECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.
Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the
More informationChapter 7. Hypothesis Testing
Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function
More informationTheoretical Statistics. Lecture 23.
Theoretical Statistics. Lecture 23. Peter Bartlett 1. Recall: QMD and local asymptotic normality. [vdv7] 2. Convergence of experiments, maximum likelihood. 3. Relative efficiency of tests. [vdv14] 1 Local
More informationThe University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80
The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple
More informationStat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota
Stat 5102 Lecture Slides Deck 3 Charles J. Geyer School of Statistics University of Minnesota 1 Likelihood Inference We have learned one very general method of estimation: method of moments. the Now we
More informationSTA 260: Statistics and Probability II
Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition
More informationTheory of Statistics.
Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased
More information7. Estimation and hypothesis testing. Objective. Recommended reading
7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationTesting Statistical Hypotheses
E.L. Lehmann Joseph P. Romano Testing Statistical Hypotheses Third Edition 4y Springer Preface vii I Small-Sample Theory 1 1 The General Decision Problem 3 1.1 Statistical Inference and Statistical Decisions
More information14.30 Introduction to Statistical Methods in Economics Spring 2009
MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationChapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics
Chapter 6 Order Statistics and Quantiles 61 Extreme Order Statistics Suppose we have a finite sample X 1,, X n Conditional on this sample, we define the values X 1),, X n) to be a permutation of X 1,,
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationLecture 32: Asymptotic confidence sets and likelihoods
Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence
More informationCh. 5 Hypothesis Testing
Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,
More informationFinal Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.
1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically
More informationStatistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm
Statistics GIDP Ph.D. Qualifying Exam Theory Jan, 06, 9:00am-:00pm Instructions: Provide answers on the supplied pads of paper; write on only one side of each sheet. Complete exactly 5 of the 6 problems.
More informationMaximum Likelihood Tests and Quasi-Maximum-Likelihood
Maximum Likelihood Tests and Quasi-Maximum-Likelihood Wendelin Schnedler Department of Economics University of Heidelberg 10. Dezember 2007 Wendelin Schnedler (AWI) Maximum Likelihood Tests and Quasi-Maximum-Likelihood10.
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationStatistics and econometrics
1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample
More information1. Fisher Information
1. Fisher Information Let f(x θ) be a density function with the property that log f(x θ) is differentiable in θ throughout the open p-dimensional parameter set Θ R p ; then the score statistic (or score
More informationStatistics Ph.D. Qualifying Exam
Department of Statistics Carnegie Mellon University May 7 2008 Statistics Ph.D. Qualifying Exam You are not expected to solve all five problems. Complete solutions to few problems will be preferred to
More informationTABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1
TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1 1.1 The Probability Model...1 1.2 Finite Discrete Models with Equally Likely Outcomes...5 1.2.1 Tree Diagrams...6 1.2.2 The Multiplication Principle...8
More informationIntroduction to Estimation Methods for Time Series models Lecture 2
Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:
More informationLet us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided
Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or
More informationStatistics. Statistics
The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,
More informationLecture 17: Likelihood ratio and asymptotic tests
Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f
More information1 Likelihood. 1.1 Likelihood function. Likelihood & Maximum Likelihood Estimators
Likelihood & Maximum Likelihood Estimators February 26, 2018 Debdeep Pati 1 Likelihood Likelihood is surely one of the most important concepts in statistical theory. We have seen the role it plays in sufficiency,
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationTesting Statistical Hypotheses
E.L. Lehmann Joseph P. Romano, 02LEu1 ttd ~Lt~S Testing Statistical Hypotheses Third Edition With 6 Illustrations ~Springer 2 The Probability Background 28 2.1 Probability and Measure 28 2.2 Integration.........
More informationStatistics 135 Fall 2008 Final Exam
Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationCherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants
18.650 Statistics for Applications Chapter 5: Parametric hypothesis testing 1/37 Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009
More informationStatistical Inference: Estimation and Confidence Intervals Hypothesis Testing
Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire
More informationChapters 10. Hypothesis Testing
Chapters 10. Hypothesis Testing Some examples of hypothesis testing 1. Toss a coin 100 times and get 62 heads. Is this coin a fair coin? 2. Is the new treatment on blood pressure more effective than the
More informationParametric Inference
Parametric Inference Moulinath Banerjee University of Michigan April 14, 2004 1 General Discussion The object of statistical inference is to glean information about an underlying population based on a
More informationMathematics Ph.D. Qualifying Examination Stat Probability, January 2018
Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from
More informationIntroducing the Normal Distribution
Department of Mathematics Ma 3/13 KC Border Introduction to Probability and Statistics Winter 219 Lecture 1: Introducing the Normal Distribution Relevant textbook passages: Pitman [5]: Sections 1.2, 2.2,
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationexp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1
4 Hypothesis testing 4. Simple hypotheses A computer tries to distinguish between two sources of signals. Both sources emit independent signals with normally distributed intensity, the signals of the first
More informationProblems ( ) 1 exp. 2. n! e λ and
Problems The expressions for the probability mass function of the Poisson(λ) distribution, and the density function of the Normal distribution with mean µ and variance σ 2, may be useful: ( ) 1 exp. 2πσ
More informationSTA 732: Inference. Notes 4. General Classical Methods & Efficiency of Tests B&D 2, 4, 5
STA 732: Inference Notes 4. General Classical Methods & Efficiency of Tests B&D 2, 4, 5 1 Going beyond ML with ML type methods 1.1 When ML fails Example. Is LSAT correlated with GPA? Suppose we want to
More informationStatistical Theory MT 2006 Problems 4: Solution sketches
Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine
More informationStatistical Theory MT 2007 Problems 4: Solution sketches
Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the
More informationThe outline for Unit 3
The outline for Unit 3 Unit 1. Introduction: The regression model. Unit 2. Estimation principles. Unit 3: Hypothesis testing principles. 3.1 Wald test. 3.2 Lagrange Multiplier. 3.3 Likelihood Ratio Test.
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationHypothesis Testing: The Generalized Likelihood Ratio Test
Hypothesis Testing: The Generalized Likelihood Ratio Test Consider testing the hypotheses H 0 : θ Θ 0 H 1 : θ Θ \ Θ 0 Definition: The Generalized Likelihood Ratio (GLR Let L(θ be a likelihood for a random
More informationStatistics Ph.D. Qualifying Exam: Part I October 18, 2003
Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer
More informationProbability and Statistics
Probability and Statistics 1 Contents some stochastic processes Stationary Stochastic Processes 2 4. Some Stochastic Processes 4.1 Bernoulli process 4.2 Binomial process 4.3 Sine wave process 4.4 Random-telegraph
More informationSummary of Chapters 7-9
Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two
More informationFor iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions.
Large Sample Theory Study approximate behaviour of ˆθ by studying the function U. Notice U is sum of independent random variables. Theorem: If Y 1, Y 2,... are iid with mean µ then Yi n µ Called law of
More informationNonparametric hypothesis tests and permutation tests
Nonparametric hypothesis tests and permutation tests 1.7 & 2.3. Probability Generating Functions 3.8.3. Wilcoxon Signed Rank Test 3.8.2. Mann-Whitney Test Prof. Tesler Math 283 Fall 2018 Prof. Tesler Wilcoxon
More informationSolution: First note that the power function of the test is given as follows,
Problem 4.5.8: Assume the life of a tire given by X is distributed N(θ, 5000 ) Past experience indicates that θ = 30000. The manufacturere claims the tires made by a new process have mean θ > 30000. Is
More informationMATH4427 Notebook 2 Fall Semester 2017/2018
MATH4427 Notebook 2 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
More informationPractice Problems Section Problems
Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationRelative efficiency. Patrick Breheny. October 9. Theoretical framework Application to the two-group problem
Relative efficiency Patrick Breheny October 9 Patrick Breheny STA 621: Nonparametric Statistics 1/15 Relative efficiency Suppose test 1 requires n 1 observations to obtain a certain power β, and that test
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More information