Tests and Their Power

Similar documents
F79SM STATISTICAL METHODS

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Maximum-Likelihood Estimation: Basic Ideas

TUTORIAL 8 SOLUTIONS #

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

STATISTICS SYLLABUS UNIT I

Lecture 21. Hypothesis Testing II

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Summary of Chapters 7-9

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

MATH 240. Chapter 8 Outlines of Hypothesis Tests

STAT 830 Hypothesis Testing

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

Evaluating Hypotheses

Lecture 12 November 3

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

14.30 Introduction to Statistical Methods in Economics Spring 2009

simple if it completely specifies the density of x

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

ECE531 Lecture 4b: Composite Hypothesis Testing

Hypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques

LECTURE 5 HYPOTHESIS TESTING

Mathematical Statistics

hypothesis testing 1

4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49

STAT 830 Hypothesis Testing

Testing Statistical Hypotheses

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Topic 15: Simple Hypotheses

Recall the Basics of Hypothesis Testing

Hypothesis Testing - Frequentist

Hypothesis testing: theory and methods

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4

Hypothesis Testing Chap 10p460

Ch. 5 Hypothesis Testing

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Physics 403. Segev BenZvi. Classical Hypothesis Testing: The Likelihood Ratio Test. Department of Physics and Astronomy University of Rochester

Hypothesis Testing. Testing Hypotheses MIT Dr. Kempthorne. Spring MIT Testing Hypotheses

Chapter 4. Theory of Tests. 4.1 Introduction

Practice Problems Section Problems

A Note on Hypothesis Testing with Random Sample Sizes and its Relationship to Bayes Factors

CHAPTER EVALUATING HYPOTHESES 5.1 MOTIVATION

Institute of Actuaries of India

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing

A Very Brief Summary of Statistical Inference, and Examples

6.4 Type I and Type II Errors

DA Freedman Notes on the MLE Fall 2003

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

Comparison of Accident Rates Using the Likelihood Ratio Testing Technique

Testing Statistical Hypotheses

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?

Lecture 4: Probability and Discrete Random Variables

Lecture 4: Statistical Hypothesis Testing

CH.9 Tests of Hypotheses for a Single Sample

INTERVAL ESTIMATION AND HYPOTHESES TESTING

H 2 : otherwise. that is simply the proportion of the sample points below level x. For any fixed point x the law of large numbers gives that

ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules

Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma

Lecture 2. G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1

Topic 10: Hypothesis Testing

Statistical Inference

Introduction to Statistical Inference

8 Testing of Hypotheses and Confidence Regions

FYST17 Lecture 8 Statistics and hypothesis testing. Thanks to T. Petersen, S. Maschiocci, G. Cowan, L. Lyons

Chapter 9: Hypothesis Testing Sections

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Topic 17: Simple Hypotheses

Introductory Econometrics. Review of statistics (Part II: Inference)

z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests

Lecture 8: Information Theory and Statistics

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Topic 10: Hypothesis Testing

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,

Nuisance parameters and their treatment

Detection Theory. Composite tests

Chapter 7. Hypothesis Testing

EC2001 Econometrics 1 Dr. Jose Olmo Room D309

CSE 312 Final Review: Section AA

280 CHAPTER 9 TESTS OF HYPOTHESES FOR A SINGLE SAMPLE Tests of Statistical Hypotheses

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Question Points Score Total: 76

Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Master s Written Examination

Fundamental Probability and Statistics

On likelihood ratio tests

Application of Variance Homogeneity Tests Under Violation of Normality Assumption

17. Convergence of Random Variables

Lecture 8: Information Theory and Statistics

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

CHI SQUARE ANALYSIS 8/18/2011 HYPOTHESIS TESTS SO FAR PARAMETRIC VS. NON-PARAMETRIC

BTRY 4090: Spring 2009 Theory of Statistics

Parametric Models: from data to models

Transcription:

Tests and Their Power Ling Kiong Doong Department of Mathematics National University of Singapore 1. Introduction In Statistical Inference, the two main areas of study are estimation and testing of hypotheses. In this workshop, we confine ourselves to the ~rea of testing of hypotheses. Basic concepts needed in testing of hypotheses will be introduced in the next section. Section 3 considers the problems of testing a simple null hypothesis against a simple alternative hypothesis. The well-known Neyman Pearson Lemma will also be discussed. Applications of the Neyman Pearson Lemma to some test problems which involve composite hypotheses are given in Section 4. Finally, in Section 5, we introduce the likelihood ratio method of finding a test. 2. Basic Concepts A standard form of problems in testing of hypotheses is Test a null hypothesis, usually denoted by H 0, against an alternative hypothesis, usually denoted by H 1, where H 0 and -H 1 are hypotheses concerning the distribution of a random variable X. A hypothesis is said to be simple if it completely determines the distribution of X. Hypotheses which are not simple are called composite hypotheses. For example, if X has a normal distribution with mean p, and variable u 2, i.e., X"" N(p,, a 2 ), then the hypothesis that p, = 0 and a 2 = 1 is simple, while the hypothesis that p, = 5 is composite. Lecture given at the Workshop on Hypothesis Testing and Non-Parametric Statistics for school teachers, 7 September 1987 18

In many practical situation, the probability density function (pdf) of a random variable X is partly known in form, but involves some unknown characteristics or parameters. A hypothesis which is concerned with the parameters of the distribution of X is referred to as a parametric hypothesis. Hypotheses that are not parametric are said to be non-parametric.. For example, the two hypotheses given in the preceding paragraph are parametric. The hypothesis that X is normally distributed and the hypothesis that X has a distribution which is symmetric about the origin are examples of non-parametric hypotheses. In this talk we shall deal with only those test problems which involve parametric hypotheses. Now let X= (X 1, X 2,, Xn) be a random sample of size n drawn on a random variable X whose distribution is indexed by parameter(s) () E 0, the parameter space. Suppose that we wish to test the null hypothesis Ho : () E Wo against the alternative hypothesis Hl : () E wl = n \ Wo. That is, when we are given sample observations x = (x 1, x 2,, xn), we want to take one of the following two actions : (i) to accept H 0, or equivalently to reject H 1 ; or (ii) to accept H 1, or equivalently to reject H 0 Let us assume that a certain test procedure or rule is used so that to each sample point x in Xx the sample space of X which is usually taken to be the n-dimensional Euclidean space R n, one and only one of the above two actions is taken. Denote by C the set of those sample points which leads to a rejection of H 0 and A = Xx \C. The regions C and A are called the critical (or rejection) and the acceptance region of the test precedure respectively. Therefore finding a test procedure is equivalent to partitioning Xx into two non-overlapping regions C and A. Once a decision is made, then we may commit either a Type I error or a Type II error which are best illustrated by the following table. ~n e Accept Ho Reject H 0 Ho is true a correct decision Type I Error H1 is true Type II Error a correct decision 19

Let us consider now the case where both H 0 and H 1 are simple. Define : a= P(Committing a Type I error), and {3* = P(Committing a Type II error). Thus a is the probability of wrongly rejecting H 0 and {3* is the probability of wrongly accepting H 0 Ideally, it is natural to seek a test procedure which will give the values of a and {3* as small as possible. Unfortunately that is not possible in most cases. Thus, instead, we look for a test procedure which has a given a and a small {3* or equivalently a large {3 = 1-{3*. ({3 gives the probability of correctly rejecting H 0.) In the literature, a is called the size or the significance level ~nd {3 is called the power of the test procedure. The above consideration leads to the following definitions of a test. Definition 1 : A non-randomized test is a function >: Xx --+ {0, 1} such that >(x) = { ~ ifx E C if x EA. Definition 2 : A randomized test is a function > : Xx --+ [0, 1]. For a given randomized test >, if 0 < >(x) = p < 1, then the probability of rejecting H 0 is p. That is, if x 0 is a sample point with >(x 0 ) = p, then we flip a biased coin with P(head) = p. If a head appears then we reject H 0 and accept H 0 if a tail appears. (Sometimes, we consult a table of random digits instead of flipping a biased coin.) Also we note that a= E(</>(X) I Ho) and {3 = E( >(X) I H1 ). 20

When H 0 is composite, we define and when H 1 is composite we define a= sup (E( (X) 18)) 9Ewo {3(8) = E( (X) 18), for 8.E w 1 For composite H 0, a is the size of the test and for composite H 1, {3(8 ) is called the power function of the test. The functional value {3(0. 1 ) at a point 8 1 E w 1 is called the power of the test at 8 = 8 1 3. Neyman-Pearson Lemma This section deals with a test problem whose null hypothesis and alternative hypothesis are simple. Let us begin with a definition.. Definition 3 : A test is said to be a most powerful {MP} test of size a in testing a simple H 0 against a simple H 1 if (i) E( (X) I Ho) = a; and (ii) E( (X) I H 1 ) ;:::: E( 4>* (X) I H 1 ) for any other test * satisfying (i). To every test, we can associate with it a pair ( a,{3) where a and {3 are respectively the size and the power of. Now let N be the set of points (a, {3) such that there exists a test whose associated pair is (a, {3). Then it is not difficult to see that (i) (0, 0) E N and (1, 1) EN; (ii) (a 1,{3 1 ) EN and (a 2,{3 2 ) EN, and 0 ~ t ~ 1, then t(a 1,{3I) + (1- t)(a 2,{3 2 ) EN, (i.e., N is convex.); (iii) (a, {3) E N implies (1 - respect to the point (i, i ).); a~d (iv) N is closed. a, 1 - {3) E N. (i.e., N is symmetric with 21

Graphically, N has the shape of the following diagram. The point A (in the preceding diagram) corresponds to amp test of size a 0 From the diagram it is also clear that if </> is a MP test of size a and power (3, then a < (3. For a given significance level a, our main objective is to determine a MP test of size a. The solution.to this problem is given by the following Neyman-Pearson Lemma (see page 118 of Zack [5].) Let / 0 and / 1 denote the pdf of X under H 0 and H 1 respectively. Then for testing H 0 against H 1 we have the following: (a) Any test of the form (X)= { ~ af /1 (x) > kfo (x) if /1 (x) = kfo (x) af /1 (x) < kfo (x). for some 0 ~ k < oo and 0 ~ 1 ~ 1 is MP relative to all tests of its SIZeS. (b) (Existence) For testing H 0 against H 1 at level of significance a, there exists ka, 0 ~ ka < oo and Ia, 0 ~ Ia ~ 1 such that the corresponding test of the form in (a) is MP of s.ize a. (c) (Uniqueness) If a te6t </>* is MP of size a, then it is of the form in (a), except perhaps on the set {xi / 1 (x) = k/ 0 (x)}; unless there exists a test of size smaller than a and power 1. For a proof, the readers are ;eferred to pages 119-121 of Zack [5]. When X is absolutely continuous, the MP test of size a is a non-randomized 22

one. However, when X is discrete, the MP test of size a is a randomized test for most values of a. The Neyman Pearson Lemma is useful not only for testing simple H 0 against simple H 1, it is also very helpful in finding a uniformly most powerful (UMP) test (a term to be defined in the next section) for testing a simple H 0 against a one-sided composite H 1 Many examples of finding MP tests using the Neyman Pearson Lemma may be found in many books on Elementary Mathematical Statistics. To end this section, we give the following Example 1 : Let the pdf of a random variable X be given by 1 f(x 18) = exp{ -x/8}, X ~ 0, 0 where 8 > 0., is the mean of X. On the basis of a random sample X, obtain the MP test for testing H 0 : 8 = 1 against H 1 : 8 = 2 at level of significance a, (0 < a < 1). ' Solution : Here / 0 (x) = exp{ -Ex.}, x, ~ 0, i = 1, 2,, n; and /I{x)= (i)n exp{-exd2}, x, ~O,i=1,2,..,n. Hence, by applying the Neyman Pearson Lemma, the MP test > of size a is given by 1 if (11 (x)/ fo (x)) ~ k (x) = { 0 if (fdx)/fo(x)) < k. which is equivalent to (x) = {: where k* is a constant to be determined so that n E( >(X) 18 = 1) = P(LX, ~ k*) =a. 23 i= 1

Now we note when 0 = 1, then 2 I:~= 1 X, has chi-square distribution with 2n degrees of freedom. Therefore k* = ~X~ (2n), a value which can be obtained from Statistical Tables [2]. For example when a= 0.05, n = 5, we have x~.os (10) = 18.307. 4. Uniformly Most Powerful Tests When the alternative hypothesis H 1 is composite, the notion of a MP test does not apply. Instead, we introduce the concept of a uniformly most powerful (UMP) test. Definition 4: A test > is called a UMP test of size a if (i) SUPoewo E( >(X) I 0) =a; and (ii) for each 0 E w 1, E( >(X) I 0) 2: E( >* (X) I 0) for any test >* satisfying (i). When Ho is simple and H 1 is composite, we can regard H 1 as a union of simple hypotheses. In some cases we can obtain a UMP test by applying the Neyman Pearson Lemma. To do this, we consider the problem of testing the simple H 0 against a typical simple H; : 0 = 0 1 where 0 1 is a typical member of w 1 If the MP test of size a for testing H 0 against H; does not depend on the value of 0 1 in w 1, then this test is MP for testing H 0 against every simple H~ in H 1 Hence this MP test of size a becomes a UMP test of size a for testing H 0 against H 1 Example 2 : Let X be the random variable considered in Example 1. Here, we wish to establish a UMP test of size a for testing H 0 : 0 = 1 against H 1 : 0 > 1. Solution : Note that w 0 element in w 1 If we consider the problem of testing = {1} and w 1 = {0 I 0 > 1}. Let 0 1 be a fixed H 0 : 0 = 1 against H; : 0 = 0 1, (here H; is simple), then we can apply the Neyman Pearson Lemma. It leads to the following MP test of size a (x) = {: 24

(See Example 1 for the derivation.) This test is independent of the value of 0 1 and hence it is the UMP test for testing H 0 against H 1 Remark: (1) If H1 in Example 2 is replaced by H2 : 0 E w2 = {0 I 0 > 02}, wh~re 0 2 is a given real number> 1, then the UMP test of size a: for testing l, H 0 against H 2 is the one given in Example 2. (2) If H 1 is replaced by H 3 : 0 E w 3 = {0 I 0 < 0 3 }, where 0 3 is a given positive number < 1, then the UMP test of size a: for testing H 0 against H 3 is (3) When H 1 is replaced by H 4 : 0 f. 1(H 4 is a two-sided hypothesis), then a UMP test of size a: for testing H 0 against H 4 does not exist. Consider now the case in which both H 0 : 0 E w 0 and H 1 : 0 E w 1 are one-sided composite hypotheses. Assume that for any 0 0 E w 0 and any 61 E W1 we have 0 0 < 01. Let 0~ = sup 0 Ewo 0. (In the case 0 0 > 01, for every 0 0 E w 0 and every 0 1 E w 1, we take 0~ = infoewo 0.) If is the UMP test of size a: for testing H; : 0 E 0~ against H 1, and if sup E( (X) I 0) =a: OEwo then is the UMP test of size a: for testing H 0 against H 1. For additional reading material concerning UMP tests, readers are referred to the books by Roussas [3] and Zacks [5]. 5. Likelihood Ratio Method The Neyman Pearson Lemma provides us a method of finding the MP test for testing a simple H 0 against a sim}>le H 1 In some cases an application of the Lemma will also lead us to obtain the UMP tests. In order to solve test problems which do not admit UMP tests, we introduce in this section the likelihood ratio method of constructing tests. The method uses a test statistic which is analogous to likelihood ratio / 1 (x) j / 0 (x) used 25

in the Neyman Pearson Lemma. For this reason, some authors refer to this method as the generalized likelihood ratio method. When both null and alternative hypotheses are simple, this method of constructing tests should not be applied, for the resulting test obtained may not be MP. Instead, we should apply the Neyman Pearson Lemma. Consider the problem of testing H 0 : tj E w 0 against H 1 : tj E w 1 Suppose that n = w 0 U w 1 contains three or more elements, i.e., at least one of H 0 and H 1 is composite. The likelihood ratio method uses the test statistic,.a(x), usually known as the likelihood ratio statistic, which is defined by Since Wo Note that and.a(x) = supbewo f(x I 0). supbeo f(xltj) c n, it is clear that 0 :::;.A(x) :::; 1 for every sample point X E Xx. sup f(x I 0) = f(x 18..,) 8Ewo sup f(x I 0) = f(x 18) 8EO where 8.., and 8 are the maximum likelihood estimates of 0 under w 0 and n respectively. When Ho is not true, the denominator, /(x 18)' in. the definition of.a(x) tends to be much larger than the numerator, f(x 18.., ); and so the value of.a(x) tends to be small. Thus the method suggests the following test: (x) = { 1, if.a(x) :::; A0 0, if.a(x) > Ao where A 0, 0 <.A 0 < 1, is a constant to be determined so that sup E( (X) I tj) = a. BE:wo This resultant test 4> is called the likelihood ratio test. The value of Ao depends on the size, a. Generally, for any given a, we require the knowledge of the distribution of the likelihood ratio statistic under H 0 to determine A 0 In most cases the inequality.a(x) :::;.A 0 used in defining the likelihood ratio test 4> may be reduced to a simpler and equivalent inequality, say 26

T(x) ~ T 0, and the distribution of T(X) may be easier to derive than that of.x(x) when H 0 is true. Likelihood ratio tests are not necessarily UMP. But they have some desirable properties. (We will not discuss them here.) A possible drawback of the likelihood ratio method is, perhaps, that the maximum likelihood estimates 0"' and 0 may be difficult to obtain. Another drawback is that the distribution of the likelihood ratio statistic.x(x) or its equivalent statistic T(X), (needed for the determination of.x 0 or T 0 ), may not be available in the literature and may be difficult to derive. Fortunately, in most situations, the asymptotic distribution (i.e., when n tends to infinity) of -2ln.X(X) under H 0 has been proven to follow a chi-square distribution. (This result will be stated without proof as a theorem.) Therefore for sufficiently large samples, an approximate value of the cut point.x 0 be easily obtained. Theorem : For testing H 0 : () E w 0 against H 1 : () E w 1, let the dimension of 0 = w 0 U w 1 be r and the dimension of w 0 be r 1 ( < r). When H 0 is true, then under certain regularity conditions, the asymptotic distribution of -2ln.X(X), as n tends to infinity, is distributed like a chi-square distribution with r - r 1 degrees of freedom. Note: For the regularity conditions about the distribution of the random variable X and a proof to the above theorem, the readers are referred to Wilks [4]. To end the talk we give an example of the determination of the degrees of freedom of the asymptotic distribution of -2ln.X(X). Example 3 : Let X,, i = 1, 2,, k(~ 2) be k independent random variables. The pdf of each X, is known but depends on the unknown mean p,, and variance u[. Consider now the null hypothesis Ho : u~ and the alternative hypothesis Here we have = u; = = uz H 1 : not all u; 's are equal. 0 = { (J.Ll,, J.Lk, u~,, uz) I - oo < J.L < +oo, ui > 0, i = 1,, k} and Wo = {(J.Ll,, J.Lk, u 2,, u 2 ) 1-oo < J.L < +oo, u 2 > 0, i = 1,, k}. can 27

Clearly the dimension of n is 2k (i.e. r = 2k) and the dimension of Wo is k + 1, i.e., r 1 = k + 1. Therefore for large samples sizes, the distribution of -2ln.A(X) is approximately distributed as the chi-square distribution with k- 1 degrees of freedom. Remark : H 0 in Example 3 may be rewritten as H~ "/2 = "/3 = = "'~o = 1, where "'i = a;/ a~. Then a~ > 0, "ti > 0, i = 1,, k, ;' = 2,, k}. Let 0 = (J.L 1,, J.L~o, a~, 1 2,, I~<). Then to describe H~, we require k - 1 components of 0, i.e., 1 2,, "/1<, to take fixed values. In general, the number of components that take fixed values in describing H 0 is the number of degrees of freedom of the asymptotic distribution of -2ln.A(X). References [1] E. L. Lehmann, Testing Statistical Hypotheses, John Wiley, 1959. [2] J. Murdoch and J. A. Barnes, Statistical Tables, (second edition), Macmillan, 1970. [3] G. G. Roussas, A First Course in Mathematical Statistics, Addison Wesley, 1973. [4] S. Wilks, Mathematical Statistics, John Wiley, 1962. [5] S. Zacks, Parametric Statistical Inference, Pergamon, 1981. 28