STAT 830 Hypothesis Testing

Similar documents
STAT 830 Hypothesis Testing

STAT 801: Mathematical Statistics. Hypothesis Testing

Hypothesis Testing. A rule for making the required choice can be described in two ways: called the rejection or critical region of the test.

Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

Chapter 9: Hypothesis Testing Sections

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

Lecture 12 November 3

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

TUTORIAL 8 SOLUTIONS #

Lecture notes on statistical decision theory Econ 2110, fall 2013

Topic 15: Simple Hypotheses

Hypothesis Testing. Testing Hypotheses MIT Dr. Kempthorne. Spring MIT Testing Hypotheses

ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4

F79SM STATISTICAL METHODS

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Chapters 10. Hypothesis Testing

Chapter 4. Theory of Tests. 4.1 Introduction

8: Hypothesis Testing

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

CSE 312 Final Review: Section AA

Topic 10: Hypothesis Testing

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49

Topic 10: Hypothesis Testing

Chapters 10. Hypothesis Testing

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.

Hypothesis Testing - Frequentist

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Mathematical Statistics

simple if it completely specifies the density of x


8 Testing of Hypotheses and Confidence Regions

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary

ECE531 Lecture 4b: Composite Hypothesis Testing

Summary of Chapters 7-9

Lecture 21. Hypothesis Testing II

λ(x + 1)f g (x) > θ 0

14.30 Introduction to Statistical Methods in Economics Spring 2009

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

Spring 2012 Math 541B Exam 1

Theory of Statistical Tests

Transformations and Expectations

Ch. 5 Hypothesis Testing

Chapter 7. Hypothesis Testing

Lecture 16 November Application of MoUM to our 2-sided testing problem

Hypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

STAT 830 Bayesian Estimation

Hypothesis testing: theory and methods

Stat 5421 Lecture Notes Fuzzy P-Values and Confidence Intervals Charles J. Geyer March 12, Discreteness versus Hypothesis Tests

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

Fundamental Probability and Statistics

On the GLR and UMP tests in the family with support dependent on the parameter

6.4 Type I and Type II Errors

Basic Concepts of Inference

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

Statistics 3858 : Maximum Likelihood Estimators

Introduction to Machine Learning. Lecture 2

Lecture 10: Generalized likelihood ratio test

Topic 3: Hypothesis Testing

Two examples of the use of fuzzy set theory in statistics. Glen Meeden University of Minnesota.

Ch3. Generating Random Variates with Non-Uniform Distributions

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

Topic 17: Simple Hypotheses

STAT 830 Non-parametric Inference Basics

Lecture 1: August 28

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

A Note on Hypothesis Testing with Random Sample Sizes and its Relationship to Bayes Factors

Statistical Preliminaries. Stony Brook University CSE545, Fall 2016

CHAPTER 8. Test Procedures is a rule, based on sample data, for deciding whether to reject H 0 and contains:

STAT 461/561- Assignments, Year 2015

ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence

Frequentist Statistics and Hypothesis Testing Spring

14.30 Introduction to Statistical Methods in Economics Spring 2009

Statistical Data Analysis Stat 3: p-values, parameter estimation

Testing Statistical Hypotheses

On the Inefficiency of the Adaptive Design for Monitoring Clinical Trials

Optimal Tests of Hypotheses (Hogg Chapter Eight)

Master s Written Examination - Solution

Lecture 4: Probability and Discrete Random Variables

Quick Tour of Basic Probability Theory and Linear Algebra

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Institute of Actuaries of India

Uniformly and Restricted Most Powerful Bayesian Tests

Statistics 135 Fall 2008 Final Exam

Hypothesis Testing Chap 10p460

Statistics 100A Homework 5 Solutions

Statistics for the LHC Lecture 1: Introduction

Transcription:

STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30

Purposes of These Notes Describe hypothesis testing Discuss Type I and Type II error. Discuss level and power Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 2 / 30

Hypothesis Testing Hypothesis testing: a statistical problem where you must choose, on the basis of data X, between two alternatives. Formalized as problem of choosing between two hypotheses: H o : θ Θ 0 or H 1 : θ Θ 1 where Θ 0 and Θ 1 are a partition of the model P θ ;θ Θ. That is Θ 0 Θ 1 = Θ and Θ 0 Θ 1 =. A rule for making the required choice can be described in two ways: 1 In terms of rejection or critical region of the test. R = {X : we choose Θ 1 if we observe X} 2 In terms of a function φ(x) which is equal to 1 for those x for which we choose Θ 1 and 0 for those x for which we choose Θ 0. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 3 / 30

Hypothesis Testing Each φ corresponds to a unique rejection region R φ = {x : φ(x) = 1}. Neyman Pearson approach treats two hypotheses asymmetrically. Hypothesis H o referred to as the null hypothesis (traditionally the hypothesis that some treatment has no effect). Definition: The power function of a test φ (or the corresponding critical region R φ ) is π(θ) = P θ (X R φ ) = E θ (φ(x)) Interested in optimality theory, that is, the problem of finding the best φ. A good φ will evidently have π(θ) small for θ Θ 0 and large for θ Θ 1. There is generally a trade off which can be made in many ways, however. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 4 / 30

Simple versus Simple testing Finding a best test is easiest when the hypotheses are very precise. Definition: A hypothesis H i is simple if Θ i contains only a single value θ i. The simple versus simple testing problem arises when we test θ = θ 0 against θ = θ 1 so that Θ has only two points in it. This problem is of importance as a technical tool, not because it is a realistic situation. Suppose that the model specifies that if θ = θ 0 then the density of X is f 0 (x) and if θ = θ 1 then the density of X is f 1 (x). How should we choose φ? To answer the question we begin by studying the problem of minimizing the total error probability. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 5 / 30

Error Types Type I error: the error made when θ = θ 0 but we choose H 1, that is, X R φ. Type II error: when θ = θ 1 but we choose H 0. The level of a simple versus simple test is or α = P θ0 (We make a Type I error) α = P θ0 (X R φ ) = E θ0 (φ(x)) Other error probability denoted β is β = P θ1 (X R φ ) = E θ1 (1 φ(x)). Minimize α+β, the total error probability given by α+β = E θ0 (φ(x))+e θ1 (1 φ(x)) = [φ(x)f 0 (x)+(1 φ(x))f 1 (x)]dx Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 6 / 30

Proof of NP lemma Problem: choose, for each x, either the value 0 or the value 1, in such a way as to minimize the integral. But for each x the quantity is between f 0 (x) and f 1 (x). φ(x)f 0 (x)+(1 φ(x))f 1 (x) To make it small we take φ(x) = 1 if f 1 (x) > f 0 (x) and φ(x) = 0 if f 1 (x) < f 0 (x). It makes no difference what we do for those x for which f 1 (x) = f 0 (x). Notice: divide both sides of inequalities to get condition in terms of likelihood ratio f 1 (x)/f 0 (x). Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 7 / 30

Bayes procedures, in disguise Theorem For each fixed λ the quantity β +λα is minimized by any φ which has { 1 f 1 (x) f φ(x) = 0 (x) > λ f 0 1 (x) f 0 (x) < λ Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 8 / 30

Neyman-Pearson framework Neyman and Pearson suggested that in practice the two kinds of errors might well have unequal consequences. Suggestion: pick the more serious kind of error, label it Type I. Require rule to hold probability α of a Type I error to be no more than some prespecified level α 0. α 0 is typically 0.05, chiefly for historical reasons. Neyman-Pearson approach: minimize β subject to the constraint α α 0. Usually this is really equivalent to the constraint α = α 0 (because if you use α < α 0 you could make R larger and keep α α 0 but make β smaller. For discrete models, however, this may not be possible. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 9 / 30

Binomial example: effects of discreteness Example: Suppose X is Binomial(n,p) and either p = p 0 = 1/2 or p = p 1 = 3/4. If R is any critical region (so R is a subset of {0,1,...,n}) then P 1/2 (X R) = k 2 n for some integer k. Example: to get α 0 = 0.05 with n = 5: possible values of α are 0,1/32 = 0.03125,2/32 = 0.0625, etc. Possible rejection regions for α 0 = 0.05: Region α β R 1 = 0 1 R 2 = {x = 0} 0.03125 1 (1/4) 5 R 3 = {x = 5} 0.03125 1 (3/4) 5 So R 3 minimizes β subject to α < 0.05. Raise α 0 slightly to 0.0625: possible rejection regions are R 1, R 2, R 3 and R 4 = R 2 R 3. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 10 / 30

Discreteness, test functions First three have same α and β as before while R 4 has α = α 0 = 0.0625 an β = 1 (3/4) 5 (1/4) 5. Thus R 4 is optimal! Problem: if all trials are failures optimal R chooses p = 3/4 rather than p = 1/2. But: p = 1/2 makes 5 failures much more likely than p = 3/4. Problem is discreteness. Solution: Expand set of possible values of φ to [0,1]. Values of φ(x) between 0 and 1 represent the chance that we choose H 1 given that we observe x; the idea is that we actually toss a (biased) coin to decide! This tactic will show us the kinds of rejection regions which are sensible. In practice: restrict our attention to levels α 0 for which best φ is always either 0 or 1. In the binomial example we will insist that the value of α 0 be either 0 or P θ0 (X 5) or P θ0 (X 4) or... Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 11 / 30

Binomial example: n = 3 4 possible values of X and 2 4 possible rejection regions. Table of levels for each possible rejection region R: R α R α 0 {3}, {0} 1/8 {0,3} 2/8 {1}, {2} 3/8 {0,1}, {0,2}, {1,3}, {2,3} 4/8 {0,1,3}, {0,2,3} 5/8 {1,2} 6/8 {0,1,2}, {1,2,3} 7/8 {0,1,2,3} 1 Best level 2/8 test has rejection region {0,3}, β = 1 [(3/4) 3 +(1/4) 3 ] = 36/64. Best level 2/8 test using randomization rejects when X = 3 and, when X = 2 tosses a coin with P(H) = 1/3, then rejects if you get H. Level is 1/8+(1/3)(3/8) = 2/8; probability of Type II error is β = 1 [(3/4) 3 +(1/3)(3)(3/4) 2 (1/4)] = 28/64. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 12 / 30

Test functions Definition: A hypothesis test is a function φ(x) whose values are always in [0,1]. If we observe X = x then we choose H 1 with conditional probability φ(x). In this case we have π(θ) = E θ (φ(x)) α = E 0 (φ(x)) and β = 1 E 1 (φ(x)) Note that a test using a rejection region C is equivalent to φ(x) = 1(x C) Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 13 / 30

The Neyman Pearson Lemma Theorem (Jerzy Neyman, Egon Pearson) When testing f 0 vs f 1, β is minimized, subject to α α 0 by: 1 f 1 (x)/f 0 (x) > λ φ(x) = γ f 1 (x)/f 0 (x) = λ 0 f 1 (x)/f 0 (x) < λ where λ is the largest constant such that P 0 (f 1 (X)/f 0 (X) λ) α 0 and P 0 (f 1 (X)/f 0 (X) λ) 1 α 0 and where γ is any number chosen so that E 0 (φ(x)) = P 0 (f 1 (X)/f 0 (X) > λ)+γp 0 (f 1 (X)/f 0 (X) = λ) = α 0 Value γ is unique if P 0 (f 1 (X)/f 0 (X) = λ) > 0. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 14 / 30

Binomial example again Example: Binomial(n,p) with p 0 = 1/2 and p 1 = 3/4: ratio f 1 /f 0 is 3 x 2 n If n = 5 this ratio is one of 1, 3, 9, 27, 81, 243 divided by 32. Suppose we have α = 0.05. λ must be one of the possible values of f 1 /f 0. If we try λ = 243/32 then and So λ = 81/32. P 0 (3 X 2 5 243/32) = P 0 (X = 5) = 1/32 < 0.05 P 0 (3 X 2 5 81/32) = P 0 (X 4) = 6/32 > 0.05 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 15 / 30

Binomial example continued Since P 0 (3 X 2 5 > 81/32) = P 0 (X = 5) = 1/32 we must solve for γ and find P 0 (X = 5)+γP 0 (X = 4) = 0.05 γ = 0.05 1/32 5/32 NOTE: No-one ever uses this procedure. = 0.12 Instead the value of α 0 used in discrete problems is chosen to be a possible value of the rejection probability when γ = 0 (or γ = 1). When the sample size is large you can come very close to any desired α 0 with a non-randomized test. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 16 / 30

Binomial again! If α 0 = 6/32 then we can either take λ to be 243/32 and γ = 1 or λ = 81/32 and γ = 0. However, our definition of λ in the theorem makes λ = 81/32 and γ = 0. When the theorem is used for continuous distributions it can be the case that the cdf of f 1 (X)/f 0 (X) has a flat spot where it is equal to 1 α 0. This is the point of the word largest in the theorem. Example: If X 1,...,X n are iid N(µ,1) and we have µ 0 = 0 and µ 1 > 0 then f 1 (X 1,...,X n ) f 0 (X 1,...,X n ) = exp{µ 1 Xi nµ 2 1/2 µ 0 Xi +nµ 2 0/2} which simplifies to exp{µ 1 Xi nµ 2 1/2} Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 17 / 30

Normal, one tailed test for mean Now choose λ so that P 0 (exp{µ 1 Xi nµ 2 1 /2} > λ) = α 0 Can make it equal because f 1 (X)/f 0 (X) has a continuous distribution. Rewrite probability as P 0 ( ( ) log(λ)+nµ X i > [log(λ)+nµ 2 2 1 /2]/µ 1) = 1 Φ 1 /2 n 1/2 µ 1 Let z α be upper α critical point of N(0,1); then z α0 = [log(λ)+nµ 2 1/2]/[n 1/2 µ 1 ]. Solve to get a formula for λ in terms of z α0, n and µ 1. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 18 / 30

Simplifying rejection regions Rejection region looks complicated: reject if a complicated statistic is larger than λ which has a complicated formula. But in calculating λ we re-expressed the rejection region in terms of Xi n > z α0 The key feature is that this rejection region is the same for any µ 1 > 0. WARNING: in the algebra above I used µ 1 > 0. This is why the Neyman Pearson lemma is a lemma! Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 19 / 30

Back to basics Definition: In the general problem of testing Θ 0 against Θ 1 the level of a test function φ is The power function is α = sup θ Θ 0 E θ (φ(x)) π(θ) = E θ (φ(x)) A test φ is a Uniformly Most Powerful level α 0 test if 1 φ has level α α o 2 If φ has level α α 0 then for every θ Θ 1 we have E θ (φ(x)) E θ (φ (X)) Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 20 / 30

Proof of Neyman Pearson lemma Given a test φ with level strictly less than α 0 define test φ (x) = 1 α 0 1 α φ(x)+ α 0 α 1 α which has level α 0 and β smaller than that of φ. Hence we may assume without loss that α = α 0 and minimize β subject to α = α 0. However, the argument which follows doesn t actually need this. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 21 / 30

Lagrange Multipliers Suppose you want to minimize f(x) subject to g(x) = 0. Consider first the function h λ (x) = f(x)+λg(x) If x λ minimizes h λ then for any other x f(x λ ) f(x)+λ[g(x) g(x λ )] Suppose you find λ such that solution x λ has g(x λ ) = 0. Then for any x we have f(x λ ) f(x)+λg(x) and for any x satisfying the constraint g(x) = 0 we have f(x λ ) f(x) So for this value of λ quantity x λ minimizes f(x) subject to g(x) = 0. To find x λ set usual partial derivatives to 0; then to find the special x λ you add in the condition g(x λ ) = 0. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 22 / 30

Return to proof of NP lemma For each λ > 0 we have seen that φ λ minimizes λα+β where φ λ = 1(f 1 (x)/f 0 (x) λ). As λ increases the level of φ λ decreases from 1 when λ = 0 to 0 when λ =. There is thus a value λ 0 where for λ > λ 0 the level is less than α 0 while for λ < λ 0 the level is at least α 0. Temporarily let δ = P 0 (f 1 (X)/f 0 (X) = λ 0 ). If δ = 0 define φ = φ λ. If δ > 0 define 1 φ(x) = γ 0 f 1 (x) f 0 (x) > λ 0 f 1 (x) where P 0 (f 1 (X)/f 0 (X) > λ 0 )+γδ = α 0. You can check that γ [0,1]. f 0 (x) = λ 0 f 1 (x) f 0 (x) < λ 0 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 23 / 30

End of NP proof Now φ has level α 0 and according to the theorem above minimizes λ 0 α+β. Suppose φ is some other test with level α α 0. Then We can rearrange this as λ 0 α φ +β φ λ 0 α φ +β φ β φ β φ +(α φ α φ )λ 0 Since α φ α 0 = α φ the second term is non-negative and β φ β φ which proves the Neyman Pearson Lemma. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 24 / 30

NP applied to Binomial(n, p) Binomial(n,p) model: test p = p 0 versus p 1 for a p 1 > p 0 NP test is of the form where we choose k so that and γ [0,1) so that φ(x) = 1(X > k)+γ1(x = k) P p0 (X > k) α 0 < P p0 (X k) α 0 = P p0 (X > k)+γp p0 (X = k) This rejection region depends only on p 0 and not on p 1 so that this test is UMP for p = p 0 against p > p 0. Since this test has level α 0 even for the larger null hypothesis it is also UMP for p p 0 against p > p 0. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 25 / 30

NP lemma applied to N(µ,1) model In the N(µ,1) model consider Θ 1 = {µ > 0} and Θ 0 = {0} or Θ 0 = {µ 0}. UMP level α 0 test of H 0 : µ Θ 0 against H 1 : µ Θ 1 is φ(x 1,...,X n ) = 1(n 1/2 X > z α0 ) Proof: For either choice of Θ 0 this test has level α 0 because for µ 0 we have P µ (n 1/2 X > z α0 ) = P µ (n 1/2 ( X µ) > z α0 n 1/2 µ) = P(N(0,1) > z α0 n 1/2 µ) P(N(0,1) > z α0 ) = α 0 Notice the use of µ 0. Central point: critical point is determined by behaviour on edge of null hypothesis. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 26 / 30

Normal example continued Now if φ is any other level α 0 test then we have Fix a µ > 0. According to the NP lemma where φ µ rejects if E 0 (φ(x 1,...,X n )) α 0 E µ (φ(x 1,...,X n )) E µ (φ µ (X 1,...,X n )) f µ (X 1,...,X n )/f 0 (X 1,...,X n ) > λ for a suitable λ. But we just checked that this test had a rejection region of the form n 1/2 X > z α0 which is the rejection region of φ. The NP lemma produces the same test for every µ > 0 chosen as an alternative. So we have shown that φ µ = φ for any µ > 0. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 27 / 30

Monotone likelihood ratio Fairly general phenomenon: for any µ > µ 0 the likelihood ratio f µ /f 0 is an increasing function of X i. So rejection region of NP test always region of form X i > k. Value of k determined by requirement that test have level α 0 ; this depends on µ 0 not on µ 1. Definition: The family f θ ;θ Θ R has monotone likelihood ratio with respect to a statistic T(X) if for each θ 1 > θ 0 the likelihood ratio f θ1 (X)/f θ0 (X) is a monotone increasing function of T(X). Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 28 / 30

Monotone likelihood ratio Theorem For a monotone likelihood ratio family the Uniformly Most Powerful level α test of θ θ 0 (or of θ = θ 0 ) against the alternative θ > θ 0 is 1 T(x) > t α φ(x) = γ T(X) = t α 0 T(x) < t α where P θ0 (T(X) > t α )+γp θ0 (T(X) = t α ) = α 0. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 29 / 30

Two tailed tests no UMP possible Typical family where this works: one parameter exponential family. Usually there is no UMP test. Example: test µ = µ 0 against two sided alternative µ µ 0. There is no UMP level α test. If there were its power at µ > µ 0 would have to be as high as that of the one sided level α test and so its rejection region would have to be the same as that test, rejecting for large positive values of X µ 0. But it also has to have power as good as the one sided test for the alternative µ < µ 0 and so would have to reject for large negative values of X µ 0. This would make its level too large. Favourite test: usual 2 sided test rejects for large values of X µ 0. Test maximizes power subject to two constraints: first, level α; second power is minimized at µ = µ 0. Second condition means power on alternative is larger than on the null. Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 30 / 30