Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Similar documents
6.4 Type I and Type II Errors

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Hypothesis Testing. Testing Hypotheses MIT Dr. Kempthorne. Spring MIT Testing Hypotheses

F79SM STATISTICAL METHODS

Topic 15: Simple Hypotheses

280 CHAPTER 9 TESTS OF HYPOTHESES FOR A SINGLE SAMPLE Tests of Statistical Hypotheses

Chapter 7: Hypothesis Testing

Lecture 1. ABC of Probability


Hypothesis Testing Chap 10p460

Topic 10: Hypothesis Testing

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing

For use only in [the name of your school] 2014 S4 Note. S4 Notes (Edexcel)

Chapters 10. Hypothesis Testing

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Chapters 10. Hypothesis Testing

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017

Topic 10: Hypothesis Testing

Frequentist Statistics and Hypothesis Testing Spring

Chapter 5: HYPOTHESIS TESTING

Summary of Chapters 7-9

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Lecture 3. Discrete Random Variables

Introductory Econometrics. Review of statistics (Part II: Inference)

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

LECTURE 5 HYPOTHESIS TESTING

hypothesis testing 1

STAT 830 Hypothesis Testing

Mathematical Statistics

ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing

STAT 830 Hypothesis Testing

Statistical Inference

STAT Chapter 8: Hypothesis Tests

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Lecture 13: p-values and union intersection tests

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

8: Hypothesis Testing

Institute of Actuaries of India

Section 9.1 (Part 2) (pp ) Type I and Type II Errors

Topic 17: Simple Hypotheses

Mathematical statistics

Tests about a population mean

PhysicsAndMathsTutor.com

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis testing (cont d)

ECE531 Lecture 4b: Composite Hypothesis Testing

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Composite Hypotheses. Topic Partitioning the Parameter Space The Power Function

Nuisance parameters and their treatment

STAT 801: Mathematical Statistics. Hypothesis Testing

Null Hypothesis Significance Testing p-values, significance level, power, t-tests

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary

Topic 17 Simple Hypotheses

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

14.30 Introduction to Statistical Methods in Economics Spring 2009

Lecture 31. Basic Theory of First Order Linear Systems

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Hypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004

Class 19. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Announcements. Proposals graded

Chapter Three. Hypothesis Testing

Lecture 21. Hypothesis Testing II

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

Lecture 12 November 3

Detection Theory. Chapter 3. Statistical Decision Theory I. Isael Diaz Oct 26th 2010

ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants

CHAPTER 8. Test Procedures is a rule, based on sample data, for deciding whether to reject H 0 and contains:

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?

Section 5.4: Hypothesis testing for μ

Lecture notes for probability. Math 124

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

CSE 312 Final Review: Section AA

Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4

simple if it completely specifies the density of x

Problems ( ) 1 exp. 2. n! e λ and

Hypothesis tests

Statistics for the LHC Lecture 1: Introduction

Design of Engineering Experiments

Master s Written Examination - Solution

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS

Basic Concepts of Inference

14.30 Introduction to Statistical Methods in Economics Spring 2009

INTERVAL ESTIMATION AND HYPOTHESES TESTING

Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma

Introduction to Statistical Inference

CONTINUOUS RANDOM VARIABLES

Statistical Methods for Particle Physics Lecture 2: statistical tests, multivariate methods

Hypothesis testing: theory and methods

MTMS Mathematical Statistics

Introductory Econometrics

Further Remarks about Hypothesis Tests

Detection theory. H 0 : x[n] = w[n]

Physics 403. Segev BenZvi. Classical Hypothesis Testing: The Likelihood Ratio Test. Department of Physics and Astronomy University of Rochester

Statistical Preliminaries. Stony Brook University CSE545, Fall 2016

MATH 240. Chapter 8 Outlines of Hypothesis Tests

Transcription:

Math 408 - Mathematical Statistics Lecture 29-30. Testing Hypotheses: The Neyman-Pearson Paradigm April 12-15, 2013 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 1 / 12

Agenda Example: Two Coins Tossing General Framework Type I Error and Type II Error Significance Level Power Neyman-Pearson Lemma Example: the likelihood ratio test for Gaussian variables The Concept of p-value Summary Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 2 / 12

Example: Two Coins Tossing Suppose Bob has two coins: Coin 0 has probability of heads p 0 = 0.5 Coin 1 has probability of heads p 1 = 0.7 Bob chooses one of the coins, tosses it n = 10 times and tells Alice the number of heads, but does not tell her whether it was coin 0 or coin 1. On the basis of the number of heads, Alice has to decide which coin it was. How should her decision rule be? Let X denote the number of heads. X X = {0, 1, 2,..., 10} Then for each coin we can compute the probability that Bob got exactly x heads: ( n P i (X = x) = p x) i x (1 p i ) n x, i = 0, 1. Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 3 / 12

Example: Two Coins Tossing 0.35 0.3 Coin 0 Coin 1 0.25 Probabiity 0.2 0.15 0.1 0.05 0 0 1 2 3 4 5 6 7 8 9 10 Number of Heads P0(X =2) Suppose that Bob observed 2 heads. Then P 1(X =2) 30, and, therefore, coin 0 was about 30 times more likely to produce this result than was coin 1. On the other hand, if there were 8 heads, then coin 1. P0(X =8) P 1(X =8) 0.19, which would favor Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 4 / 12

Hypothesis Testing The example with two coins is an example of hypothesis testing: The Null Hypothesis H 0 : Bob tossed coin 0 The Alternative Hypothesis H 1 : Bob tossed coin 1 Alice would accept H 0 if the likelihood ratio L(Data Coin 0) L(Data Coin 1) = P 0(X = x) P 1 (X = x) > 1 and she would reject H 0 if the likelihood ratio In this example, Alice would accept H 0 if L(Data Coin 0) L(Data Coin 1) = P 0(X = x) P 1 (X = x) < 1 x 6 and she would reject H 0 if x > 6 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 5 / 12

Hypothesis Testing: General Framework More formally, suppose that we partition the parameter space Θ into two disjoint sets Θ 0 and Θ 1 and that we wish to test H 0 : θ Θ 0 versus H 1 : θ Θ 1 We call H 0 the null hypothesis and H 1 the alternative hypothesis. Let X be data and let X be the range of X. We test a hypothesis by finding an appropriate subset of outcomes R X called the rejection region. If X R we reject the null hypothesis, otherwise, we do not reject the null hypothesis: X R reject H 0 In the Two Coins Example, X is the number of heads X is {0, 1, 2,..., 10} R is {7, 8, 9, 10} X / R accept H 0 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 6 / 12

Hypothesis Testing: General Framework Usually the rejection region R is of the form R = {x X : T (x) < c} where T is a test statistic and c is a critical value. The main problem in hypothesis testing is to find an appropriate test statistic T and an appropriate cutoff value c In the Two Coins Example, T (x) = c = 1 P0(X =x) P 1(X =x) is the likelihood ratio Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 7 / 12

Main Definitions In hypothesis testing, there are two types of errors we can make: Rejecting H 0 when H 0 is true is called a type I error Accepting H 0 when H 1 is true is called a type II error Definition The probability of a type I error is called the significance level of the test and is denoted by α α = P(type I error) = P(Reject H 0 H 0 ) The probability of a type II error is denoted by β β = P(type II error) = P(Accept H 0 H 1 ) (1 β) is called the power of the test power = 1 β = 1 P(Accept H 0 H 1 ) = P(Reject H 0 H 1 ) Thus, the power of the test is the probability of rejecting H 0 when it is false. Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 8 / 12

Neyman-Pearson Lemma Definition A hypothesis of the form θ = θ 0 is called a simple hypothesis. A hypothesis of the form θ > θ 0 or θ < θ 0 is called a composite hypothesis. The Neyman-Pearson Lemma shows that the test that is based on the likelihood ratio (as in the Two Coins Example) is optimal for simple hypotheses: Neyman-Pearson Lemma Suppose that H 0 and H 1 are simple hypotheses, H 0 : θ = θ 0 and H 1 : θ = θ 1. Suppose that the likelihood ratio test that rejects H 0 whenever the likelihood ratio is less than c, Reject H 0 L(Data θ 0) L(Data θ 1 ) < c has significance level α LR. Then any other test for which the significance level α α LR has power less than or equal to that of the likelihood ratio test 1 β 1 β LR Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 9 / 12

Example Example Let X 1,..., X n N(µ, σ 2 ), where σ 2 is known. Consider two simple hypotheses: H 0 : µ = µ 0 H 1 : µ = µ 1 > µ 0 Construct the likelihood ratio test with significance level α. Answer: Reject H 0 X n > µ 0 + z α σ n Neyman-Pearson: this test is the most powerful test among all tests with significance level α. Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 10 / 12

The Concept of p-value Reporting reject H 0 or accept H 0 is not very informative. For example, if the test just reposts to reject H 0, this does not tell us how strong the evidence against H 0 is. This evidence is summarized in terms of p-value. Definition Suppose for every α (0, 1) we have a test of significance level α with rejection region R α. Then, the p-value is the smallest significance level at which we can reject H 0 : p-value = inf{α : X R α } Informally, the p-value is a measure of the evidence against H 0 : the smaller the p-value, the stronger the evidence against H 0 Typically, researchers use the following evidence scale: p-value < 0.01: very string evidence against H 0 0.01 < p-value < 0.05: strong evidence against H 0 0.05 < p-value < 0.10: weak evidence against H 0 p-value > 0.10: little or no evidence against H 0 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 11 / 12

Summary In general, we partition the parameter space Θ into two disjoint sets Θ 0 and Θ 1 and test H 0 : θ Θ 0 versus H 1 : θ Θ 1 H0 is called the null hypothesis H1 is called the alternative hypothesis If Hi : θ = θ i, then the hypothesis is called simple If X is data and X is the range of X, then we reject H 0 X R X. Rejection region R = {x : T (x) < c} For the likelihood ratio test, T (x) = P(X =x H 0 ) P(X =x H 1 ) Type I Error: Rejecting H 0 when H 0 is true α = P(Reject H0 H 0) is called significance level (small α is good) Type II Error: Accepting H 0 when H 1 is true 1 β = 1 P(Accept H0 H 1) is called power (large power is good) Neyman-Pearson Lemma: basing the test on the likelihood ratio is optimal. p-value summarizes the evidence against the null hypothesis, p-value = inf{α : X R α }. Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 12 / 12