Chapter 9: Hypothesis Testing Sections

Similar documents
Theory of Statistical Tests

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

STAT 830 Hypothesis Testing

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

Summary of Chapters 7-9

TUTORIAL 8 SOLUTIONS #

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 7 Distributions derived from the normal distribution, and comparing independent samples.

Lecture 21. Hypothesis Testing II

STAT 830 Hypothesis Testing

Statistical Inference

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Testing One Sample Tests

F79SM STATISTICAL METHODS

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Lecture 12 November 3

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

A Very Brief Summary of Statistical Inference, and Examples

Chapter 9: Hypothesis Testing Sections

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Some General Types of Tests

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Detection Theory. Composite tests

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Hypothesis Testing - Frequentist

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

H 2 : otherwise. that is simply the proportion of the sample points below level x. For any fixed point x the law of large numbers gives that

A Very Brief Summary of Statistical Inference, and Examples

2014/2015 Smester II ST5224 Final Exam Solution

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Chapter 10. Hypothesis Testing (I)

Chapter 7. Hypothesis Testing

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Hypothesis Testing. A rule for making the required choice can be described in two ways: called the rejection or critical region of the test.

Problem 1 (20) Log-normal. f(x) Cauchy

Statistics. Statistics

Qualifying Exam in Probability and Statistics.

Statistical Theory MT 2007 Problems 4: Solution sketches

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017

STAT 801: Mathematical Statistics. Hypothesis Testing

Chapter 4. Theory of Tests. 4.1 Introduction

Statistical Inference

10. Composite Hypothesis Testing. ECE 830, Spring 2014

Comparing two independent samples

INTERVAL ESTIMATION AND HYPOTHESES TESTING

Statistical Inference: Maximum Likelihood and Bayesian Approaches

Statistical Theory MT 2006 Problems 4: Solution sketches

Probability and Statistics Notes

Topic 10: Hypothesis Testing

Interval Estimation. Chapter 9

Confidence Intervals, Testing and ANOVA Summary

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

Exercises Chapter 4 Statistical Hypothesis Testing

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Institute of Actuaries of India

Solution: First note that the power function of the test is given as follows,

Frequentist Statistics and Hypothesis Testing Spring

Mathematical Statistics

Detection theory 101 ELEC-E5410 Signal Processing for Communications

Central Limit Theorem ( 5.3)

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma

Topic 10: Hypothesis Testing

4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

Composite Hypotheses and Generalized Likelihood Ratio Tests

Chapters 10. Hypothesis Testing

14.30 Introduction to Statistical Methods in Economics Spring 2009

Lecture 13: p-values and union intersection tests

Statistical Inference

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4

Uniformly and Restricted Most Powerful Bayesian Tests

simple if it completely specifies the density of x

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics

Confidence Regions For The Ratio Of Two Percentiles

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Introductory Econometrics

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X).

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.

STATISTICS SYLLABUS UNIT I

EC2001 Econometrics 1 Dr. Jose Olmo Room D309

Importance Sampling and. Radon-Nikodym Derivatives. Steven R. Dunbar. Sampling with respect to 2 distributions. Rare Event Simulation

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS

ECE531 Lecture 4b: Composite Hypothesis Testing

(a) (3 points) Construct a 95% confidence interval for β 2 in Equation 1.

Topic 15: Simple Hypotheses

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

Two Sample Hypothesis Tests

Confidence Intervals for Normal Data Spring 2014

Lecture 26: Likelihood ratio tests

Null Hypothesis Significance Testing p-values, significance level, power, t-tests

First Year Examination Department of Statistics, University of Florida

Transcription:

Chapter 9: Hypothesis Testing Sections 9.1 Problems of Testing Hypotheses 9.2 Testing Simple Hypotheses 9.3 Uniformly Most Powerful Tests Skip: 9.4 Two-Sided Alternatives 9.6 Comparing the Means of Two Normal Distributions 9.7 The F Distributions 9.8 Bayes Test Procedures 9.9 Foundational Issues Hypothesis Testing 1 / 25

9.3 Uniformly Most Powerful Tests Uniformly Most Powerful (UMP) Tests H 0 : θ Ω 0 vs H 1 : θ Ω 1 A test δ is a uniformly most powerful test at level α 0 if for any other level α 0 test δ π(θ δ) π(θ δ ) for all θ Ω 1 I.o.w: It has the lowest probability of type II error of any test, uniformly for all θ Ω 1. First control the probability of type I error by setting the level (size) of the test low, then control the probability of type II error. If π(θ δ ) is high for all θ Ω 1, the test is often called powerful Hypothesis Testing 2 / 25

UMP Tests Chapter 9 9.3 Uniformly Most Powerful Tests Example 1: Simple hypotheses. H 0 : θ = θ 0 vs. H 1 : θ = θ 1 LRT is UMP by Nayman-Pearson lemma (Theorem 9.2.2) Example 2: One-sided hypotheses: H 0 : θ θ 0 vs. H 1 : θ > θ 0 In a large class of problems (the distribution has a monotone likelihood ratio ), we can show that reject H 0 if T t" is a UMP for some T (Ch 9.3) Example 3: Two-sided hypotheses: H 0 : θ = θ 0 vs. H 1 : θ θ 0 UMP tests do not exist (Page 565) Hypothesis Testing 3 / 25

9.3 Uniformly Most Powerful Tests Chapter 9: Hypothesis Testing Sections 9.1 Problems of Testing Hypotheses Skip: 9.2 Testing Simple Hypotheses 9.3 Uniformly Most Powerful Tests Skip: 9.4 Two-Sided Alternatives 9.6 Comparing the Means of Two Normal Distributions 9.7 The F Distributions 9.8 Bayes Test Procedures 9.9 Foundational Issues Hypothesis Testing 4 / 25

The t-test The t-test is a test for hypotheses concerning the mean parameter in the normal distribution when the variance is also unknown. The test is based on the t distribution The setup for the next few slides: Let X 1,..., X n be i.i.d. N(µ, σ 2 ) and consider the hypotheses H 0 : µ µ 0 vs. H 1 : µ > µ 0 (1) The parameter space here is < µ < and σ 2 > 0, i.e. Ω = (, ) (0, ) And Ω 0 = (, µ 0 ] (0, ) and Ω 1 = (µ 0, ) (0, ) Hypothesis Testing 5 / 25

The one-sided t-test The t test: a likelihood ratio test (see p. 583-585 in the book) Let U = ( n(x n µ 0 ) σ where σ 1 = n 1 If µ = µ 0 then U has the t n 1 distribution Tests based on U are called t tests ) 1/2 n (X i X n ) 2 i=1 Hypothesis Testing 6 / 25

The one-sided t-test Chapter 9 Let Tm 1 be the quantile function of the t m distribution The test δ that rejects H 0 in (1) if U T 1 n 1 (1 α 0) has size α 0 (Theorem 9.5.1) To calculate the p-value: Theorem 9.5.2: p-values for t Tests Let u be the observed value of U. The p-value for the hypothesis in (1) is 1 T n 1 (u). Hypothesis Testing 7 / 25

Example Example: Acid Concentration in Cheese (Example 8.5.4) Have a random sample of n = 10 lactic acid measurements from cheese, assumed to be from a normal distribution with unknown mean and variance. Observed: x n = 1.379 and σ = 0.3277 Perform the level α 0 = 0.05 t-test of the hypotheses Compute the p-value H 0 : µ 1.2 vs H 1 : µ > 1.2 Hypothesis Testing 8 / 25

The complete power function Need the power function to decide the sample size n The power function π(µ, σ 2 δ) is a non-central t m distributions Def: Non-central t m distributions Let W N(ψ, 1) and Y χ 2 m be independent. The distribution of X = W (Y /m) 1/2 is called the non-central t distribution with m degrees of freedom and non-centrality parameter ψ Hypothesis Testing 9 / 25

Non-central t m distribution Hypothesis Testing 10 / 25

The complete power function For the one-sided t-test Theorem 9.5.3 U has the non-central t n 1 distribution with non-centrality parameter ψ = n(µ µ 0 )/σ. The power function of the t-test that rejects H 0 in (1) if U T 1 n 1 (1 α 0) = c 1 is π(µ, σ 2 δ) = 1 T n 1 (c 1 ψ) Hypothesis Testing 11 / 25

Power function for the one-sided t-test Example: n = 10, µ 0 = 5, α 0 = 0.05 Note that the power function is a function of both σ 2 and µ Hypothesis Testing 12 / 25

The other one-sided t-test Now consider the hypothesis H 0 : µ µ 0 vs. H 1 : µ < µ 0 (2) The test δ that rejects H 0 if U T 1 n 1 (α 0) has size α 0 (Corollary 9.5.1) Theorem 9.5.2: p-values for t Tests Let u be the observed value of U. The p-value for the hypothesis in (2) is T n 1 (u). Theorem 9.5.3 U has the non-central t n 1 distribution with non-centrality parameter ψ = n(µ µ 0 )/σ. The power function of the t-test that rejects H 0 in (2) if U T 1 n 1 (α 0) = c 2 is π(µ, σ 2 δ) = T n 1 (c 2 ψ) Hypothesis Testing 13 / 25

Two-sided t-test Consider now the test with a two-sided alternative hypothesis: H 0 : µ = µ 0 vs. H 1 : µ µ 0 (3) Size α 0 test δ: rejects H 0 iff U T 1 n 1 (1 α 0/2) = c If u is the observed value of U then the p-value is 2(1 T n 1 ( u )) The power function is π(µ, σ 2 δ) = T n 1 ( c ψ) + 1 T n 1 (c ψ) Hypothesis Testing 14 / 25

Notes on one sample t tests Paired t tests are conducted in the same way For large n, the distribution of the test statistic under H 0 is close to the standard normal, i.e., the corresponding test is close to a Z test Hypothesis Testing 15 / 25

The two-sample t-test Chapter 9 9.6 Comparing the Means of Two Normal Distributions Comparing the means of two populations X 1,..., X m i.i.d. N(µ 1, σ 2 ) and Y 1,..., Y n i.i.d. N(µ 2, σ 2 ) The variance is the same for both samples, but unknown We are interested in testing one of these hypotheses: a) H 0 : µ 1 µ 2 vs. H 1 : µ 1 > µ 2 b) H 0 : µ 1 µ 2 vs. H 1 : µ 1 < µ 2 c) H 0 : µ 1 = µ 2 vs. H 1 : µ 1 µ 2 Hypothesis Testing 16 / 25

Two-sample t statistic Chapter 9 9.6 Comparing the Means of Two Normal Distributions Let X m = 1 m X i and Y n = 1 m n i=1 n i=1 m SX 2 = n (X i X m ) 2 and SY 2 = (Y i Y n ) 2 i=1 Y i i=1 m + n 2 (X m Y n ) U = ( 1 m + 1 ) 1/2 ( ) n S 2 X + SY 2 1/2 Theorem 9.6.1: If µ 1 = µ 2 then U t m+n 2 Theorem 9.6.4: For any µ 1 and µ 2, U has the non-central t m+n 2 distribution with non-centrality parameter µ 1 µ 2 ψ = σ (1/m + 1/n) 1/2 Hypothesis Testing 17 / 25

Two-sample t test summary Proofs similar to the regular t-test 9.6 Comparing the Means of Two Normal Distributions a) H 0 : µ 1 µ 2 vs. H 1 : µ 1 > µ 2 Level α 0 test: Reject H 0 iff U T 1 m+n 2 (1 α 0) p-value: 1 T m+n 2 (u) b) H 0 : µ 1 µ 2 vs. H 1 : µ 1 < µ 2 Level α 0 test: Reject H 0 iff U T 1 m+n 2 (α 0) p-value: T m+n 2 (u) c) H 0 : µ 1 = µ 2 vs. H 1 : µ 1 µ 2 Level α 0 test: Reject H 0 iff U T 1 m+n 2 (1 α 0/2) p-value: 2(1 T m+n 2 ( u )) Hypothesis Testing 18 / 25

9.6 Comparing the Means of Two Normal Distributions Power function is now a function of 3 parameters: π(µ 1, µ 2, σ 2 δ) The two-sample t-test is a likelihood ratio test (see p. 592) Important difference: Paired t test vs. two sample t test Two-sample t test with unequal variances Proposed test-statistics do not have known distribution, but approximations have been obtained Approach 1: The Welch statistic V = X m Y n ( S 2 X m(m 1) + S2 Y n(n 1) ) 1/2 can be approximated by a t distribution Approach 2: The distribution of the likelihood ratio statistic can be approximated by the χ 2 1 distribution if the sample size is large enough Hypothesis Testing 19 / 25

F-distributions Chapter 9 9.7 The F Distributions In light of the previous slide, it would be nice to have a test of whether the variances in the two normal populations are equal need the F m,n distributions Def: F m,n -distributions Let Y χ 2 m and W χ 2 n be independent. The distribution of X = Y /m W /n = ny mw is called the F distribution with m and n degrees of freedom The pdf of the F m,n distribution is f (x) = Γ ((m + n)/2) mm/2 n n/2 Γ(m/2)Γ(n/2) x m/2 1 (mx + n) (m+n)/2 x > 0 Hypothesis Testing 20 / 25

F-distributions Chapter 9 9.7 The F Distributions Hypothesis Testing 21 / 25

9.7 The F Distributions Properties of the F-distributions The 0.95 and 0.975 quantiles of the F m,n distribution is tabulated in the back of the book for a few combinations of m and n Theorem 9.7.2: Two properties (i) If X F m,n then 1/X F n,m (ii) If Y t n then Y 2 F 1,n Hypothesis Testing 22 / 25

9.7 The F Distributions Comparing the variances of two normals Comparing the variances of two populations X 1,..., X m i.i.d. N(µ 1, σ1 2) and Y 1,..., Y n i.i.d. N(µ 2, σ2 2) All four parameters unknown Consider the hypotheses: (I) H 0 : σ 2 1 σ2 2 vs. H 1 : σ 2 1 > σ2 2 and the test that rejects H 0 if V c, where This test is called an F-test σ 2 2 σ 2 1 V F m 1,n 1 If σ 2 1 = σ2 2 then V F m 1,n 1 V = S2 X /(m 1) SY 2 /(n 1) Hypothesis Testing 23 / 25

The F test Chapter 9 9.7 The F Distributions Let G m,n (x) be the cdf of the F m,n distribution Theorem 9.7.4 Let δ be the test that rejects H 0 in H 0 : σ 2 1 σ2 2 vs. H 1 : σ 2 1 > σ2 2 if V c = G 1 m 1,n 1 (1 α 0). Then δ is of size α 0 and ( ) π(µ 1, µ 2, σ1 2, σ2 2 δ) = 1 G σ 2 2 m 1,n 1 c and σ1 2 p-value = 1 G m 1,n 1 (v), where v is the observed value of V Hypothesis Testing 24 / 25

9.7 The F Distributions The F test two sided alternative Two sided alternative: H 0 : σ 2 1 = σ2 2 vs. H 1 : σ 2 1 σ2 2 Equal-tailed two-sided F test Let δ be the F-test that rejects H 0 when V c 1 = G 1 m 1,n 1 (α 0/2) or V c 2 = G 1 m 1,n 1 (1 α 0/2). Then δ is a level α 0 test and the p-value is 2 min{1 G m 1,n 1 (v), G m 1,n 1 (v)}. Hypothesis Testing 25 / 25