Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Similar documents
Composite Hypotheses. Topic Partitioning the Parameter Space The Power Function

Topic 15: Simple Hypotheses

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Topic 10: Hypothesis Testing

Topic 10: Hypothesis Testing

Statistical Inference

Topic 17 Simple Hypotheses

F79SM STATISTICAL METHODS

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Topic 19 Extensions on the Likelihood Ratio

Topic 17: Simple Hypotheses

LECTURE 5 HYPOTHESIS TESTING

Chapter 9: Hypothesis Testing Sections


280 CHAPTER 9 TESTS OF HYPOTHESES FOR A SINGLE SAMPLE Tests of Statistical Hypotheses

Summary of Chapters 7-9

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

On the Inefficiency of the Adaptive Design for Monitoring Clinical Trials

Hypothesis Testing. Testing Hypotheses MIT Dr. Kempthorne. Spring MIT Testing Hypotheses

Hypothesis Testing Chap 10p460

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

STAT 801: Mathematical Statistics. Hypothesis Testing

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

CHAPTER 8. Test Procedures is a rule, based on sample data, for deciding whether to reject H 0 and contains:

MATH 240. Chapter 8 Outlines of Hypothesis Tests

Composite Hypotheses

Tests about a population mean

Introductory Econometrics. Review of statistics (Part II: Inference)

ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing

STAT 515 fa 2016 Lec Statistical inference - hypothesis testing

Introductory Econometrics

TUTORIAL 8 SOLUTIONS #

9-7: THE POWER OF A TEST

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

Ch. 5 Hypothesis Testing

Institute of Actuaries of India

Quality Control Using Inferential Statistics In Weibull Based Reliability Analyses S. F. Duffy 1 and A. Parikh 2

Applied Statistics for the Behavioral Sciences

Visual interpretation with normal approximation

INTERVAL ESTIMATION AND HYPOTHESES TESTING

Chapter 7. Hypothesis Testing

8: Hypothesis Testing

The problem of base rates

Mathematical Statistics

Introduction to Statistical Hypothesis Testing

Hypothesis Testing. A rule for making the required choice can be described in two ways: called the rejection or critical region of the test.

Mathematical statistics

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

Hypothesis testing (cont d)

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.

Hypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004

Topic 21 Goodness of Fit

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

Introduction to Statistics

6.4 Type I and Type II Errors

Statistical Inference

Performance Evaluation and Comparison

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Lecture 12 November 3

Lecture 5: ANOVA and Correlation

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Preliminary Statistics. Lecture 5: Hypothesis Testing

FYST17 Lecture 8 Statistics and hypothesis testing. Thanks to T. Petersen, S. Maschiocci, G. Cowan, L. Lyons

8.1-4 Test of Hypotheses Based on a Single Sample

Detection and Estimation Chapter 1. Hypothesis Testing

Statistics: CI, Tolerance Intervals, Exceedance, and Hypothesis Testing. Confidence intervals on mean. CL = x ± t * CL1- = exp

Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques

Preliminary Statistics Lecture 5: Hypothesis Testing (Outline)

Chapters 10. Hypothesis Testing

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

9-6. Testing the difference between proportions /20

Stat 529 (Winter 2011) Experimental Design for the Two-Sample Problem. Motivation: Designing a new silver coins experiment

Statistical Inference

Mathematical statistics

Spearman Rho Correlation

Hypothesis testing: theory and methods

Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

Hypothesis for Means and Proportions

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

Chapter 7 Comparison of two independent samples

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS

Logistic Regression Analysis

Statistical Inference

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

Two Sample Hypothesis Tests

Finite sample size optimality of GLR tests. George V. Moustakides University of Patras, Greece

hypothesis testing 1

Review. December 4 th, Review

Basic Concepts of Inference

Chapter 7: Hypothesis Testing

4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49

Solution: First note that the power function of the test is given as follows,

14.30 Introduction to Statistical Methods in Economics Spring 2009

Chapter 8 of Devore , H 1 :

Transcription:

Topic 18 Composite Hypotheses Partitioning the Parameter Space 1 / 10

Outline Partitioning the Parameter Space 2 / 10

Partitioning the Parameter Space Simple hypotheses limit us to a decision between one of two possible states of nature. This limitation does not allow us, under the procedures of hypothesis testing to address the basic question: Does the parameter value θ 0 increase, decrease or change at all under under a different experimental condition? This leads us to consider composite hypotheses. In this case, the parameter space Θ is divided into two disjoint regions, Θ 0 and Θ 1. The hypothesis test is now written H 0 : θ Θ 0 versus H 1 : θ Θ 1. Again, H 0 is called the null hypothesis and H 1 the alternative hypothesis. 3 / 10

Partitioning the Parameter Space For the three alternatives to the question posed above, we have, for a one dimensional parameter space increase would lead to the choices Θ 0 = {θ; θ θ 0 } and Θ 1 = {θ; θ > θ 0 }, decrease would lead to the choices Θ 0 = {θ; θ θ 0 } and Θ 1 = {θ; θ < θ 0 }, and change would lead to the choices Θ 0 = {θ 0 } and Θ 1 = {θ; θ θ 0 } for some choice of parameter value θ 0.The effect that we are meant to show, here the nature of the change, is contained in Θ 1. The first two options given above are called one-sided tests. The third is called a two-sided test. Rejecting the null hypothesis, critical regions, and type I and type II errors have the same meaning for a composite hypotheses. Significance level and power will necessitate an extension of the ideas for simple hypotheses. 4 / 10

Power is now a function of the parameter value θ. If our test is to reject H 0 whenever the data fall in a critical region C, then the power function is defined as π(θ) = P θ {X C}. that gives the probability of rejecting the null hypothesis for a given parameter value. For θ Θ 0, π(θ) is the probability of making a type I error, i.e., rejecting the null hypothesis when it is indeed true. For θ Θ 1, 1 π(θ) is the probability of making a type II error, i.e., failing to reject the null hypothesis when it is false. The ideal power function has π(θ) 0 for all θ Θ 0 and π(θ) 1 for all θ Θ 1. 5 / 10

The goal is to make the chance for error small. One strategy is to consider a method analogous to that employed in the Neyman-Pearson lemma. Thus, we must simultaneously, fix a (significance) level α, now defined to be the largest value of π(θ) in the region Θ 0 defined by the null hypothesis, and. By focusing on the value of the parameter in Θ 0 that is most likely to result in an error, we insure that the probability of a type I error is no more that α irrespective of the value for θ Θ 0. Look for a critical region that makes the power function as large as possible for values of the parameter θ Θ 1. 6 / 10

Example. Let X 1, X 2,..., X n be independent N(µ, σ 0 ) random variables with σ 0 known and µ unknown. For the composite hypothesis for the one-sided test H 0 : µ µ 0 versus H 1 : µ > µ 0, we use the test statistic from the likelihood ratio test and reject H 0 if the statistic x is too large. Thus, the critical region C = {x; x k(µ 0 )}. If µ is the true mean, then the power function π(µ) = P µ {X C} = P µ { X k(µ 0 )}. The value of k(µ 0 ) depends on the level of the test. 7 / 10

As the actual mean µ increases, then the probability that the sample mean X exceeds a particular value k(µ 0 ) also increases. In other words, π is an increasing function. Thus, the maximum value of π on Θ 0 = {µ; µ µ 0 } takes place for µ = µ 0. Consequently, to obtain level α for the hypothesis test, set α = π(µ 0 ) = P µ0 { X k(µ 0 )}. We now use this to find the value k(µ 0 ). When µ 0 is the value of the mean, we standardize to give a standard normal random variable Z = X µ 0 σ 0 / n. Choose z α so that P{Z z α } = α. Thus, P µ0 {Z z α } = P µ0 { X µ 0 + σ 0 n z α } and k(µ 0 ) = µ 0 + (σ 0 / n)z α. 8 / 10

Example, If µ is the true state of nature, then Z = X µ σ 0 / n is a standard normal random variable. Use this to show that the power function ( π(µ) = 1 Φ z α µ µ ) 0 σ 0 / n where Φ is the distribution function for the standard normal. pi 0.0 0.2 0.4 0.6 0.8 1.0-0.5 0.0 0.5 1.0 Power function for the one-sided test with alternative greater. The size of the test α is given by the height of the red segment. Notice that π(µ) < α for all µ < µ 0 and π(µ) > α for all µ > µ 0. mu 9 / 10

We have seen the expression π(µ) = 1 Φ ( z α µ µ ) 0 σ 0 / n in several contexts. If we fix n, the number of observations and the alternative value µ = µ 1 > µ 0 and determine the power 1 β as a function of the significance level α, then we have the receiver operator characteristic. If we fix µ 1 the alternative value and the significance level α, then we can determine the power as a function of n the number of observations. If we fix n and the significance level α, then we can determine the power function π(µ), the power as a function of the alternative value µ. Exercise. Give the appropriate expression for π for a less than alternative and use this to plot the power function for the example with a model species and its mimic. Take α = 0.05, µ 0 = 10, σ 0 = 3, and n = 16 observations, 10 / 10