Section 9.1 (Part 2) (pp ) Type I and Type II Errors

Similar documents
Section 10.1 (Part 2 of 2) Significance Tests: Power of a Test

Section 5.4: Hypothesis testing for μ

280 CHAPTER 9 TESTS OF HYPOTHESES FOR A SINGLE SAMPLE Tests of Statistical Hypotheses

Announcements. Unit 3: Foundations for inference Lecture 3: Decision errors, significance levels, sample size, and power.

Unit 19 Formulating Hypotheses and Making Decisions

PSY 305. Module 3. Page Title. Introduction to Hypothesis Testing Z-tests. Five steps in hypothesis testing

Preliminary Statistics Lecture 5: Hypothesis Testing (Outline)

ECO220Y Review and Introduction to Hypothesis Testing Readings: Chapter 12

Statistical Process Control (contd... )

Hypothesis Testing The basic ingredients of a hypothesis test are

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals

Chapter 7: Hypothesis Testing

Chapter Three. Hypothesis Testing

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

23. MORE HYPOTHESIS TESTING

Hypothesis Testing. ECE 3530 Spring Antonio Paiva

ECO220Y Hypothesis Testing: Type I and Type II Errors and Power Readings: Chapter 12,

Topic 17 Simple Hypotheses

Introduction to Statistics

Preliminary Statistics. Lecture 5: Hypothesis Testing

Hypothesis testing. Data to decisions

Tests about a population mean

9-7: THE POWER OF A TEST

Statistical Inference. Why Use Statistical Inference. Point Estimates. Point Estimates. Greg C Elvers

STAT 515 fa 2016 Lec Statistical inference - hypothesis testing

LECTURE 5. Introduction to Econometrics. Hypothesis testing

Hypothesis tests

HYPOTHESIS TESTING. Hypothesis Testing

Statistical Inference: Uses, Abuses, and Misconceptions

First we look at some terms to be used in this section.

MTMS Mathematical Statistics

Hypotheses and Errors

LAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2

In any hypothesis testing problem, there are two contradictory hypotheses under consideration.

Fundamental Statistical Concepts and Methods Needed in a Test-and-Evaluator s Toolkit. Air Academy Associates

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing

Hypotheses Test Procedures. Is the claim wrong?

STAT Chapter 8: Hypothesis Tests

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur

Hypothesis Testing. Testing Hypotheses MIT Dr. Kempthorne. Spring MIT Testing Hypotheses

Quantitative Methods for Economics, Finance and Management (A86050 F86050)

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

HYPOTHESIS TESTING: THE CHI-SQUARE STATISTIC

Statistical Inference. Section 9.1 Significance Tests: The Basics. Significance Test. The Reasoning of Significance Tests.

Mathematical statistics

Mathematical statistics

ME3620. Theory of Engineering Experimentation. Spring Chapter IV. Decision Making for a Single Sample. Chapter IV

Hypothesis Tests and Estimation for Population Variances. Copyright 2014 Pearson Education, Inc.

Tests for Two Coefficient Alphas

Introduction 1. STA442/2101 Fall See last slide for copyright information. 1 / 33

Inferential statistics

Applied Statistics for the Behavioral Sciences

hypothesis a claim about the value of some parameter (like p)

Dr. Allen Back. Nov. 21, 2016

Econ 325: Introduction to Empirical Economics

The Purpose of Hypothesis Testing

14.30 Introduction to Statistical Methods in Economics Spring 2009

Hypothesis Testing. ) the hypothesis that suggests no change from previous experience

Verification and Validation. CS1538: Introduction to Simulations

CHAPTER 9, 10. Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities:

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Mathematical Statistics

Chapter 5: HYPOTHESIS TESTING

The problem of base rates

Relating Graph to Matlab

WISE Power Tutorial Answer Sheet

5.2 Tests of Significance

Introductory Econometrics. Review of statistics (Part II: Inference)

8: Hypothesis Testing

CHAPTER 9: HYPOTHESIS TESTING

(right tailed) or minus Z α. (left-tailed). For a two-tailed test the critical Z value is going to be.

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wilcoxon Test and Calculating Sample Sizes


Tests for the Odds Ratio in Logistic Regression with One Binary X (Wald Test)

Math 101: Elementary Statistics Tests of Hypothesis

One-Way Repeated Measures Contrasts

Statistics for IT Managers

Review of Statistics 101

Multiple Testing. Gary W. Oehlert. January 28, School of Statistics University of Minnesota

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

For use only in [the name of your school] 2014 S4 Note. S4 Notes (Edexcel)

Analysis of Variance (ANOVA)

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

hypothesis testing 1

Topic 17: Simple Hypotheses

LECTURE 12 CONFIDENCE INTERVAL AND HYPOTHESIS TESTING

20 Hypothesis Testing, Part I

Visual interpretation with normal approximation

8.1-4 Test of Hypotheses Based on a Single Sample

Chapter 23. Inferences About Means. Monday, May 6, 13. Copyright 2009 Pearson Education, Inc.

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

Do students sleep the recommended 8 hours a night on average?

Sampling Distributions

Statistics for Managers Using Microsoft Excel/SPSS Chapter 8 Fundamentals of Hypothesis Testing: One-Sample Tests

Statistical Inference

PASS Sample Size Software. Poisson Regression

16.3 One-Way ANOVA: The Procedure

Chapter 10 Verification and Validation of Simulation Models. Banks, Carson, Nelson & Nicol Discrete-Event System Simulation

Transcription:

Section 9.1 (Part 2) (pp. 547-551) Type I and Type II Errors Because we are basing our conclusion in a significance test on sample data, there is always a chance that our conclusions will be in error. It turns out that there are two types of errors that can be committed in a significance test: If we reject H 0 when H 0 is true, it is called a Type I Error If we fail to reject H 0 when H 0 is false, it is called a Type II Error Memory aid: Type II error occurs when we fail II reject the null hypothesis when we really should. Example - The manager of a fast-food restaurant wants to reduce the proportion of drive-through customers who have to wait for more than 2 minutes to receive their food once their order is placed. Based on store records, the proportion of customers who had to wait at least 2 minutes was p = 0.63. To reduce this proportion, the manager will collect a random sample of drive-through times and test the following hypotheses: H 0: p = 0.63 H A: p < 0.63 Where p is the true proportion of drive-through customers who have to wait more than 2 minutes after their order is placed to receive their food. Describe a Type I and a Type II error in this setting and explain the consequences of each. Application - In the United States court system, defendants are presumed innocent and to be proven guilty, it must be done beyond a shadow of a doubt. In this situation, what constitutes a Type I error? What constitutes a Type II error? Which error is worse?

Error Probabilities - We can assess the performance of a significance test by looking at the probabilities of the two types of errors. Example (cont) - In the fast-food example we were testing the following hypotheses H 0: p = 0.63 H A: p < 0.63 where p = the true proportion of drive-through customers who have to wait more than two minutes after their order is placed to receive their food. Suppose that the manager decided to carry out this test using a random sample of 250 orders with a significance level of = 0.10. What is the probability of making a Type I error? A Type I error is to reject H 0 when. If our sample results in a p that is much smaller than 0.63, we will. How small would p have to be? The 10% significance level tells us to count results that could happen less than 10% of the time by chance if. Let s assume H 0: p = 0.63 is true. Then the sampling distribution of p will have: Shape: Center: Spread: From this example, what can we conclude about the probability of a Type I error? Significance and Type I Error The significance level of any fixed level test is the probability of a Type I error. That is, is the probability that the test will reject the null hypothesis H 0 when H 0 is in fact true. Consider the consequences of a Type I error before choosing a significance level. P(Type I error) = P(Reject H 0 H 0 is true) =

What about Type II errors? A significance test makes a Type II error when. A high probability of a Type II error for a particular alternative hypothesis means that the test is not sensitive enough to usually detect the alternative. Rather than being concerned with the probability of failing to reject H 0 when H 0 is false we really want the probability of getting it right. In other words, we want the probability of rejecting H 0 when H 0 is in fact false. This probability (of getting it right ) is referred to as the power of a significance test. The power of a test against a specific alternative is the probability that the test will reject H 0 at a chosen significance level when the specified alternative value of the parameter is true. The power of a test against any alternative is 1 minus the probability of a Type II error for that alternative, that is Power = 1 - = P(Rejecting H 0 H A is true) There are many ways to explain power: Power is the probability of rejecting the null hypothesis when in fact it is false. Power is the probability of making a correct decision (to reject the null hypothesis) when the null hypothesis is false. Power is the probability that a test of significance will pick up on an effect that is present. Power is the probability that a test of significance will detect a deviation from the null hypothesis, should such a deviation exist. Power is the probability of avoiding a Type II error.

Planning Studies: The Power of a Statistical Test - How large a sample should we take to carry out a significance test? The answer depends on what alternative values of the parameter are important to detect. There are basically four things that affect power: (1) The significance level of the test. (2) The sample size n. (3) The inherent variability in the measured response variable. (4) The difference between the hypothesized value of a parameter and its true value.

Example (cont) - Suppose the manager of the fast-food restaurant wants to change some aspects of his study about the proportion of drive-through customers who have to wait at least two minutes to receive their food after they place their order. Significance level: To reduce the possibility of Type I error and avoid the possibility of unnecessarily paying an extra employee, the manager reduces the significance level from 0.10 to 0.01. Practical importance: To justify the additional cost of the extra employee, the manager decides that the true proportion must be reduced to at most 0.53. To get faster results, the manager reduces the sample size from 250 to 100. How will these changes affect the power of the test of H 0: p = 0.63 versus H A: p < 0.63? We will use the Improved Batting Averages (Power) Applet at www.rossmanchance.com to see. 1. Hypothesized value = 0.63 Alternative value = 0.53 n = 250, 10,000 samples, = 0.10 3. Hypothesized value = 0.63 Alternative value = 0.50 n = 250, 10,000 samples, = 0.01 2. Hypothesized value = 0.63 Alternative value = 0.53 n = 250, 10,000 samples, change = 0.01 4. Hypothesized value = 0.63 Alternative value = 0.53 n = 100, 10,000 samples, = 0.01 Ways to Reduce Error (1) If you insist on a smaller significance level, you have to take a larger sample. A smaller significance level requires stronger evidence to reject the null hypothesis. (2) If you insist on higher power, you will need a larger sample. Higher power gives a better chance of detecting a difference when it is really there. (3) At any significance level and desired power, detecting small a small difference requires a larger sample than detecting a large difference. The best advice for maximizing the power of a test is to choose as high an level (Type I error probability) as you are willing to risk and as large a sample size as you can afford. HW: Read pp. Read pp 547-551; problems 17, 21, 23, 25-30.

Improved Batting Averages Applet Enter 0.63 for hypothesized value Enter 0.53 for alternative value Enter 250 for sample size Enter 10,000 for number of samples Press Draw Samples Enter 0.1 for alpha, press count to verify that the power when p=0.53 is about 97% Change alpha to 0.01 and press count Change the alternative value to 0.50 Press draw samples Choose level of significance and press count

Change the sample size to 100 Press draw samples Choose level of significance and press count