Because it tests for differences between multiple pairs of means in one test, it is called an omnibus test.

Similar documents
Chapter 13, Part A Analysis of Variance and Experimental Design

Overview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions

MOST PEOPLE WOULD RATHER LIVE WITH A PROBLEM THEY CAN'T SOLVE, THAN ACCEPT A SOLUTION THEY CAN'T UNDERSTAND.

Comparing Two Populations. Topic 15 - Two Sample Inference I. Comparing Two Means. Comparing Two Pop Means. Background Reading

2 1. The r.s., of size n2, from population 2 will be. 2 and 2. 2) The two populations are independent. This implies that all of the n1 n2

Common Large/Small Sample Tests 1/55

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Chapter 6 Sampling Distributions

Topic 9: Sampling Distributions of Estimators

Recall the study where we estimated the difference between mean systolic blood pressure levels of users of oral contraceptives and non-users, x - y.

This chapter focuses on two experimental designs that are crucial to comparative studies: (1) independent samples and (2) matched pair samples.

Sample Size Determination (Two or More Samples)

Topic 9: Sampling Distributions of Estimators

Agreement of CI and HT. Lecture 13 - Tests of Proportions. Example - Waiting Times

STA Learning Objectives. Population Proportions. Module 10 Comparing Two Proportions. Upon completing this module, you should be able to:

Math 140 Introductory Statistics

This is an introductory course in Analysis of Variance and Design of Experiments.

Chapter 22. Comparing Two Proportions. Copyright 2010, 2007, 2004 Pearson Education, Inc.

Final Examination Solutions 17/6/2010

Topic 9: Sampling Distributions of Estimators

Properties and Hypothesis Testing

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10

Chapter 22. Comparing Two Proportions. Copyright 2010 Pearson Education, Inc.

Lecture 5: Parametric Hypothesis Testing: Comparing Means. GENOME 560, Spring 2016 Doug Fowler, GS

Sampling Distributions, Z-Tests, Power

Module 1 Fundamentals in statistics

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

UNIVERSITY OF TORONTO Faculty of Arts and Science APRIL/MAY 2009 EXAMINATIONS ECO220Y1Y PART 1 OF 2 SOLUTIONS

Class 23. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Math 152. Rumbos Fall Solutions to Review Problems for Exam #2. Number of Heads Frequency

A quick activity - Central Limit Theorem and Proportions. Lecture 21: Testing Proportions. Results from the GSS. Statistics and the General Population

Comparing your lab results with the others by one-way ANOVA

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

Introduction to Econometrics (3 rd Updated Edition) Solutions to Odd- Numbered End- of- Chapter Exercises: Chapter 3

Hypothesis Testing. Evaluation of Performance of Learned h. Issues. Trade-off Between Bias and Variance

Stat 200 -Testing Summary Page 1

PSYCHOLOGICAL RESEARCH (PYC 304-C) Lecture 9

ST 305: Exam 3 ( ) = P(A)P(B A) ( ) = P(A) + P(B) ( ) = 1 P( A) ( ) = P(A) P(B) σ X 2 = σ a+bx. σ ˆp. σ X +Y. σ X Y. σ X. σ Y. σ n.

t distribution [34] : used to test a mean against an hypothesized value (H 0 : µ = µ 0 ) or the difference

Chapter 13: Tests of Hypothesis Section 13.1 Introduction

HYPOTHESIS TESTS FOR ONE POPULATION MEAN WORKSHEET MTH 1210, FALL 2018

Power and Type II Error

Section 9.2. Tests About a Population Proportion 12/17/2014. Carrying Out a Significance Test H A N T. Parameters & Hypothesis

Chapter 2 The Monte Carlo Method

Tests of Hypotheses Based on a Single Sample (Devore Chapter Eight)

If, for instance, we were required to test whether the population mean μ could be equal to a certain value μ

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

1 Inferential Methods for Correlation and Regression Analysis

MA238 Assignment 4 Solutions (part a)

3/3/2014. CDS M Phil Econometrics. Types of Relationships. Types of Relationships. Types of Relationships. Vijayamohanan Pillai N.

BIOS 4110: Introduction to Biostatistics. Breheny. Lab #9

7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals

University of California, Los Angeles Department of Statistics. Hypothesis testing

Data Analysis and Statistical Methods Statistics 651

Lecture 8: Non-parametric Comparison of Location. GENOME 560, Spring 2016 Doug Fowler, GS

Stat 139 Homework 7 Solutions, Fall 2015

Statistics. Chapter 10 Two-Sample Tests. Copyright 2013 Pearson Education, Inc. publishing as Prentice Hall. Chap 10-1

Interval Estimation (Confidence Interval = C.I.): An interval estimate of some population parameter is an interval of the form (, ),

Chapter 11: Asking and Answering Questions About the Difference of Two Proportions

Formulas and Tables for Gerstman

Sample Size Estimation in the Proportional Hazards Model for K-sample or Regression Settings Scott S. Emerson, M.D., Ph.D.

Statistics 511 Additional Materials

Biostatistics for Med Students. Lecture 2

5. A formulae page and two tables are provided at the end of Part A of the examination PART A

Lecture 7: Non-parametric Comparison of Location. GENOME 560, Spring 2016 Doug Fowler, GS

NCSS Statistical Software. Tolerance Intervals

LESSON 20: HYPOTHESIS TESTING

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

Lecture 6 Simple alternatives and the Neyman-Pearson lemma

Chapter 22: What is a Test of Significance?

Lecture 7: Properties of Random Samples

ENGI 4421 Confidence Intervals (Two Samples) Page 12-01

Table 12.1: Contingency table. Feature b. 1 N 11 N 12 N 1b 2 N 21 N 22 N 2b. ... a N a1 N a2 N ab

Big Picture. 5. Data, Estimates, and Models: quantifying the accuracy of estimates.

Topic 18: Composite Hypotheses

GG313 GEOLOGICAL DATA ANALYSIS

6 Sample Size Calculations

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

- E < p. ˆ p q ˆ E = q ˆ = 1 - p ˆ = sample proportion of x failures in a sample size of n. where. x n sample proportion. population proportion

Lecture 2: Monte Carlo Simulation

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)

Sampling, Sampling Distribution and Normality

Day 8-3. Prakash Balachandran Department of Mathematics & Statistics Boston University. Friday, October 28, 2011

Frequentist Inference

(7 One- and Two-Sample Estimation Problem )

CHAPTER 8 FUNDAMENTAL SAMPLING DISTRIBUTIONS AND DATA DESCRIPTIONS. 8.1 Random Sampling. 8.2 Some Important Statistics

Chapter 23: Inferences About Means

10. Comparative Tests among Spatial Regression Models. Here we revisit the example in Section 8.1 of estimating the mean of a normal random

A statistical method to determine sample size to estimate characteristic value of soil parameters

Expectation and Variance of a random variable

MIT : Quantitative Reasoning and Statistical Methods for Planning I

General IxJ Contingency Tables

Describing the Relation between Two Variables

AP Statistics Review Ch. 8

(6) Fundamental Sampling Distribution and Data Discription

October 25, 2018 BIM 105 Probability and Statistics for Biomedical Engineers 1

Last Lecture. Wald Test

BIOSTATISTICS. Lecture 5 Interval Estimations for Mean and Proportion. dr. Petr Nazarov

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

Economics Spring 2015

Transcription:

Math 308 Sprig 018 Classes 19 ad 0: Aalysis of Variace (ANOVA) Page 1 of 6 Itroductio ANOVA is a statistical procedure for determiig whether three or more sample meas were draw from populatios with equal meas. I everyday laguage, ANOVA tests the ull hypothesis that the populatio meas (estimated by the sample meas) are all equal. If this ull hypothesis is reected, the we coclude that the populatio meas are ot all equal. A more precise formulatio of the ull ad alterative hypotheses for comparig meas is: H0 : 1... H : 1 at least oe pair of meas is differet, i, i, 1,..., ; i Because it tests for differeces betwee multiple pairs of meas i oe test, it is called a omibus test. Motivatio for ANOVA: You may as, Why do t we ust test if the meas are equal with t-tests? After all, we ow about t-tests. The problem with t-tests o eve three sample meas is that the t-tests are ot idepedet because you use each mea more tha oce. Cosider a simple example. Example 1: A chemical egieer must decide betwee 3 catalysts i terms of reactio completio times. The chemical egieer taes three samples, each with 5 completio times. If she tries to mae ifereces about the populatio meas usig t-tests o the three sample meas, she will use each sample mea twice: x1vs x, x1vs x3ad x vs x 3. Because she used each sample mea i two t-tests the t-tests are ot idepedet ad, therefore, she caot ow the probability of a Type I error. Recall that a Type I error is reectig oe of the three ull hypotheses whe it is true. Clearly, the situatio becomes worse whe we compare more tha three meas. Ca you thi of a way to remedy this situatio? Problem: Iflatio of Type I error with multiple t-tests. Oe remedy is to use ANOVA. The ull hypothesis of ANOVA assumes that all meas are equal. This is equivalet to statig that all samples were tae from the same populatio. The ANOVA test elimiates the problem of multiple t-tests o the same sample meas by testig all the meas at oce to see if ay of the meas are differet. If we reect the ull hypothesis, all we ow is that at least oe pair of meas is differet. We caot tell which pair of meas is differet util we ru a multiple compariso test. Solutio: Completely Radomized Desigs (also called Oe-Way ANOVA or a sigle-factor ANOVA).Example : A miig egieer wishes to compare five strais of bacteria with respect to mea copper yield (lb/to).

Math 308 Sprig 018 Classes 19 ad 0: Aalysis of Variace (ANOVA) Page of 6 Example 3: A chemical egieer assesses the effect of four pressure settigs o product yield (grams/liter). Pressure (i grams/liter) Sample Batch Low (1) Moderate () Strog (3) High (4) 1 8 30 31 9 6 9 9 7 3 9 30 33 30 4 30 30 33 31 5 8 8 9 7 6 31 3 33 3 7 6 9 8 7 8 3 3 3 3 9 5 8 7 7 10 9 30 3 30 The hypothesis-testig procedure for ANOVA 1. We formulate a ull hypothesis ad a appropriate alterative hypothesis. Cosider the sigle-factor ANOVA model for a sigle factor with 4 levels, such as the four differet pressure settigs i Example B. Each populatio mea may be represeted as: 4 1. B, where. is the populatio mea of the product yield at the four 4 product settigs ad B is the effect of pressure settig. Note that B may be positive if a pressure settig icreases product yield over the mea settig or egative if a product settig decreases product yield. Uder the ull hypothesis, B 0 for each. The sigle-factor ANOVA ull hypothesis for our example, H0 : 1 3 4, icludes the assumptio that the variaces are equal; so, the ull hypothesis is equivalet to assumig that all samples were draw from a oe populatio distributed as N(, ). Uder this assumptio we ow that the stadard error of the sample meas is. Cosequetly, we ow how the sample meas should be X distributed if the samples are draw from a sigle populatio or idetical populatios. Assumptios of sigle-factor ANOVA: 1. The depedet variable is ormally distributed.. Each populatio has the same or equal variace (homogeeity of variace). The factor has equal variaces at all levels of the factor. 3. A o-mathematical assumptio is that the samples represet idepedet radom samples.

Math 308 Sprig 018 Classes 19 ad 0: Aalysis of Variace (ANOVA) Page 3 of 6. We specify the probability of Type I error, P(Type I error). P(Type I error) is usually set at 0.05 or 0.01. 3. Based o the samplig distributio of a appropriate statistic, we costruct a criterio for testig the ull hypothesis agaist the give alterative. The F-statistic is used to compare the variability displayed by the sample meas (the amog-sample treatmet variace) to the variability displayed withi the samples (the withi-sample error variace). The form of the test or observed F-statistic is MSTreatmet F. MS error The distributio of the F-statistic is geerated from the ull hypothesis which assumes that all samples were tae from equivalet populatios. Note that the umerator of the F statistic estimates the variace of the mea. We ow that the variace of the mea is a fuctio of the variace of X:. The deomiator estimates the variace of X X iside the samples divided by the sample size:. So, the umerator ad the deomiator estimate the variace of the mea usig two differet sample statistics. So, if the ull hypothesis is true, the F value should radomly vary aroud the value 1. If the ull hypothesis is NOT TRUE, the umerator will ot be a good estimate of the variace of the mea ad will be larger tha 1. So, if the variability amog the sample meas is greater tha the variability withi the samples, the F is larger tha the critical value ad the ull hypothesis is reected. I that case we coclude that there is at least oe pair of populatio meas that are ot equal. Criterio or decisio rule: For the oe-factor ANOVA, the degrees of freedom for the umerator of the F statistic v1 1 ad the degrees of freedom for the deomiator v T c, where T is the sum of all sample sizes. The F-test is always a oe-sided test. We reect the ull hypothesis oly if the obtaied F value is too large. This meas that the differeces betwee the sample meas are too large for all the sample meas to represet the sigle mea of oe populatio. 4.We calculate from the data the value of the (obtaied) F statistic. ANOVA Table. A ANOVA table is used to calculate a F statistic. Here is a summary of relevat otatio. = umber of groups (samples from distict populatios) i = sample size of ith group N = sum of the sample sizes (total umber of observatios) x i = mea of ith sample x = grad mea or mea of all observatios s i = stadard deviatio of ith sample

Math 308 Sprig 018 Classes 19 ad 0: Aalysis of Variace (ANOVA) Page 4 of 6 v 1 = df treatmet = -1 = degrees of freedom amog sample meas v = df error = - =degrees of freedom withi samples SS treatmet = SS error = i ( xi - x ) (treatmet sum of squares amog sample meas) ( i -1)s i (error sum of squares withi samples) If you already have the sample variace, the followig formula is much easier to use. SS total = i 1 ( xi x) = SS treatmet + SS error MS treatmet = SS treatmet /(-1) = mea square amog sample meas MS error = SS error /(-) = mea square withi samples F = MS amog / MS withi The p-value is the probability of a F as large or larger tha the oe obtaied. 5. We decide to reect the ull hypothesis or fail to reect it. Decisio: We reect or FTR the ull hypothesis H0 : 1 3 4 at the specified α level by usig the decisio rule. Example 3: 1. Hypotheses: The ull hypothesis for comparig = 4 meas is: H :... 0 1 4 : 1 H at least oe pair of meas is differet, i. P(Type I error) ad Test Statistic: The egieer chose α = 0.05. Assumptios for the F statistic: The egieer had evidece that larger samples at various pressures displayed symmetric histograms; so, the assumptio of ormality seemed appropriate. Note: The sample stadard deviatios may vary by a factor of without producig ivalid results; so, give the stadard deviatios i a table for step 4, the assumptio of homogeeity of variaces seems reasoable. The egieer, therefore, chose the F statistic. 3. Decisio rule: The F for v 1 = 4-1 =3 ad v = 40 4 = 36 degrees of freedom for α = 0.05 is.86. The egieer s decisio rule is to reect the ull hypothesis if the obtaied F >.86. (Note that I foud the umber.86 i a table which icludes v = 36. Usig a table which preset v i icremets of 10, the closest you could come to the correct criterio F is by fidig the critical F value for v 1 = 3 ad v = 30. (Why ot 40?) This value is.9. Note that this will be coservative i the sese that a obtaied F of.87 will ot be reected for degrees of freedom 3 ad 30, but will be (correctly) reected for degrees of freedom 3 ad 36.)

Math 308 Sprig 018 Classes 19 ad 0: Aalysis of Variace (ANOVA) Page 5 of 6 4. Calculate F value. Note: I displayed oly the first three decimal places i this example. I actually stored the stadard deviatios i the memory of my calculator at greater precisio. Therefore, if you wor through this example, you may differ from my aswers begiig i the hudredths place. Recall that the formula for the sample stadard deviatio is: 1 1 s X X X X 1 1 ( i ) ( i ) Pressure Mea Yield N Std. Deviatio 1.00 8.400 10.706.00 9.800 10 1.3984 3.00 30.700 10.63 4.00 9.00 10.0976 Total 9.55 40.136 SS treatmet = i ( xi - x ) = 10[(8.4 9.55) (9.8 9.55) (30.7 9.55) (9. 9.55) ] SS error = = 8.75 ( i -1)s i = 9[.706 1.3984.63.0976 ] 149.700 MS treatmet = SS treatmet /(-1)= 8.75/(4-1)= 9.45 MS error = SS error /(N-)= 149.700/(40-4)= 4.158 F = MS amog / MS withi = 9.45/4.158 =.67 ANOVA Table for Example 3 Sum of Squares df Mea Square F Treatmets 8.75 3 9.45.67 Error 149.700 36 4.158 Total 177.975 39 5. Decisio: We FTR the ull hypothesis H0 : 1 3 4 at α=0.05 because the obtaied F.67 is less tha.86. Note that computer programs give you a p-value. Usig the p-value, you reect the ull hypothesis if the obtaied p-value is less tha or equal to 0.05. I our example the obtaied p-value is greater tha 0.05; so, we FTR the ull hypothesis. Multiple comparisos. If we reect the ull hypothesis, the how do we fid out which meas are differet? Note that we use multiple compariso tests oly after we have reected the ull hypothesis i the F test. There are may approaches to multiple compariso tests, may of them very techical. The Boferroi method is based o adusted probabilities. If you, experimeter, have meas to ( 1) compare, the there possible two-sample t-tests. So, the Boferroi method

Math 308 Sprig 018 Classes 19 ad 0: Aalysis of Variace (ANOVA) Page 6 of 6 uses istead of for the two-sided t-tests. For the Boferroi method, we actually do ( 1) all the t-tests but we use the modified value for probability of a Type I error. As stated previously, there are may differet types of multiple compariso ad follow-up tests. The Boferroi method is a respected method ad it follows from the treatmet of probability i this course Boferroi method of followig up reectio of the Null Hypothesis i ANOVA The Boferroi obtaied statistic is MS x x i error 1 1 ( ) i, i, i, 1,,...,. The critical t is obtaied by usig a two-tailed t test at the level ( 1). So for the ( 1) two-tailed test with the degrees of freedom=-, the degrees of freedom for the MS error.