One-way Analysis of Variance. Major Points. T-test. Ψ320 Ainsworth

Similar documents
10/31/2012. One-Way ANOVA F-test

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

One-way ANOVA. Experimental Design. One-way ANOVA

8/23/2018. One-Way ANOVA F-test. 1. Situation/hypotheses. 2. Test statistic. 3.Distribution. 4. Assumptions

One-way between-subjects ANOVA. Comparing three or more independent means

The One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)

Analysis of Variance

INTRODUCTION TO ANALYSIS OF VARIANCE

Introduction to the Analysis of Variance (ANOVA)

One-way between-subjects ANOVA. Comparing three or more independent means

Hypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -

psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

One-Way ANOVA Cohen Chapter 12 EDUC/PSY 6600

Analysis of Variance (ANOVA)

Calculating Fobt for all possible combinations of variances for each sample Calculating the probability of (F) for each different value of Fobt

Factorial Within Subjects. Psy 420 Ainsworth

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

Introduction to Analysis of Variance. Chapter 11

Notes for Week 13 Analysis of Variance (ANOVA) continued WEEK 13 page 1

STAT 115:Experimental Designs

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table

Unit 27 One-Way Analysis of Variance

8/28/2017. Repeated-Measures ANOVA. 1. Situation/hypotheses. 2. Test statistic. 3.Distribution. 4. Assumptions

Comparing Several Means: ANOVA

ANCOVA. Lecture 9 Andrew Ainsworth

The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization.

Lecture 6: Single-classification multivariate ANOVA (k-group( MANOVA)

DETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics

Battery Life. Factory

1. What does the alternate hypothesis ask for a one-way between-subjects analysis of variance?

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants.

An Old Research Question

We need to define some concepts that are used in experiments.

COMPARING SEVERAL MEANS: ANOVA

Analysis of Variance: Part 1

Chap The McGraw-Hill Companies, Inc. All rights reserved.

Solutions to Final STAT 421, Fall 2008

Analysis of Variance. One-way analysis of variance (anova) shows whether j groups have significantly different means. X=10 S=1.14 X=5.8 X=2.4 X=9.

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

ANCOVA. Psy 420 Andrew Ainsworth

Introduction to Analysis of Variance (ANOVA) Part 2

DESAIN EKSPERIMEN Analysis of Variances (ANOVA) Semester Genap 2017/2018 Jurusan Teknik Industri Universitas Brawijaya

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =

What If There Are More Than. Two Factor Levels?

Multiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600

Analysis of Variance and Contrasts

13: Additional ANOVA Topics

One-Way ANOVA Source Table J - 1 SS B / J - 1 MS B /MS W. Pairwise Post-Hoc Comparisons of Means

Hypothesis Testing hypothesis testing approach

Research Methods II MICHAEL BERNSTEIN CS 376

20.0 Experimental Design

Group comparison test for independent samples

Statistics For Economics & Business

Chapter 10: Analysis of variance (ANOVA)

PLSC PRACTICE TEST ONE

IEOR165 Discussion Week 12

STAT Chapter 10: Analysis of Variance

Chapter 14: Repeated-measures designs

Econ 3790: Business and Economic Statistics. Instructor: Yogesh Uppal

WELCOME! Lecture 13 Thommy Perlinger

COMPLETELY RANDOM DESIGN (CRD) -Design can be used when experimental units are essentially homogeneous.

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018

Extending the Robust Means Modeling Framework. Alyssa Counsell, Phil Chalmers, Matt Sigal, Rob Cribbie

Comparing the means of more than two groups

Multiple Comparisons

Lecture 7: Hypothesis Testing and ANOVA

One-Way Analysis of Variance. With regression, we related two quantitative, typically continuous variables.

CHAPTER 4 Analysis of Variance. One-way ANOVA Two-way ANOVA i) Two way ANOVA without replication ii) Two way ANOVA with replication

Chapter Seven: Multi-Sample Methods 1/52

N J SS W /df W N - 1

Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance ECON 509. Dr.

Analysis of variance

ANOVA: Comparing More Than Two Means

Repeated Measures Analysis of Variance

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI

Explanation of R 2, and Other Stories

Hypothesis testing: Steps

Multiple comparisons - subsequent inferences for two-way ANOVA

Sociology 593 Exam 1 Answer Key February 17, 1995

Contrasts and Multiple Comparisons Supplement for Pages

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

Variance Estimates and the F Ratio. ERSH 8310 Lecture 3 September 2, 2009

Chapter 8 Student Lecture Notes 8-1. Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance

Week 14 Comparing k(> 2) Populations

Note: k = the # of conditions n = # of data points in a condition N = total # of data points

Hypothesis testing: Steps

Analysis of Variance ANOVA. What We Will Cover in This Section. Situation

H0: Tested by k-grp ANOVA

Much of the material we will be covering for a while has to do with designing an experimental study that concerns some phenomenon of interest.

QUEEN MARY, UNIVERSITY OF LONDON


What Is ANOVA? Comparing Groups. One-way ANOVA. One way ANOVA (the F ratio test)

Analysis of Variance. ภาว น ศ ร ประภาน ก ล คณะเศรษฐศาสตร มหาว ทยาล ยธรรมศาสตร

While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1

Statistical methods for comparing multiple groups. Lecture 7: ANOVA. ANOVA: Definition. ANOVA: Concepts

H0: Tested by k-grp ANOVA

Transcription:

One-way Analysis of Variance Ψ30 Ainsworth Major Points Problem with t-tests and multiple groups The logic behind ANOVA Calculations Multiple comparisons Assumptions of analysis of variance Effect Size for ANOVA T-test So far, we have made comparisons between a single group and population, -related samples and independent samples What if we want to compare more than groups? One solution: multiple t-tests 3

T-test With 3 groups, you would perform 3 t- tests Not so bad, but what if you had 10 groups? You would need 45 comparisons to analyze all pairs That s right 45!!! 4 The Danger of Multiple t-tests Each time you conduct a t-test on a single set of data, what is the probability of rejecting a true null hypothesis? Assume that H 0 is true. You are conducting 45 tests on the same set of data. How many rejections will you have? Roughly or 3 false rejections! So, multiple t-tests on the same set of data artificially inflate α 5 Summary: The Problems With Multiple t-tests Inefficient - too many comparisons when we have even modest numbers of groups. Imprecise - cannot discern patterns or trends of differences in subsets of groups. Inaccurate - multiple tests on the same set of data artificially inflate α What is needed: a single test for the overall difference among all means e.g. ANOVA 6

Logic of the Analysis of Variance Null hypothesis h 0 : Population means equal µ 1 = µ = µ 3 = µ 4 Alternative hypothesis: h 1 Not all population means equal. 7 Logic Create a measure of variability among group means MS BetweenGroups AKA s BetweenGroups Create a measure of variability within groups MS WithinGroups AKA s WithinGroups 8 Logic MS BetweenGroups /MS WithinGroups Ratio approximately 1 if null true Ratio significantly larger than 1 if null false approximately 1 can actually be as high as or 3, but not much higher 9

So, why is it called analysis of variance anyway? Aren t we interested in mean differences? Variance revisited Basic variance formula s ( ) X X SS = = n 1 df 10 Why is it called analysis of variance anyway? What if data comes from groups? We can have different sums of squares ( ij ) ( ij j ) j ( j ) SS = Y Y 1.. SS = Y Y. SS3 = n Y. Y.. Where i represents the individual and j represents the groups 11 Logic of ANOVA Grand Mean (Ungrouped Mean) X John s Score YGroup1 Y Group Y Group3 X-Axis 1

Sums of Squares The total variability can be partitioned into between groups variability and within groups variability. ( Yij Y.. ) = nj ( Y. j Y.. ) + ( Yij Y. j ) SSTotal = SSBetweenGroups + SSWithinGroups SST = SSBG + SSWG SS = SS + SS T Effect Error 13 Degrees of Freedom (df ) Number of observations free to vary df T = N -1 Variability of N observations df BG = g -1 Variability of g means df WG = g (n -1) or N - g n observations in each group = n -1 df times g groups df T = df BG + df WG 14 Mean Square (i.e. Variance) MS MS MS T BG WG = s = T = s = BG = s = WG ( Yij Y.. ) N 1 j (. j..) n Y Y g 1 ( Yij Y. j ) gn ( 1) 15

F-test MS WG contains random sampling variation among the participants MS BG also contains random sampling variation but it can also contain systematic (real) variation between the groups (either naturally occurring or manipulated) 16 F-test Grand Mean (Ungrouped Mean) Y Y Y X-Axis The F-test is a ratio of the MS BG /MS WG and if the group differences are just random the ratio will equal 1 (e.g. random/random) 17 F-test YGroup1 Y Group Y Group3 If there are real differences between the groups the difference will be larger than 1 and we can calculate the probability and hypothesis test 18

F distribution 1 F-ratio There is a separate F distribution for every df like t but we need both df bg and df wg to calculate the F CV from the F table E.3 for alpha =.05 and E.4 for alpha =.01 19 Example A researcher is interested in knowing which brand of baby food babies prefer: Beechnut, Del Monte or Gerber. He randomly selects 15 babies and assigns each to try strained pees from one of the three brands Liking is measured by the number of spoonfuls the baby takes before getting upset (e.g. crying, screaming, throwing the food, etc.) 0 Hypothesis Testing 1. H o : µ Beechnut = µ Del Monte = µ Gerber. At least µs are different 3. α =.05 4. More than groups ANOVA F 5. For F cv you need both df BG = 3 1 = and df WG = g (n - 1) = 3(5 1) = 1 Table E.3 F cv (,1) = 3.89, if F o > 3.89 reject the null hypothesis 1

Step 6 Calculate F-test Start with Sum of Squares (SS) We need: SS T SS BG SS WG Then, use the SS and df to compute mean squares and F `Brand Baby Spoonfuls (Y) Group Means ( Y Y ) ( Y Y ) n ( Y Y ) Beechnut Del Monte Gerber Step 6 Calculate F-test ij.. ij. j j. j.. 1 3 (3-6.333) = 11.109.56 4 5.443 (4-4.6) = 0.36 3 4 4.6 5.443 0.36 [5 * (4.6-6.333) ] = 15.015 4 4 5.443 0.36 5 8.779 11.56 6 7 0.445 1 7 4 5.443 4 8 8 6.779 4 9 6 0.111 0 10 5 1.777 1 11 9 7.113 0.36 1 6 0.111 5.76 13 10 8.4 13.447.56 14 8.779 0.16 15 9 7.113 0.36 [5 * (6-6.333) ] = 0.555 [5 * (8.4-6.333) ] = 1.36 Mean 6.333 Sum 71.335 34.4 36.93 ANOVA summary table and Step 7 Source SS df MS F BG 36.93 18.465 6.441 WG 34.4 1.867 Total 71.335 14 Remember MS = SS/df F = MS BG /MS WG Step 7 Since 6.441 > 3.89, reject the null hypothesis 4

Conclusions The F for groups is significant. We would obtain an F of this size, when H 0 true, less than 5% of the time. The difference in group means cannot be explained by random error. The baby food brands were rated differently by the sample of babies. 5 Alternative Analysis computational approach to SS Equations ( Y ) T SS Y Y N ( ) aj T SSBG = n N ( a ) = T = = j SSWG Y n Under each part of the equations, you divide by the number of scores it took to get the number in the numerator N 6 Computational Approach Example Brand Baby Spoonfuls (Y) 1 3 4 3 4 4 4 5 8 Sum 3 6 7 7 4 8 8 9 6 10 5 Sum 30 11 9 1 6 13 10 14 8 15 9 Sum 4 Beechnut Del Monte Gerber Total Sum Y Squared 95 673 T 95 SST = Y = 673 = 71.33 N 15 ( ) aj T 3 + 30 + 4 95 SSBG = = = n N 5 15 = 638.6 601.667 = 36.93 ( a ) j SSWG = Y = n 3 + 30 + 4 = 673 = 673 638.6 = 34.4 5 Note: You get the same SS using this method

Unequal Sample Sizes With one-way, no particular problem Multiply mean deviations by appropriate n i as you go The problem is more complex with more complex designs, as shown in next chapter. Equal samples only simplify the equation because when n 1 = n = = n g (...) = (...) n Y Y n Y Y j j j 8 Multiple Comparisons Significant F only shows that not all groups are equal We want to know what groups are different. Such procedures are designed to control familywise error rate. Familywise error rate defined Contrast with per comparison error rate 9 More on Error Rates Most tests reduce significance level (α) for each t test. The more tests we run the more likely we are to make Type I error. Good reason to hold down number of tests 30

Fisher s LSD Procedure Requires significant overall F, or no tests Run standard t tests between pairs of groups. Often we replace s pooled with MS error from overall analysis It is really just a pooled error term, but with more degrees of freedom (pooled across all treatment groups) 31 Fisher s LSD Procedure Comparing Beechnut and Gerber * MSWG MSWG.867.867 s = + = + XBeechnut XGerber nbeechnut ngerber 5 5 X X 8.4 4.6 t = = = 3.55 1.071 Beechnut Gerber * sxbeechnut XGerber t cv (5+5-=8) α =.05 = 1.860 Since 3.55 > 1.860, the groups are significantly different. 3 Bonferroni t Test Run t tests between pairs of groups, as usual Hold down number of t tests Reject if t exceeds critical value in Bonferroni table Works by using a more strict value of α for each comparison 33

Bonferroni t Critical value of α for each test set at.05/c, where c = number of tests run Assuming familywise α =.05 e. g. with 3 tests, each t must be significant at.05/3 =.0167 level. With computer printout, just make sure calculated probability <.05/c 34 Assumptions for Analysis of Variance Assume: Observations normally distributed within each population Population variances are equal Homogeneity of variance or homoscedasticity Observations are independent 35 Assumptions Analysis of variance is generally robust to first two A robust test is one that is not greatly affected by violations of assumptions. 36

Magnitude of Effect Eta squared (η ) Easy to calculate Somewhat biased on the high side Formula SSBG η = SS Total Percent of variation in the data that can be attributed to treatment differences 37 Magnitude of Effect Omega squared (ω ) Much less biased than η Not as intuitive We adjust both numerator and denominator with MS error Formula SSBG ( k 1) MS ω = SS + MS T WG WG 38 η and ω for Foa, et al. SSBG 36.93 η = = =.50 SST 74.335 SSBG ( k 1) MSWG 36.93 (.867) ω = = =.40 SS + MS 74.335 +.867 T WG η =.50: 50% of variability in symptoms can be accounted for by treatment ω =.40: This is a less biased estimate, and note that it is 5% smaller. 39

Other Measures of Effect Size We can use the same kinds of measures we talked about with t tests (e.g. d and γ) Usually makes most sense to talk about groups at a time or effect size between the largest and smallest groups, etc. And there are methods for converting η to d and vice versa 40