Analysis of Variance (ANOVA)

Similar documents
WELCOME! Lecture 13 Thommy Perlinger

Statistics for EES Factorial analysis of variance

Analysis of variance (ANOVA) Comparing the means of more than two groups

More about Single Factor Experiments

Analysis of Variance (ANOVA) Cancer Research UK 10 th of May 2018 D.-L. Couturier / R. Nicholls / M. Fernandes

A posteriori multiple comparison tests

Solutions - Homework #2

In ANOVA the response variable is numerical and the explanatory variables are categorical.

One-Way ANOVA. Some examples of when ANOVA would be appropriate include:

Stat 6640 Solution to Midterm #2

Solutions - Homework #2

Chapters 4-6: Inference with two samples Read sections 4.2.5, 5.2, 5.3, 6.2

Lecture 7: Hypothesis Testing and ANOVA

Statistics for Managers Using Microsoft Excel Chapter 10 ANOVA and Other C-Sample Tests With Numerical Data

Design of Experiments. Factorial experiments require a lot of resources

Multiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600

STAT 115:Experimental Designs

ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS

Lecture 5: ANOVA and Correlation

Analysis of Variance

An Analysis of College Algebra Exam Scores December 14, James D Jones Math Section 01

DETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics

Lec 1: An Introduction to ANOVA

Introduction to Statistical Inference Lecture 10: ANOVA, Kruskal-Wallis Test

DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective

Stats fest Analysis of variance. Single factor ANOVA. Aims. Single factor ANOVA. Data

sphericity, 5-29, 5-32 residuals, 7-1 spread and level, 2-17 t test, 1-13 transformations, 2-15 violations, 1-19

Statistical methods for comparing multiple groups. Lecture 7: ANOVA. ANOVA: Definition. ANOVA: Concepts

STAT 263/363: Experimental Design Winter 2016/17. Lecture 1 January 9. Why perform Design of Experiments (DOE)? There are at least two reasons:

Introduction. Chapter 8

COMPARING SEVERAL MEANS: ANOVA

Inferences About the Difference Between Two Means

Chapter 10. Design of Experiments and Analysis of Variance

Week 7.1--IES 612-STA STA doc

DESAIN EKSPERIMEN Analysis of Variances (ANOVA) Semester Genap 2017/2018 Jurusan Teknik Industri Universitas Brawijaya

Stat 705: Completely randomized and complete block designs

Outline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013

Week 14 Comparing k(> 2) Populations

Chapter 8 Student Lecture Notes 8-1. Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance

Analysis of Variance

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)

Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance ECON 509. Dr.

CHI SQUARE ANALYSIS 8/18/2011 HYPOTHESIS TESTS SO FAR PARAMETRIC VS. NON-PARAMETRIC

The entire data set consists of n = 32 widgets, 8 of which were made from each of q = 4 different materials.

One-way ANOVA (Single-Factor CRD)

Unit 12: Analysis of Single Factor Experiments

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

Analysing and presenting data: practical hints

Analysing qpcr outcomes. Lecture Analysis of Variance by Dr Maartje Klapwijk

10 One-way analysis of variance (ANOVA)

H0: Tested by k-grp ANOVA

A discussion on multiple regression models

MATH Notebook 3 Spring 2018

df=degrees of freedom = n - 1

The General Linear Model

Multiple Comparisons

Hypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table

ANOVA: Comparing More Than Two Means

Chapter Seven: Multi-Sample Methods 1/52

Analysis of Covariance

Ch 2: Simple Linear Regression

4.1. Introduction: Comparing Means

Introduction to Analysis of Variance (ANOVA) Part 2

Regression: Main Ideas Setting: Quantitative outcome with a quantitative explanatory variable. Example, cont.

(Where does Ch. 7 on comparing 2 means or 2 proportions fit into this?)

1. The (dependent variable) is the variable of interest to be measured in the experiment.

Introduction to Design and Analysis of Experiments with the SAS System (Stat 7010 Lecture Notes)

Biostatistics for physicists fall Correlation Linear regression Analysis of variance

1 Introduction to One-way ANOVA

ANOVA: Analysis of Variation

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College

Chap The McGraw-Hill Companies, Inc. All rights reserved.

Example - Alfalfa (11.6.1) Lecture 14 - ANOVA cont. Alfalfa Hypotheses. Treatment Effect

Factorial designs. Experiments

The General Linear Model

General Linear Models (GLM) for Fixed Factors

What is a Hypothesis?

36-309/749 Experimental Design for Behavioral and Social Sciences. Sep. 22, 2015 Lecture 4: Linear Regression

Experimental Design and Data Analysis for Biologists

Linear Combinations of Group Means

Inference for Regression Inference about the Regression Model and Using the Regression Line

Sleep data, two drugs Ch13.xls

What Is ANOVA? Comparing Groups. One-way ANOVA. One way ANOVA (the F ratio test)

H0: Tested by k-grp ANOVA

Chapter 1 Statistical Inference

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

1. (Problem 3.4 in OLRT)

(1) The explanatory or predictor variables may be qualitative. (We ll focus on examples where this is the case.)

ANOVA CIVL 7012/8012

Lecture 15 - ANOVA cont.

Chapter 3 Multiple Regression Complete Example

Specific Differences. Lukas Meier, Seminar für Statistik

Inference with Simple Regression

STAT Chapter 10: Analysis of Variance

Degrees of freedom df=1. Limitations OR in SPSS LIM: Knowing σ and µ is unlikely in large

Analysis of Variance (ANOVA)

Inference for Regression Simple Linear Regression

Turning a research question into a statistical question.

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Transcription:

Analysis of Variance (ANOVA) Much of statistical inference centers around the ability to distinguish between two or more groups in terms of some underlying response variable y. Sometimes, there are but 2 groups (males vs. females or timing at PFI vs. 24 hours before PFI), sometimes more groups, and sometimes multiple sets of groups from more than one factor (categorical explanatory variable). These different situations & their corresponding analyses are given below. 1. 2 groups: Use a (Welch s) t-test or Wilcoxon rank sum test if normality assumptions are not met. 2. 3 or more groups, 1 factor: Use a one-way ANOVA or the Kruskal-Wallis test if normality assumptions are not met. 3. 2 or more factors: Use an n-way ANOVA to address the effects of the factors and their interactions on the response (y). The type of ANOVA performed depends on how the data being analyzed were collected. In general, there are many types of designs used in practice, all of which can be analyzed by some type of ANOVA. Some of these are: 123 1. Observational studies: weakest inference, data collected without formal randomization 2. Completely randomized designs (CRD): treatments (groups) randomly assigned to units 3. Randomized block designs (RBD): CRD s performed within heterogeneous blocks of similar units, can remove variation in the treatment by blocking 4. Factorial (crossed) designs: CRD s or RBD s on two or more factors 5. Nested designs: Designs where levels of one factor are nested within (i.e.: only have meaning within the levels of another factor 6. Repeated measures designs: Repeated measures on the same units over time 7. Split plot designs, latin squares, etc. Regardless of what type of ANOVA is needed, all ANOVA does is partition the total variation in the response variable y (called SST) into those parts explained by the various factors under study. In other words, it quantifies how much of the variability in y is due to the factors of interest. 124

Conceptual ANOVA: Why is it called analysis of variance? Example: Insects such as fruit flies rest, but is this rest sleep? One experiment to investigate this question gave caffeine to fruit flies to see if it affected their rest. We know that caffeine reduces sleep in mammals, so if it reduces rest in fruit flies, this might be a hint that the rest is really sleep. The amounts of rest (in minutes) for the four caffeine treatment groups appear below: Control 1 mg/ml 2.5 mg/ml 5 mg/ml 542 482 500 250 627 479 450 236 520 486 466 315 529 575 377 299 556 464 491 323 y 1 = 554.8 y 2 = 497.2 y 3 = 456.8 y 4 = 284.6 s 1 = 42.588 s 2 = 44.280 s 3 = 48.823 s 4 = 39.259 n 1 = 5 n 2 = 5 n 3 = 5 n 4 = 5 y = 448.35 125 We want to test for any differences among the population rest times for the 4 groups, i.e.: we want to test: Two Types of Variation Present 1. Within-sample Variation - variation within a sample MSW = from one group. How would you measure this? Letting n be the total sample size and t be the number of treatments (groups): (ni 1)s 2 i n t = (n 1 1)s 2 1 +(n 2 1)s 2 2 +(n 3 1)s 2 3 +(n 4 1)s 2 4 n t = 4(42.588)2 + +4(39.259) 2 20 4 = 1924.85 with 16 d.f. = 30797.6 16 2. Between-sample Variation - variation between the samples. How would you measure this? 1 MSB = t 1 t n i (y i y ) 2 i=1 126

= 1 3 5(554.8 448.35) 2 + +5(284.6 448.35) 2 = 1 [203016.95]=67672.32 with 3 d.f.. 3 Note that, as expected, the between group mean sum of squares is considerably larger than the within group mean SS. This indicates emphatically that there are significant differences in mean rest time between the groups relative to the variation in rest time within a group. How do we formally test for such a difference? We wish to test: H 0 :µ 1 =µ 2 =µ 3 =µ 4 µ 1 =µ 2 =µ 3 H 0 : µ 1 =µ 2 =µ 4 µ 1 =µ 4 3 =µ 4 2 µ 3 µ µ vs. H a : At least one of these is not true. Supposeσ 2 1 =σ2 2 =σ2 3 =σ2 4 (homogeneity of variance assumption). 127 Aside: We could test each of the 6 cases (H 0 :µ i =µ j vs. Ha :µ i µ j, i j) with a t-test, atα=0.05. For example, we could test H 0 :µ 1 =µ 2 vs. H a :µ 1 µ 2 using: t = = y 1 y 2 s 2 1 n 1 + s 2 2 n 2 = 57.6 27.475 = 2.096. 554.8 497.2 42.588 2 44.280 2 + 5 5 The p-value for this test is: P( T 8 >2.096)=0.069 indicating suggestive evidence of a difference in the mean rest time between the control grp. and the 1 mg/ml group. If we were to fail to reject H 0 for each case (atα=0.05), can we conclude thatµ 1 =µ 2 =µ 3 =µ 4? As the number of tests increases, the probability of falsely rejecting H 0 in at least one of those tests increases. = Combined Type I errorαfor 6 t-tests is larger than the α-level for a single test. 128

This test hinges on comparing the within-sample variance (s 2 W ) to the between-sample variance (s2 B ). Fact: Under H 0, both s 2 B & s2 W are independent unbiased estimates of the population varianceσ 2 (since then, all 16 observations are from the same population). Hence, we expect s 2 B /s2 W = 1 under H 0. If s 2 B >> s2 W (as in the fly example), we have evidence against H 0. How do we quantify the degree of evidence against H 0? Result: Under H 0, if the rest times are normal, the test statistic is: t F= s2 B s 2 F t 1, n i t. W i=1 So, large values of F= To test H 0 :µ 1 =µ 2 =µ 3 =µ 4 vs. H a : At least one is different, the test statistic is: F= s2 B s 2 = 67672.32 1924.85 = 35.16>> 9.01= F.001(3, 16), W 129 so the p-value<.001, and we reject H 0 concluding there are highly significant differences in the mean rest times. This is called the main effects F-test. Example 2: An observational study was performed in a large statistics class where the hours of sleep the previous night (y) and the seating position (front, middle, back) in the lecture hall (x 1 ) were recorded. The teacher wanted to see if there was any relationship between these two variables. For this example, I want to demonstrate how to use MatLab to run an ANOVA model. An ANOVA model was run using the anovan function below and produced the following ANOVA table: [p1,tab1,stats1,terms1] = anovan(sleep.hours,... {sleep.seat}); Source Sum Sq. d.f. Mean Sq. F Prob>F X1 9.44 2 4.71993 3.21 0.0429 Error 250.135 170 1.47138 Total 259.575 172 Constrained (Type III) sums of squares. 130

What does this indicate about any differences in mean sleeping time according to seating position? Since there appears to be an overall effect of seating position on mean sleeping time, we can also look at the magnitude and significance of all pairwise differences between the three means using the multcompare function to perform multiple comparisons of the factor level means as follows: [ci,means] = multcompare(stats1, ctype, hsd ); The [ci means] piece returns the group means and 95% (default) confidence intervals for the group means. The ctype, hsd piece on the RHS tells MatLab to use the Tukey Honest Significance Difference (HSD) 95% joint confidence intervals for the three group means. Tukey s HSD method is the most commonly used method to find joint confidence intervals when looking at pairwise differences. 131 Click on the group you want to test X1=Front X1=Middle X1=Back 6.2 6.4 6.6 6.8 7 7.2 7.4 7.6 7.8 The population marginal means of groups X1=Back and X1=Front are significantly different There are other methods of performing multiple comparisons among the group means. Some of the more popular choices are: 1. Fisher s least significant difference (LSD). 2. Bonferroni method 3. Scheffe s method The Tukey HSD method tends not to be too conservative so is generally the preferred choice of comparison among any two groups when multiple 132

133 Predicted Values 4 6.6 6.7 6.8 6.9 7 7.1 7.2 7.3 7.4 3 2 Residuals 1 0 1 2 3 4 Residual Plot below with the MatLab code following. A residual plot and normal quantile plot are shown 3. 2. 1. ANOVA model? What assumptions are being made in running this comparisons are to be made. 134 What do these tell you? Standard Normal Quantiles 4 3 2 1 0 1 2 3 3 2 Residuals 1 0 1 2 3 4 Normal Quantile Plot

figure(1) % 1st Figure yhat1 = sleep.hours-stats1.resid; plot(yhat1,stats1.resid, ko ); % Plots residuals vs. predicted y-values xlabel( Predicted Values, fontsize,14,... fontweight, b ); ylabel( Residuals, fontsize,14,... fontweight, b ); title( Residual Plot, fontsize,14,... fontweight, b ); figure(2); % 2nd Figure qqplot(stats1.resid); % Normal quantile plot of residuals xlabel( Standard Normal Quantiles,... fontsize,14, fontweight, b ); ylabel( Residuals, fontsize,14,... fontweight, b ); title( Normal Quantile Plot, fontsize,... 14, fontweight, b ); Back to the Sleep Example: Suppose in addition to the mean hours of sleep and the seating position, the gender of the student was also recorded. Here then, we have two factors (seat, gender) and want to examine their effect and 135 possible interactive effect on mean hours of sleep. In this case, we can use a 2-way ANOVA. What does such a model really look like? Let: y i = the mean hours of sleep for the i th student, x 1i = x 2i = x 3i = i= 1, 2,..., 173, 1 if the i th seat is in the Front 0 otherwise 1 if the i th seat is in the Middle 0 otherwise, 1 if the i th student is Female 0 otherwise,, x 4i = x 1i x 3i, x 5i = x 2i x 3i. Then, this 2-way ANOVA model can be written as: y i =β 0 +β 1 x 1i +β 2 x 2i +β 3 x 3i +β 4 x 4i +β 5 x 5i +ε i. Although there were 173 subjects in this sleep study, suppose for a moment that there were but 12 subjects with 136

2 at each of the seat/gender combinations. In matrix form, the model would appear as follows: y 1 1 1 0 1 1 0 1 1 0 1 1 0 y 2 y 3 y 4 y 5 y 6 y 7 y 8 y 9 y 10 y 11 y 12 = 1 1 0 0 0 0 1 1 0 0 0 0 1 0 1 1 0 1 1 0 1 1 0 1 1 0 1 0 0 0 1 0 1 0 0 0 1 0 0 1 0 0 1 0 0 1 0 0 1 0 0 0 0 0 1 0 0 0 0 0 β 0 β 1 β 2 β 3 β 4 β 5 + ε 1 ε 2 ε 3 ε 4 ε 5 ε 6 ε 7 ε 8 ε 9 ε 10 ε 11 ε 12 Hence, any ANOVA model is nothing more than a linear model with the design matrix defined in the right way. Specifically, continuous variables (such as mean hours of sleep) contain the values themselves, whereas categorical variables (such as seat & gender) 137 can be expressed as indicator variables. If a categorical variable has t levels, then we need t 1 indicator variables to describe the possible outcomes for that variable. Note that in the sleep example, a first-order model with an interaction requires 6 parameters to be estimated. In general, for 2 factors at a levels and b levels respectively, the model contains ab parameters, and hence there are ab degrees of freedom in the explanatory part of the model. Although we could run any ANOVA model as a general linear model, software is written to construct these indicator variables behind the scenes and simply report the ANOVA table with the relevant components to allow the user to examine the effects of the factors and their interactions on the response variable. Yeah!!! Before running the 2-way ANOVA model, we first compute some basic group statistics and look at some plots of the main effects as well as an interaction plot expressing the relationship between seat and gender on mean hours of sleep. 138

% Computes group statistics for both seat % location and gender [order numel means se] = grpstats(sleep.hours,... {sleep.seat sleep.gender},... { gname, numel, mean, sem },.05); order = numel = means = se = Front Female 24 7.1250 0.2762 Front Male 20 6.0000 0.2206 Middle Female 37 7.0541 0.1547 Middle Male 31 6.8710 0.2086 Back Female 29 7.0690 0.1836 Back Male 32 7.3594 0.2733 139 mean % Constructs plots of the group means for % both groups separately maineffectsplot(sleep.hours,{sleep.seat,... sleep.gender}, varnames,... { Seat Location, Gender }); 7.2 7.1 7 6.9 6.8 6.7 6.6 Front Middle Back Seat Location Female Male Gender 140

7 6.5 6 % Constructs an interaction plot of seat location % and gender on sleep hours interactionplot(sleep.hours,... {sleep.seat sleep.gender},... varnames,{ Seating Location, Gender }); Female Male 7 Seating Location = Front Seating Location = Middle Seating Location = Back Seating Location 6.5 6 Gender = Female Gender = Male Gender Front Middle Back 141 Finally, the 2-way ANOVA is run with the following MatLab code and produces the ANOVA table which follows. Are the results surprising given the plots created? % Runs the 2-way ANOVA with an interaction (2) [p1,tab1,stats1,terms1] = anovan(sleep.hours,... {sleep.seat,sleep.gender},... varnames,{ Seat, Gender },... model, interaction ); Source Sum Sq. d.f. Mean Sq. F Prob>F ------------------------------------------------ Seat 10.809 2 5.40474 3.85 0.023208 Gender 4.78 1 4.7796 3.4 0.066806 Seat*Gender 12.87 2 6.43489 4.58 0.011543 Error 234.48 167 1.40407 Total 259.575 172 ------------------------------------------------ Constrained (Type III) sums of squares. What can we say about a gender effect? Do the various sums of squares sum to the total sum of squares? 142

Multiple comparisons among the different levels of individual groups and between levels of combinations of groups can be made to compare group means. If some of the explanatory variables are continuous, we can include these as additional variables in the model as well. Formally, if an experiment is performed where units are randomly assigned to combinations of the (categorical) model factors and then measurements are made on the response variable y and other continuous covariates, the resulting analysis is known as an analysis of covariance (ANCOVA). 143