ANOVA Analysis of Variance

Similar documents
Difference in two or more average scores in different groups

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)

The One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

One-Way ANOVA Source Table J - 1 SS B / J - 1 MS B /MS W. Pairwise Post-Hoc Comparisons of Means

Prepared by: Prof. Dr Bahaman Abu Samah Department of Professional Development and Continuing Education Faculty of Educational Studies Universiti

N J SS W /df W N - 1

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization.

One-way between-subjects ANOVA. Comparing three or more independent means

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

Analysis of Variance

Analysis of variance

Multiple Comparisons

ANOVA: Comparing More Than Two Means

Introduction to Business Statistics QM 220 Chapter 12

Data Analysis and Statistical Methods Statistics 651

One-Way Analysis of Variance. With regression, we related two quantitative, typically continuous variables.

Statistical methods for comparing multiple groups. Lecture 7: ANOVA. ANOVA: Definition. ANOVA: Concepts

Two-Way ANOVA. Chapter 15

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

Comparing Several Means: ANOVA

Econ 3790: Business and Economic Statistics. Instructor: Yogesh Uppal

Lec 1: An Introduction to ANOVA

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)

Introduction to Analysis of Variance. Chapter 11

Notes for Week 13 Analysis of Variance (ANOVA) continued WEEK 13 page 1

Analysis of Variance (ANOVA)

Analysis of variance

Statistiek II. John Nerbonne using reworkings by Hartmut Fitz and Wilbert Heeringa. February 13, Dept of Information Science

ANOVA: Comparing More Than Two Means

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants.

Multiple comparisons - subsequent inferences for two-way ANOVA

Disadvantages of using many pooled t procedures. The sampling distribution of the sample means. The variability between the sample means

Unit 27 One-Way Analysis of Variance

Statistics for Managers Using Microsoft Excel Chapter 10 ANOVA and Other C-Sample Tests With Numerical Data

CHAPTER 4 Analysis of Variance. One-way ANOVA Two-way ANOVA i) Two way ANOVA without replication ii) Two way ANOVA with replication

Hypothesis Testing. Hypothesis: conjecture, proposition or statement based on published literature, data, or a theory that may or may not be true

Analysis of Variance (ANOVA)

Note: k = the # of conditions n = # of data points in a condition N = total # of data points

Hypothesis testing: Steps

Chapter 10: Analysis of variance (ANOVA)

ANOVA continued. Chapter 10

Tukey Complete Pairwise Post-Hoc Comparison

SEVERAL μs AND MEDIANS: MORE ISSUES. Business Statistics

4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES

What If There Are More Than. Two Factor Levels?

16.3 One-Way ANOVA: The Procedure

WELCOME! Lecture 13 Thommy Perlinger

13: Additional ANOVA Topics

Hypothesis testing: Steps

STAT 135 Lab 9 Multiple Testing, One-Way ANOVA and Kruskal-Wallis

One-way between-subjects ANOVA. Comparing three or more independent means

Introduction to the Analysis of Variance (ANOVA)

ANOVA continued. Chapter 10

22s:152 Applied Linear Regression. 1-way ANOVA visual:

ANOVA continued. Chapter 11

Levene's Test of Equality of Error Variances a

STAT 263/363: Experimental Design Winter 2016/17. Lecture 1 January 9. Why perform Design of Experiments (DOE)? There are at least two reasons:

22s:152 Applied Linear Regression. Take random samples from each of m populations.

T-Test QUESTION T-TEST GROUPS = sex(1 2) /MISSING = ANALYSIS /VARIABLES = quiz1 quiz2 quiz3 quiz4 quiz5 final total /CRITERIA = CI(.95).

An inferential procedure to use sample data to understand a population Procedures

STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5)

1. The (dependent variable) is the variable of interest to be measured in the experiment.

1 DV is normally distributed in the population for each level of the within-subjects factor 2 The population variances of the difference scores

22s:152 Applied Linear Regression. There are a couple commonly used models for a one-way ANOVA with m groups. Chapter 8: ANOVA

Analysis of Variance: Repeated measures

STATS Analysis of variance: ANOVA

Oct Simple linear regression. Minimum mean square error prediction. Univariate. regression. Calculating intercept and slope

An Analysis of College Algebra Exam Scores December 14, James D Jones Math Section 01

One-Way ANOVA. Some examples of when ANOVA would be appropriate include:

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

Independent Samples ANOVA

M A N O V A. Multivariate ANOVA. Data

These are all actually contrasts (the coef sum to zero). What are these contrasts representing? What would make them large?

One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups

DESAIN EKSPERIMEN Analysis of Variances (ANOVA) Semester Genap 2017/2018 Jurusan Teknik Industri Universitas Brawijaya

Linear Combinations of Group Means

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

1 The Randomized Block Design

Chapter 10: STATISTICAL INFERENCE FOR TWO SAMPLES. Part 1: Hypothesis tests on a µ 1 µ 2 for independent groups

Two-Way Analysis of Variance - no interaction

Multiple Testing. Gary W. Oehlert. January 28, School of Statistics University of Minnesota

Contrasts (in general)

Chapter 11 - Lecture 1 Single Factor ANOVA

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)

CHAPTER 10 Comparing Two Populations or Groups

CHAPTER 10 Comparing Two Populations or Groups

Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model

COMPARING SEVERAL MEANS: ANOVA

(Where does Ch. 7 on comparing 2 means or 2 proportions fit into this?)

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table

Week 14 Comparing k(> 2) Populations

What Does the F-Ratio Tell Us?

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =

Analysis of Variance and Contrasts

ANALYTICAL COMPARISONS AMONG TREATMENT MEANS (CHAPTER 4)

Transcription:

ANOVA Analysis of Variance

ANOVA Analysis of Variance Extends independent samples t test

ANOVA Analysis of Variance Extends independent samples t test Compares the means of groups of independent observations

ANOVA Analysis of Variance Extends independent samples t test Compares the means of groups of independent observations Don t be fooled by the name. ANOVA does not compare variances.

ANOVA Analysis of Variance Extends independent samples t test Compares the means of groups of independent observations Don t be fooled by the name. ANOVA does not compare variances. Can compare more than two groups

ANOVA Null and Alternative Hypotheses Say the sample contains K independent groups

ANOVA Null and Alternative Hypotheses Say the sample contains K independent groups ANOVA tests the null hypothesis H 0 : μ 1 = μ 2 = = μ K That is, the group means are all equal

ANOVA Null and Alternative Hypotheses Say the sample contains K independent groups ANOVA tests the null hypothesis H 0 : μ 1 = μ 2 = = μ K That is, the group means are all equal The alternative hypothesis is H 1 : μ i μ j for some i, j or, the group means are not all equal

Example: Accuracy of Implant Placement Implants were placed in a manikin using placement guides of various widths. 15 implants were placed using each guide. Error (discrepancies with a reference implant) was measured for each implant.

Example: Accuracy of Implant Placement The overall mean of the entire sample was 0.248 mm. This is called the grand mean, and is often denoted by X. If H 0 were true then we d expect the group means to be close to the grand mean.

Example: Accuracy of Implant Placement The ANOVA test is based on the combined distances from X. If the combined distances are large, that indicates we should reject H 0.

The Anova Statistic To combine the differences from the grand mean we Square the differences Multiply by the numbers of observations in the groups Sum over the groups ( ) 2 ( ) 2 ( ) X X + 15 X X + X 2 SSB = 15 X 4mm 6mm 15 where the are the group means. X * 8mm SSB = Sum of Squares Between groups

The Anova Statistic To combine the differences from the grand mean we Square the differences Multiply by the numbers of observations in the groups Sum over the groups ( ) 2 ( ) 2 ( ) X X + 15 X X + X 2 SSB = 15 X 4mm 6mm 15 where the are the group means. X * 8mm SSB = Sum of Squares Between groups Note: This looks a bit like a variance.

How big is big? For the Implant Accuracy Data, SSB = 0.0047 Is that big enough to reject H 0? As with the t test, we compare the statistic to the variability of the individual observations. In ANOVA the variability is estimated by the Mean Square Error, or MSE

MSE Mean Square Error The Mean Square Error is a measure of the variability after the group effects have been taken into account. 1 MSE = N K ( x ij X j ) where x ij is the i th observation in the j th group. j i 2

MSE Mean Square Error The Mean Square Error is a measure of the variability after the group effects have been taken into account. 1 MSE = N K ( x ij X j ) where x ij is the i th observation in the j th group. j i 2

MSE Mean Square Error The Mean Square Error is a measure of the variability after the group effects have been taken into account. 1 MSE = N K ( x ij X j ) Note that the variation of the means seems quite small compared to the variance of observations within groups j i 2

Notes on MSE If there are only two groups, the MSE is equal to the pooled estimate of variance used in the equalvariance t test. ANOVA assumes that all the group variances are equal. Other options should be considered if group variances differ by a factor of 2 or more.

ANOVA F Test The ANOVA F test is based on the F statistic F = SSB ( K 1) MSE where K is the number of groups. Under H 0 the F statistic has an F distribution, with K 1 and N K degrees of freedom (N is the total number of observations)

Implant Data: F test p value To get a p value we compare our F statistic to an F(2, 42) distribution.

Implant Data: F test p value To get a p value we compare our F statistic to an F(2, 42) distribution. In our example.0047 2 F =.0467 42 =.211

Implant Data: F test p value To get a p value we compare our F statistic to an F(2, 42) distribution. In our example.0047 2 F =.0467 42 =.211 The p value is P ( F(2,42) >. 211) = 0. 81

ANOVA Table Results are often displayed using an ANOVA Table Sum of Squares df Mean Square F Sig. Between Groups.005 2.002.211.811 Within Groups.466 42.011 Total.470 44

ANOVA Table Results are often displayed using an ANOVA Table Sum of Squares df Mean Square F Sig. Between Groups.005 2.002.211.811 Within Groups.466 42.011 Total.470 44 Pop Quiz!: Where are the following quantities presented in this table? Sum of Squares Between (SSB) Mean Square Error (MSE) F Statistic p value

ANOVA Table Results are often displayed using an ANOVA Table Sum of Squares df Mean Square F Sig. Between Groups.005 2.002.211.811 Within Groups.466 42.011 Total.470 44 Sum of Squares Between (SSB) Mean Square Error (MSE) F Statistic p value

ANOVA Table Results are often displayed using an ANOVA Table Sum of Squares df Mean Square F Sig. Between Groups.005 2.002.211.811 Within Groups.466 42.011 Total.470 44 Sum of Squares Between (SSB) Mean Square Error (MSE) F Statistic p value

ANOVA Table Results are often displayed using an ANOVA Table Sum of Squares df Mean Square F Sig. Between Groups.005 2.002.211.811 Within Groups.466 42.011 Total.470 44 Sum of Squares Between (SSB) Mean Square Error (MSE) F Statistic p value

ANOVA Table Results are often displayed using an ANOVA Table Sum of Squares df Mean Square F Sig. Between Groups.005 2.002.211.811 Within Groups.466 42.011 Total.470 44 Sum of Squares Between (SSB) Mean Square Error (MSE) F Statistic p value

Post Hoc Tests NHANES I data, women 40-60 yrs old. Compare cholesterol between periodontal groups. The ANOVA shows good evidence (p = 0.002) that the means are not all the same. Sum of Squares Mean Square F Sig. df Between Groups 33383 3 11128 5.1.002 Within Groups 4417119 2007 2201 Total 4450502 2010 Which means are different? Can directly compare the subgroups using post hoc tests.

Least Significant Difference test Std. N Mean Deviation Healthy 802 221.5 46. 2 Gingivitis 490 223.5 45.3 Periodontitis 347 227.3 48.9 Edentulous 372 232.4 48. 8 Sum of Squares Mean Square F Sig. df Between Groups 33383 3 11128 5.1.002 Within Groups 4417119 2007 2201 Total 4450502 2010 The most simple post hoc test is called the Least Significant Difference Test. The computation is very similar to the equalvariance t test. Compute an equal-variance t test, but replace the pooled variance (s 2 ) with the MSE.

Least Significant Difference Test: Examples Std. N Mean Deviation Healthy 802 221.5 46. 2 Gingivitis 490 223.5 45.3 Periodontitis 347 227.3 48.9 Edentulous 372 232.4 48. 8 Compare Healthy group to Periodontitis group: T = 221.5 227.3 2201 1 ( 802 + 1 347) = 1.92 p = 2 P( t1147 > 1.92) = 0.055 Sum of Squares Mean Square F Sig. df Between Groups 33383 3 11128 5.1.002 Within Groups 4417119 2007 2201 Total 4450502 2010 Compare Gingivitis group to Periodontitis group: 223.5 227.3 T = = 1.15 2201 1 490 + 1 347 ( ) p = 2 P( t835 > 1.15) = 0.25

Post Hoc Tests: Multiple Comparisons Post hoc testing usually involves multiple comparisons.

Post Hoc Tests: Multiple Comparisons Post hoc testing usually involves multiple comparisons. For example, if the data contain 4 groups, then 6 different pairwise comparisons can be made Healthy Gingivitis Periodontitis Edentulous

Post Hoc Tests: Multiple Comparisons Post hoc testing usually involves multiple comparisons. For example, if the data contain 4 groups, then 6 different pairwise comparisons can be made Healthy Gingivitis Periodontitis Edentulous

Post Hoc Tests: Multiple Comparisons Post hoc testing usually involves multiple comparisons. For example, if the data contain 4 groups, then 6 different pairwise comparisons can be made Healthy Gingivitis Periodontitis Edentulous

Post Hoc Tests: Multiple Comparisons Post hoc testing usually involves multiple comparisons. For example, if the data contain 4 groups, then 6 different pairwise comparisons can be made Healthy Gingivitis Periodontitis Edentulous

Post Hoc Tests: Multiple Comparisons Post hoc testing usually involves multiple comparisons. For example, if the data contain 4 groups, then 6 different pairwise comparisons can be made Healthy Gingivitis Periodontitis Edentulous

Post Hoc Tests: Multiple Comparisons Post hoc testing usually involves multiple comparisons. For example, if the data contain 4 groups, then 6 different pairwise comparisons can be made Healthy Gingivitis Periodontitis Edentulous

Post Hoc Tests: Multiple Comparisons Post hoc testing usually involves multiple comparisons. For example, if the data contain 4 groups, then 6 different pairwise comparisons can be made Healthy Gingivitis Periodontitis Edentulous

Post Hoc Tests: Multiple Comparisons Each time a hypothesis test is performed at significance level α, there is probability α of rejecting in error. Performing multiple tests increases the chances of rejecting in error at least once.

Post Hoc Tests: Multiple Comparisons Each time a hypothesis test is performed at significance level α, there is probability α of rejecting in error. Performing multiple tests increases the chances of rejecting in error at least once. For example: if you did 6 independent hypothesis tests at the α = 0.05

Post Hoc Tests: Multiple Comparisons Each time a hypothesis test is performed at significance level α, there is probability α of rejecting in error. Performing multiple tests increases the chances of rejecting in error at least once. For example: if you did 6 independent hypothesis tests at the α = 0.05 If, in truth, H 0 were true for all six.

Post Hoc Tests: Multiple Comparisons Each time a hypothesis test is performed at significance level α, there is probability α of rejecting in error. Performing multiple tests increases the chances of rejecting in error at least once. For example: if you did 6 independent hypothesis tests at the α = 0.05 If, in truth, H 0 were true for all six. The probability that at least one test rejects H 0 is 26%

Post Hoc Tests: Multiple Comparisons Each time a hypothesis test is performed at significance level α, there is probability α of rejecting in error. Performing multiple tests increases the chances of rejecting in error at least once. For example: if you did 6 independent hypothesis tests at the α = 0.05 If, in truth, H 0 were true for all six. The probability that at least one test rejects H 0 is 26% P(at least one rejection) = 1 P(no rejections) = 1.95 6 =.26

Bonferroni Correction for Multiple Comparisons The Bonferroni correction is a simple way to adjust for the multiple comparisons. Bonferroni Correction Perform each test at significance level α. Multiply each p-value by the number of tests performed. The overall significance level (chance of any of the tests rejecting in error) will be less than α.

Example: Cholesterol Data post hoc comparisons Group 1 Group 2 Mean Difference (Group 1 - Group 2) Least Significant Difference p-value Bonferroni p-value Healthy Gingivitis -2.0.46 1.0 Healthy Periodontitis -5.8.055.330 Healthy Edentulous -10.9.00021.00126 Gingivitis Periodontitis -3.9.25 1.0 Gingivitis Edentulous -8.9.0056.0336 Periodontitis Edentulous -5.1.147.88

Example: Cholesterol Data post hoc comparisons Group 1 Group 2 Mean Difference (Group 1 - Group 2) Least Significant Difference p-value Bonferroni p-value Healthy Gingivitis -2.0.46 1.0 Healthy Periodontitis -5.8.055.330 Healthy Edentulous -10.9.00021.00126 Gingivitis Periodontitis -3.9.25 1.0 Gingivitis Edentulous -8.9.0056.0336 Periodontitis Edentulous -5.1.147.88 Conclusion: The Edentulous group is significantly different than the Healthy group and the Gingivitis group (p < 0.05), after adjustment for multiple comparisons