One-factor analysis of variance (ANOVA)

Size: px
Start display at page:

Download "One-factor analysis of variance (ANOVA)"

Transcription

1 One-factor analysis of variance (ANOVA) March 1, 2017 psych10.stanford.edu

2 Announcements / Action Items Schedule update: final R lab moved to Week 10 Optional Survey 5 coming soon, due on Saturday

3 Last time The surprising-ness of an event depends on whether we were interested in that specific event occurring or one of many possible events occurring implications for multiple comparisons in hypothesis testing Choosing the correct conditional probability matters We can use Bayes rule to find these correct conditional probabilities from partial information We need to account for whether our observations are independent

4 This time Recap: what types of questions do we ask when we compare two means? How much variance should we expect between group means? How can we compare multiple means?

5 This time Recap: what types of questions do we ask when we compare two means? How much variance should we expect between group means? How can we compare multiple means?

6 Memorizing letters Can people memorize more letters when they are grouped to have conceptual meaning than when they are not? n = 50 college students were given a sequence of thirty letters to memorize in twenty seconds grouping variable: randomly assigned to either recognizable three-letter groupings: JFK-CIA-FBI-USA- unrecognizable grouping: JFKC-IAF-BIU- response variable: immediately listed as many letters as they could remember, score was number of letters correctly listed before first mistake

7 What might we want to know? What s the best guess for # of items memorized, given that a person (past or future) is in a specific group? What our best guess for the difference between (population) group means? What are our the lowest and highest plausible guesses for the difference between (population) group means? Would it be likely to observe a difference between groups that is this large if there actually is no difference in the populations? Is recognizability of the words a major determinant in how many words people can memorize? Can these data address questions about a cause-and-effect relationship between recognizability and memory? What people and what situations do these findings generalize to? Lots of other things! Generate a question, find a tool (not v.v.)

8 What (and why) might we want to know? What s the best guess for # of items memorized, given that a person (past or future) is in a specific group? group sample means (11.15 words and words) What our best guess for the difference between (population) group means? difference in group sample means (3.17 words) 30 clean # memorized

9 What (and why) might we want to know? Is recognizability of the words a major determinant in how many words people can memorize? r 2 is roughly (-1.48) 2 ((-1.48) ) =.046 in our sample, it explains ~ 5% of the variability in number of words memorized 30 clean # memorized

10 What (and why) might we want to know? Would it be likely to observe a difference between groups that is this large if there actually is no difference in the populations? t = -1.48, p =.15 reasonably likely to observe this difference in sample means if there is no difference in population means don t have enough evidence to infer that population means are different 30 clean # memorized

11 What (and why) might we want to know? What are our the lowest and highest plausible guesses for the difference between (population) group means? confidence interval (-7.47 words, 1.14 words) there might be a small advantage for not recognizable strings or there might be a really big advantage for recognizable strings probably worth studying some more 30 clean # memorized

12 What (and why) might we want to know? Can these data address questions about a cause-and-effect relationship between recognizability and memory? What people and what situations do these findings generalize to? experiment with random assignment can make cause-andeffect conclusions might want to generalize to people similar to college students, situations similar to memorizing letter strings and recalling immediately after

13 This time Recap: what types of questions do we ask when we compare two means? How much variance should we expect between group means? How can we compare multiple means?

14 A note on terminology Analysis of variance (ANOVA) is a broad term that describes a general procedure This procedure can be used to analyze many situations, with multiple variables, termed factors Today, I m referring to a one-factor, independentsamples analysis of variance as simply ANOVA

15 Reminder x x pool x x 1 x pool x 2 x 1 x x 2 total variance variance we can explain variance we cannot explain (1) Sum of squares (SS): for each value, calculate its distance from the mean and square it, then sum these squared values (2) Variance (s 2 or σ 2 ): mean of these squared values (3) Standard deviation (s or σ): square root of the variance

16 Ratio of variances When comparing two means, a t-statistic is a ratio numerator: the difference in sample means that we observed denominator: a typical distance in sample means that we would expect to observe, if the population means were equal A difference can only compare two means how can we summarize distances between many values? variance: a typical squared distance from the mean We ll look at a new ratio numerator: the variance in sample means that we observed denominator: the variance in sample means that we would expect to observe, if the population means (and variances) were equal (after some rearranging)

17 we re imagining that our observed data are generated by the following process: we have a single distribution of values we select n1 values as group 1, n2 values as group 2, etc. but group is arbitrary because all of the values came from the same distribution any differences in means between groups are from sampling error this is a distribution of means problem

18 How can we calculate this ratio? central limit theorem describes a distribution of sample means: has a standard deviation equal to the population standard deviation divided by the square root of the sample size has a variance equal to the population variance divided by the sample size (n) (variance of sample means we observed) divided by (variance of sample means we would expect to observe if the population means and variances were equal) (variance of sample means we observed) divided by (variance of individual values / sample size (n) corresponding to each sample mean) (n * variance of sample means we observed) divided by (variance of individual values)

19 How can we calculate this ratio? numerator: n * variance of sample means we observed n Pgroup ( x group x pooled ) 2 df betweengroup # groups - 1 Pgroup n group( x group x pooled ) 2 df betweengroup Pvalues ( x group x pooled ) 2 SSbetween (aka SSexplained) df betweengroup denominator: variance of individual values we observed Pvalues (x i x group ) 2 df withingroup SSwithin (aka SSunexplained) # values - # groups

20 F-ratio F = SS between df between SS within df within clean total 10 Ratio of independent estimates of the population variance (if the null hypothesis is true) Memorized between within Using SSbetween 10 Using SSwithin 0 Observation

21 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations 30 clean total Source SS sum of squares Between Ʃ(x group - x pooled) 2 k - 1 Within Ʃ(xi - x group) 2 N - k Total Ʃ(xi - x pooled) 2 (= SSb + SSw) 20 df MS degrees of mean square freedom 10 (variance) Memorized N - 1 (= dfb + dfw) 0 between SSb / dfb MSw = SSw / dfw MSb = Observation F MSb / MSw within

22 SS with dplyr

23 SS with dplyr

24 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between Ʃ(x group - x pooled) 2 k - 1 Within Ʃ(xi - x group) 2 N - k MSb = SSb / dfb MSw = SSw / dfw MSb / MSw Total Ʃ(xi - x pooled) 2 (= SSb + SSw) N - 1 (= dfb + dfw)

25 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between k - 1 Within N - k MSb = SSb / dfb MSw = SSw / dfw MSb / MSw Total N - 1 (= dfb + dfw)

26 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between = 1 Within = 49 MSb = SSb / dfb MSw = SSw / dfw MSb / MSw Total = 50

27 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between = 1 Within = / 1 = / 49 = MSb / MSw Total = 50

28 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Is 2.22 unlikely if the null hypothesis is true? Source SS sum of squares df degrees of freedom MS mean square (variance) F Between = 1 Within = / 1 = / 49 = / = 2.22 Total = 50

29 F distributions F is a ratio: Observed variance in sample means divided by expected variance in sample means Two independent estimates of the population variance If there are no differences between groups (in the population): What would we expect F to be? ratio of two things that we believe are equal 1 Will it always be exactly that? no, due to sampling error How high could F get? When should the numerator be greater than the denominator? no limit; the numerator should be greater than the denominator if the variability in our group means is the combination of random variability and different population means How low could F get? When should the numerator be less than the denominator? 0; random chance, but it is not consistent with any situation of interest only interested in upper (right) tail

30 F distributions What could we do if we didn t know the shape of the sampling distribution of F statistics we could have observed if there were no differences between groups (null hypothesis was true)? Already defined: a family of F distributions, defined by dfbetween (df1) and dfwithin (df2) this order matters Only interested in upper tail, which includes any ordering of differences between means

31 Statistical significance of F Where k = # of groups, N = total number of participants all sums are sums over all individual observations Is 2.22 unlikely if the null hypothesis is true? no Source SS sum of squares p > α df degrees of freedom Between = 1 Within = 49 MS mean square (variance) / 1 = / 49 = F / = 2.22 Total = 50 Fobserved < Fcritical

32 Relationship with t-test What if the t-test and ANOVA give us different results? (Shouldn t they give us the same result)? We assumed the population variances were equal (in addition to the mean), so the right comparison is to a Student s t-test that assumes equal variances between groups When we have k = 2, F = t 2 (The F-statistic that we get from our sample is exactly equal to the square of the Student s t-statistic that we get from our sample).

33 Relationship with t-test What if the t-test and ANOVA give us different results? (Shouldn t they give us the same result)? We assumed the population variances were equal (in addition to the mean), so the right comparison is to a Student s t-test that assumes equal variances between groups When we have k = 2, F = t 2 (The F-statistic that we get from our sample is exactly equal to the square of the Student s t-statistic that we get from our sample).

34 Relationship between F and t If you re curious The numerator of t 2 is (x 1 - x 2) 2 = ((x 1 - x 2) - (μ1 - μ2)) 2 / 1 This is an unbiased estimate of the variance of sample mean differences if H0 is true (proof omitted) The denominator of t 2 is sx 1-x 2 2 This is the expected variance of sample mean differences if H0 is true (estimated from individual observations) So t 2 is a ratio of variances, i.e., an F-statistic

35 Effect size in ANOVA r 2 = variance we can explain / total variance r 2 = SSbetween / SStotal r 2 = / =.043 Source SS sum of squares r 2 = t 2 / (t 2 + df) r 2 = (-1.49) 2 / ((-1.49) 2 + df) =.043 Between Within Total

36 Mini recap F is a ratio of variances that we would expect to see if all population groups had the same mean and variance If this ratio is large, it suggests that the group means are different from each other for reasons other than random variability If we have two groups, this maps on to the (two-tailed) Student s t-statistic (p-value is equal, F = t 2 ) Then why did we do this? procedure is not restricted to two groups partitioning variance is a useful strategy for other questions that cannot be answered with a t-test

37 This time Recap: what types of questions do we ask when we compare two means? How much variance should we expect between group means? How can we compare multiple means?

38 Comprehending passages Bransford & Johnson, If the balloons popped, the sound wouldn t be able to carry since everything would be too far away from the correct floor. 2. A closed window would also prevent the sound from carrying, since most buildings tend to be well-insulated. 3. Since the whole operation depends on a steady flow of electricity, a break in the middle of the wire would also cause problems. 4. Of course, the fellow could shout, but the human voice is not loud enough to carry that far. 5. An additional problem is that a string could break on the instrument. 6. Then there could be no accompaniment to the message. 7. It is clear that the best situation would involve less distance. 8. Then there would be fewer potential problems. 9. With face to face contact, the least number of things could go wrong.

39 Comprehending passages Bransford & Johnson, If the balloons popped, the sound wouldn t be able to carry since everything would be too far away from the correct floor. 2. A closed window would also prevent the sound from carrying, since most buildings tend to be well-insulated. 3. Since the whole operation depends on a steady flow of electricity, a break in the middle of the wire would also cause problems. 4. Of course, the fellow could shout, but the human voice is not loud enough to carry that far. 5. An additional problem is that a string could break on the instrument. 6. Then there could be no accompaniment to the message. 7. It is clear that the best situation would involve less distance. 8. Then there would be fewer potential problems. 9. With face to face contact, the least number of things could go wrong.

40 Comprehending passages Grouping variable: 57 participants randomly assigned to: hear passage alone (no picture) see picture before hearing passage see picture after hearing passage Response variable: test of comprehension of the passage (ranges from 1-7) Two hypotheses, use family-wise α =.05 H 0 : μ none = μ before = μ after H A : it is not the case that μ none = μ before = μ after (careful, this is not the same thing as as μ none μ before μ after )

41 F (statistical significance), ratio of (SSbetween / dfbetween) to (SSwithin / dfwithin) r 2 (effect size), ratio of SSbetween / SStotal what is the role of sample size (n) in each of these? between within clean total Observation Comprehension Comprehending passages

42 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between Ʃ(x group - x pooled) 2 k - 1 Within Ʃ(xi - x group) 2 N - k MSb = SSb / dfb MSw = SSw / dfw MSb / MSw Total Ʃ(xi - x pooled) 2 (= SSb + SSw) N - 1 (= dfb + dfw)

43 Sums of squares

44 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between k - 1 Within N - k MSb = SSb / dfb MSw = SSw / dfw MSb / MSw Total N - 1 (= dfb + dfw)

45 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between = 2 Within = 54 MSb = SSb / dfb MSw = SSw / dfw MSb / MSw Total = 56

46 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between = 2 Within = / 2 = / 54 = 1.75 MSb / MSw Total = 56

47 ANOVA table Where k = # of groups, N = total number of participants all sums are sums over all individual observations Source SS sum of squares df degrees of freedom MS mean square (variance) F Between = 2 Within = / 2 = / 54 = / 1.75 = Total = 56

48 ANOVA table in R

49 Effect size r 2 = SS between / SS total = / =.27 We can explain 27% of the variance between comprehension scores by taking into account which condition a person was in Source SS sum of squares df degrees of freedom MS mean square (variance) F Between = 2 Within = / 2 = / 54 = / 1.75 = Total = 56

50 Comprehending passages Two hypotheses, use family-wise α =.05 H 0 : μ none = μ before = μ after infer that it is not the case that μ none = μ before = μ after (careful, this is not the same thing as as μ none μ before μ after ) We have inferred that some population means are different, and we have described that taking condition into account explains 27% of the variance across scores But we want to know which groups are different Follow up pair-wise confidence intervals for differences in population means: 95% CI for difference between after - before: [-2.63, -0.85] 95% CI for difference between after - none: [-1.03, +0.72] 95% CI for difference between before - none: [+0.73, +2.42] which pairwise comparison are significant at α =.05?

51 Recap We can make inferences based on analysis of ratios of variance we compare two estimates of variance that we would expect to be equal if there were no differences between groups in the population We can use analysis of variance to compare means between multiple groups (control family-wise α)

52 Quiz 3 Some comments 30 Mean = 0.72 count Median = 0.74 SD = 0.16 IQR = (truncated) score

53 Questions

Two-Sample Inferential Statistics

Two-Sample Inferential Statistics The t Test for Two Independent Samples 1 Two-Sample Inferential Statistics In an experiment there are two or more conditions One condition is often called the control condition in which the treatment is

More information

Hypothesis testing: Steps

Hypothesis testing: Steps Review for Exam 2 Hypothesis testing: Steps Exam 2 Review 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region 3. Compute

More information

Hypothesis testing: Steps

Hypothesis testing: Steps Review for Exam 2 Hypothesis testing: Steps Repeated-Measures ANOVA 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region

More information

Sampling Distributions: Central Limit Theorem

Sampling Distributions: Central Limit Theorem Review for Exam 2 Sampling Distributions: Central Limit Theorem Conceptually, we can break up the theorem into three parts: 1. The mean (µ M ) of a population of sample means (M) is equal to the mean (µ)

More information

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique

More information

Disadvantages of using many pooled t procedures. The sampling distribution of the sample means. The variability between the sample means

Disadvantages of using many pooled t procedures. The sampling distribution of the sample means. The variability between the sample means Stat 529 (Winter 2011) Analysis of Variance (ANOVA) Reading: Sections 5.1 5.3. Introduction and notation Birthweight example Disadvantages of using many pooled t procedures The analysis of variance procedure

More information

Introduction to Business Statistics QM 220 Chapter 12

Introduction to Business Statistics QM 220 Chapter 12 Department of Quantitative Methods & Information Systems Introduction to Business Statistics QM 220 Chapter 12 Dr. Mohammad Zainal 12.1 The F distribution We already covered this topic in Ch. 10 QM-220,

More information

Calculating Fobt for all possible combinations of variances for each sample Calculating the probability of (F) for each different value of Fobt

Calculating Fobt for all possible combinations of variances for each sample Calculating the probability of (F) for each different value of Fobt PSY 305 Module 5-A AVP Transcript During the past two modules, you have been introduced to inferential statistics. We have spent time on z-tests and the three types of t-tests. We are now ready to move

More information

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

Review. One-way ANOVA, I. What s coming up. Multiple comparisons Review One-way ANOVA, I 9.07 /15/00 Earlier in this class, we talked about twosample z- and t-tests for the difference between two conditions of an independent variable Does a trial drug work better than

More information

10/31/2012. One-Way ANOVA F-test

10/31/2012. One-Way ANOVA F-test PSY 511: Advanced Statistics for Psychological and Behavioral Research 1 1. Situation/hypotheses 2. Test statistic 3.Distribution 4. Assumptions One-Way ANOVA F-test One factor J>2 independent samples

More information

Chapter 24. Comparing Means. Copyright 2010 Pearson Education, Inc.

Chapter 24. Comparing Means. Copyright 2010 Pearson Education, Inc. Chapter 24 Comparing Means Copyright 2010 Pearson Education, Inc. Plot the Data The natural display for comparing two groups is boxplots of the data for the two groups, placed side-by-side. For example:

More information

4.1. Introduction: Comparing Means

4.1. Introduction: Comparing Means 4. Analysis of Variance (ANOVA) 4.1. Introduction: Comparing Means Consider the problem of testing H 0 : µ 1 = µ 2 against H 1 : µ 1 µ 2 in two independent samples of two different populations of possibly

More information

The One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)

The One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs) The One-Way Repeated-Measures ANOVA (For Within-Subjects Designs) Logic of the Repeated-Measures ANOVA The repeated-measures ANOVA extends the analysis of variance to research situations using repeated-measures

More information

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares K&W introduce the notion of a simple experiment with two conditions. Note that the raw data (p. 16)

More information

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies The t-test: So Far: Sampling distribution benefit is that even if the original population is not normal, a sampling distribution based on this population will be normal (for sample size > 30). Benefit

More information

CHAPTER 13: F PROBABILITY DISTRIBUTION

CHAPTER 13: F PROBABILITY DISTRIBUTION CHAPTER 13: F PROBABILITY DISTRIBUTION continuous probability distribution skewed to the right variable values on horizontal axis are 0 area under the curve represents probability horizontal asymptote

More information

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z). Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z). For example P(X 1.04) =.8508. For z < 0 subtract the value from

More information

We need to define some concepts that are used in experiments.

We need to define some concepts that are used in experiments. Chapter 0 Analysis of Variance (a.k.a. Designing and Analysing Experiments) Section 0. Introduction In Chapter we mentioned some different ways in which we could get data: Surveys, Observational Studies,

More information

Statistical methods for comparing multiple groups. Lecture 7: ANOVA. ANOVA: Definition. ANOVA: Concepts

Statistical methods for comparing multiple groups. Lecture 7: ANOVA. ANOVA: Definition. ANOVA: Concepts Statistical methods for comparing multiple groups Lecture 7: ANOVA Sandy Eckel seckel@jhsph.edu 30 April 2008 Continuous data: comparing multiple means Analysis of variance Binary data: comparing multiple

More information

The t-statistic. Student s t Test

The t-statistic. Student s t Test The t-statistic 1 Student s t Test When the population standard deviation is not known, you cannot use a z score hypothesis test Use Student s t test instead Student s t, or t test is, conceptually, very

More information

PSY 216. Assignment 12 Answers. Explain why the F-ratio is expected to be near 1.00 when the null hypothesis is true.

PSY 216. Assignment 12 Answers. Explain why the F-ratio is expected to be near 1.00 when the null hypothesis is true. PSY 21 Assignment 12 Answers 1. Problem 1 from the text Explain why the F-ratio is expected to be near 1.00 when the null hypothesis is true. When H0 is true, the treatment had no systematic effect. In

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

Hypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -

Hypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos - Hypothesis Testing with One-Way ANOVA Statistics Arlo Clark-Foos Conceptual Refresher 1. Standardized z distribution of scores and of means can be represented as percentile rankings. 2. t distribution

More information

Introduction to the Analysis of Variance (ANOVA)

Introduction to the Analysis of Variance (ANOVA) Introduction to the Analysis of Variance (ANOVA) The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique for testing for differences between the means of multiple (more

More information

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table Your schedule of coming weeks One-way ANOVA, II 9.07 //00 Today: One-way ANOVA, part II Next week: Two-way ANOVA, parts I and II. One-way ANOVA HW due Thursday Week of May Teacher out of town all week

More information

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent: Activity #10: AxS ANOVA (Repeated subjects design) Resources: optimism.sav So far in MATH 300 and 301, we have studied the following hypothesis testing procedures: 1) Binomial test, sign-test, Fisher s

More information

Variance Estimates and the F Ratio. ERSH 8310 Lecture 3 September 2, 2009

Variance Estimates and the F Ratio. ERSH 8310 Lecture 3 September 2, 2009 Variance Estimates and the F Ratio ERSH 8310 Lecture 3 September 2, 2009 Today s Class Completing the analysis (the ANOVA table) Evaluating the F ratio Errors in hypothesis testing A complete numerical

More information

WELCOME! Lecture 13 Thommy Perlinger

WELCOME! Lecture 13 Thommy Perlinger Quantitative Methods II WELCOME! Lecture 13 Thommy Perlinger Parametrical tests (tests for the mean) Nature and number of variables One-way vs. two-way ANOVA One-way ANOVA Y X 1 1 One dependent variable

More information

An inferential procedure to use sample data to understand a population Procedures

An inferential procedure to use sample data to understand a population Procedures Hypothesis Test An inferential procedure to use sample data to understand a population Procedures Hypotheses, the alpha value, the critical region (z-scores), statistics, conclusion Two types of errors

More information

ONE FACTOR COMPLETELY RANDOMIZED ANOVA

ONE FACTOR COMPLETELY RANDOMIZED ANOVA MALLOY PSYCH 3000 1-ANOVA PAGE 1 ONE FACTOR COMPLETELY RANDOMIZED ANOVA Sampling Distribution of F F is a test statistic [ ][ ][ ][ ] Test Statistic: F = MALLOY PSYCH 3000 1-ANOVA PAGE 2 ONE WAY ANOVA

More information

Review for Final. Chapter 1 Type of studies: anecdotal, observational, experimental Random sampling

Review for Final. Chapter 1 Type of studies: anecdotal, observational, experimental Random sampling Review for Final For a detailed review of Chapters 1 7, please see the review sheets for exam 1 and. The following only briefly covers these sections. The final exam could contain problems that are included

More information

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College 1-Way ANOVA MATH 143 Department of Mathematics and Statistics Calvin College Spring 2010 The basic ANOVA situation Two variables: 1 Categorical, 1 Quantitative Main Question: Do the (means of) the quantitative

More information

Lecture (chapter 10): Hypothesis testing III: The analysis of variance

Lecture (chapter 10): Hypothesis testing III: The analysis of variance Lecture (chapter 10): Hypothesis testing III: The analysis of variance Ernesto F. L. Amaral March 19 21, 2018 Advanced Methods of Social Research (SOCI 420) Source: Healey, Joseph F. 2015. Statistics:

More information

Harvard University. Rigorous Research in Engineering Education

Harvard University. Rigorous Research in Engineering Education Statistical Inference Kari Lock Harvard University Department of Statistics Rigorous Research in Engineering Education 12/3/09 Statistical Inference You have a sample and want to use the data collected

More information

ANOVA: Comparing More Than Two Means

ANOVA: Comparing More Than Two Means 1 ANOVA: Comparing More Than Two Means 10.1 ANOVA: The Completely Randomized Design Elements of a Designed Experiment Before we begin any calculations, we need to discuss some terminology. To make this

More information

While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1

While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1 While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1 Chapter 12 Analysis of Variance McGraw-Hill, Bluman, 7th ed., Chapter 12 2

More information

Chapter 22. Comparing Two Proportions 1 /29

Chapter 22. Comparing Two Proportions 1 /29 Chapter 22 Comparing Two Proportions 1 /29 Homework p519 2, 4, 12, 13, 15, 17, 18, 19, 24 2 /29 Objective Students test null and alternate hypothesis about two population proportions. 3 /29 Comparing Two

More information

Independent Samples ANOVA

Independent Samples ANOVA Independent Samples ANOVA In this example students were randomly assigned to one of three mnemonics (techniques for improving memory) rehearsal (the control group; simply repeat the words), visual imagery

More information

Chapter 5 Contrasts for one-way ANOVA

Chapter 5 Contrasts for one-way ANOVA Chapter Contrasts for one-way NOV Page. What is a contrast? -. Types of contrasts -. Significance tests of a single contrast -0. Brand name contrasts -. Relationships between the omnibus F and contrasts

More information

Chapter 24. Comparing Means

Chapter 24. Comparing Means Chapter 4 Comparing Means!1 /34 Homework p579, 5, 7, 8, 10, 11, 17, 31, 3! /34 !3 /34 Objective Students test null and alternate hypothesis about two!4 /34 Plot the Data The intuitive display for comparing

More information

8/23/2018. One-Way ANOVA F-test. 1. Situation/hypotheses. 2. Test statistic. 3.Distribution. 4. Assumptions

8/23/2018. One-Way ANOVA F-test. 1. Situation/hypotheses. 2. Test statistic. 3.Distribution. 4. Assumptions PSY 5101: Advanced Statistics for Psychological and Behavioral Research 1 1. Situation/hypotheses 2. Test statistic One-Way ANOVA F-test One factor J>2 independent samples H o :µ 1 µ 2 µ J F 3.Distribution

More information

Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance ECON 509. Dr.

Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance ECON 509. Dr. Department of Economics Business Statistics Chapter 1 Chi-square test of independence & Analysis of Variance ECON 509 Dr. Mohammad Zainal Chapter Goals After completing this chapter, you should be able

More information

ANOVA Analysis of Variance

ANOVA Analysis of Variance ANOVA Analysis of Variance ANOVA Analysis of Variance Extends independent samples t test ANOVA Analysis of Variance Extends independent samples t test Compares the means of groups of independent observations

More information

Lecture 11: Two Way Analysis of Variance

Lecture 11: Two Way Analysis of Variance Lecture 11: Two Way Analysis of Variance Review: Hypothesis Testing o ANOVA/F ratio: comparing variances o F = s variance between treatment effect + chance s variance within sampling error (chance effects)

More information

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES BIOL 458 - Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES PART 1: INTRODUCTION TO ANOVA Purpose of ANOVA Analysis of Variance (ANOVA) is an extremely useful statistical method

More information

9 One-Way Analysis of Variance

9 One-Way Analysis of Variance 9 One-Way Analysis of Variance SW Chapter 11 - all sections except 6. The one-way analysis of variance (ANOVA) is a generalization of the two sample t test to k 2 groups. Assume that the populations of

More information

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z). Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z). For example P(X 1.04) =.8508. For z < 0 subtract the value from

More information

STAT Chapter 9: Two-Sample Problems. Paired Differences (Section 9.3)

STAT Chapter 9: Two-Sample Problems. Paired Differences (Section 9.3) STAT 515 -- Chapter 9: Two-Sample Problems Paired Differences (Section 9.3) Examples of Paired Differences studies: Similar subjects are paired off and one of two treatments is given to each subject in

More information

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs) The One-Way Independent-Samples ANOVA (For Between-Subjects Designs) Computations for the ANOVA In computing the terms required for the F-statistic, we won t explicitly compute any sample variances or

More information

1 Descriptive statistics. 2 Scores and probability distributions. 3 Hypothesis testing and one-sample t-test. 4 More on t-tests

1 Descriptive statistics. 2 Scores and probability distributions. 3 Hypothesis testing and one-sample t-test. 4 More on t-tests Overall Overview INFOWO Statistics lecture S3: Hypothesis testing Peter de Waal Department of Information and Computing Sciences Faculty of Science, Universiteit Utrecht 1 Descriptive statistics 2 Scores

More information

Inference for Regression Simple Linear Regression

Inference for Regression Simple Linear Regression Inference for Regression Simple Linear Regression IPS Chapter 10.1 2009 W.H. Freeman and Company Objectives (IPS Chapter 10.1) Simple linear regression p Statistical model for linear regression p Estimating

More information

Note that we are looking at the true mean, μ, not y. The problem for us is that we need to find the endpoints of our interval (a, b).

Note that we are looking at the true mean, μ, not y. The problem for us is that we need to find the endpoints of our interval (a, b). Confidence Intervals 1) What are confidence intervals? Simply, an interval for which we have a certain confidence. For example, we are 90% certain that an interval contains the true value of something

More information

Chapter 8 Student Lecture Notes 8-1. Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance

Chapter 8 Student Lecture Notes 8-1. Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance Chapter 8 Student Lecture Notes 8-1 Department of Economics Business Statistics Chapter 1 Chi-square test of independence & Analysis of Variance ECON 509 Dr. Mohammad Zainal Chapter Goals After completing

More information

ANOVA: Analysis of Variation

ANOVA: Analysis of Variation ANOVA: Analysis of Variation The basic ANOVA situation Two variables: 1 Categorical, 1 Quantitative Main Question: Do the (means of) the quantitative variables depend on which group (given by categorical

More information

psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests

psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests last lecture: introduction to factorial designs next lecture: factorial between-ps ANOVA II: (effect sizes and follow-up tests) 1 general

More information

z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests

z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests Chapters 3.5.1 3.5.2, 3.3.2 Prof. Tesler Math 283 Fall 2018 Prof. Tesler z and t tests for mean Math

More information

Using SPSS for One Way Analysis of Variance

Using SPSS for One Way Analysis of Variance Using SPSS for One Way Analysis of Variance This tutorial will show you how to use SPSS version 12 to perform a one-way, between- subjects analysis of variance and related post-hoc tests. This tutorial

More information

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions Introduction to Analysis of Variance 1 Experiments with More than 2 Conditions Often the research that psychologists perform has more conditions than just the control and experimental conditions You might

More information

Design of Experiments. Factorial experiments require a lot of resources

Design of Experiments. Factorial experiments require a lot of resources Design of Experiments Factorial experiments require a lot of resources Sometimes real-world practical considerations require us to design experiments in specialized ways. The design of an experiment is

More information

Advanced Experimental Design

Advanced Experimental Design Advanced Experimental Design Topic Four Hypothesis testing (z and t tests) & Power Agenda Hypothesis testing Sampling distributions/central limit theorem z test (σ known) One sample z & Confidence intervals

More information

Difference between means - t-test /25

Difference between means - t-test /25 Difference between means - t-test 1 Discussion Question p492 Ex 9-4 p492 1-3, 6-8, 12 Assume all variances are not equal. Ignore the test for variance. 2 Students will perform hypothesis tests for two

More information

Confidence Intervals. - simply, an interval for which we have a certain confidence.

Confidence Intervals. - simply, an interval for which we have a certain confidence. Confidence Intervals I. What are confidence intervals? - simply, an interval for which we have a certain confidence. - for example, we are 90% certain that an interval contains the true value of something

More information

Statistics For Economics & Business

Statistics For Economics & Business Statistics For Economics & Business Analysis of Variance In this chapter, you learn: Learning Objectives The basic concepts of experimental design How to use one-way analysis of variance to test for differences

More information

Analysis of variance (ANOVA) Comparing the means of more than two groups

Analysis of variance (ANOVA) Comparing the means of more than two groups Analysis of variance (ANOVA) Comparing the means of more than two groups Example: Cost of mating in male fruit flies Drosophila Treatments: place males with and without unmated (virgin) females Five treatments

More information

Data Analysis and Statistical Methods Statistics 651

Data Analysis and Statistical Methods Statistics 651 Data Analysis and Statistical Methods Statistics 651 http://www.stat.tamu.edu/~suhasini/teaching.html Suhasini Subba Rao Motivations for the ANOVA We defined the F-distribution, this is mainly used in

More information

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015 AMS7: WEEK 7. CLASS 1 More on Hypothesis Testing Monday May 11th, 2015 Testing a Claim about a Standard Deviation or a Variance We want to test claims about or 2 Example: Newborn babies from mothers taking

More information

Explanation of R 2, and Other Stories

Explanation of R 2, and Other Stories Explanation of R 2, and Other Stories I accidentally gave an incorrect formula for R 2 in class. This summary was initially just going to be me correcting my error, but I've taken the opportunity to clarify

More information

One-way between-subjects ANOVA. Comparing three or more independent means

One-way between-subjects ANOVA. Comparing three or more independent means One-way between-subjects ANOVA Comparing three or more independent means Data files SpiderBG.sav Attractiveness.sav Homework: sourcesofself-esteem.sav ANOVA: A Framework Understand the basic principles

More information

Statistics Primer. ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong

Statistics Primer. ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong Statistics Primer ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong 1 Quick Overview of Statistics 2 Descriptive vs. Inferential Statistics Descriptive Statistics: summarize and describe data

More information

CHAPTER 13: F PROBABILITY DISTRIBUTION

CHAPTER 13: F PROBABILITY DISTRIBUTION CHAPTER 13: F PROBABILITY DISTRIBUTION continuous probability distribution skewed to the right variable values on horizontal axis are 0 area under the curve represents probability horizontal asymptote

More information

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS In our work on hypothesis testing, we used the value of a sample statistic to challenge an accepted value of a population parameter. We focused only

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the

More information

One-Way ANOVA Cohen Chapter 12 EDUC/PSY 6600

One-Way ANOVA Cohen Chapter 12 EDUC/PSY 6600 One-Way ANOVA Cohen Chapter 1 EDUC/PSY 6600 1 It is easy to lie with statistics. It is hard to tell the truth without statistics. -Andrejs Dunkels Motivating examples Dr. Vito randomly assigns 30 individuals

More information

Cuckoo Birds. Analysis of Variance. Display of Cuckoo Bird Egg Lengths

Cuckoo Birds. Analysis of Variance. Display of Cuckoo Bird Egg Lengths Cuckoo Birds Analysis of Variance Bret Larget Departments of Botany and of Statistics University of Wisconsin Madison Statistics 371 29th November 2005 Cuckoo birds have a behavior in which they lay their

More information

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last week: Sample, population and sampling

More information

appstats27.notebook April 06, 2017

appstats27.notebook April 06, 2017 Chapter 27 Objective Students will conduct inference on regression and analyze data to write a conclusion. Inferences for Regression An Example: Body Fat and Waist Size pg 634 Our chapter example revolves

More information

Review of Statistics 101

Review of Statistics 101 Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods

More information

One-Way Analysis of Variance. With regression, we related two quantitative, typically continuous variables.

One-Way Analysis of Variance. With regression, we related two quantitative, typically continuous variables. One-Way Analysis of Variance With regression, we related two quantitative, typically continuous variables. Often we wish to relate a quantitative response variable with a qualitative (or simply discrete)

More information

Chapter 12. ANalysis Of VAriance. Lecture 1 Sections:

Chapter 12. ANalysis Of VAriance. Lecture 1 Sections: Chapter 1 ANalysis Of VAriance Lecture 1 Sections: 1.1 1. ANOVA test is an extension of two sample independent t-test demonstrated how to compare differences of means between two groups. The t-test is

More information

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression

More information

CHAPTER 10 Comparing Two Populations or Groups

CHAPTER 10 Comparing Two Populations or Groups CHAPTER 10 Comparing Two Populations or Groups 10.1 Comparing Two Proportions The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers Comparing Two Proportions

More information

Unit 27 One-Way Analysis of Variance

Unit 27 One-Way Analysis of Variance Unit 27 One-Way Analysis of Variance Objectives: To perform the hypothesis test in a one-way analysis of variance for comparing more than two population means Recall that a two sample t test is applied

More information

Introduction to Statistics for the Social Sciences Review for Exam 4 Homework Assignment 27

Introduction to Statistics for the Social Sciences Review for Exam 4 Homework Assignment 27 Introduction to Statistics for the Social Sciences Review for Exam 4 Homework Assignment 27 Name: Lab: The purpose of this worksheet is to review the material to be represented in Exam 4. Please answer

More information

Chapter 7 Factorial ANOVA: Two-way ANOVA

Chapter 7 Factorial ANOVA: Two-way ANOVA Chapter 7 Factorial ANOVA: Two-way ANOVA Page Two-way ANOVA: Equal n. Examples 7-. Terminology 7-6 3. Understanding main effects 7- and interactions 4. Structural model 7-5 5. Variance partitioning 7-6.

More information

Chapter 22. Comparing Two Proportions 1 /30

Chapter 22. Comparing Two Proportions 1 /30 Chapter 22 Comparing Two Proportions 1 /30 Homework p519 2, 4, 12, 13, 15, 17, 18, 19, 24 2 /30 3 /30 Objective Students test null and alternate hypothesis about two population proportions. 4 /30 Comparing

More information

Analysis of Variance

Analysis of Variance Statistical Techniques II EXST7015 Analysis of Variance 15a_ANOVA_Introduction 1 Design The simplest model for Analysis of Variance (ANOVA) is the CRD, the Completely Randomized Design This model is also

More information

Lab #12: Exam 3 Review Key

Lab #12: Exam 3 Review Key Psychological Statistics Practice Lab#1 Dr. M. Plonsky Page 1 of 7 Lab #1: Exam 3 Review Key 1) a. Probability - Refers to the likelihood that an event will occur. Ranges from 0 to 1. b. Sampling Distribution

More information

ANOVA Situation The F Statistic Multiple Comparisons. 1-Way ANOVA MATH 143. Department of Mathematics and Statistics Calvin College

ANOVA Situation The F Statistic Multiple Comparisons. 1-Way ANOVA MATH 143. Department of Mathematics and Statistics Calvin College 1-Way ANOVA MATH 143 Department of Mathematics and Statistics Calvin College An example ANOVA situation Example (Treating Blisters) Subjects: 25 patients with blisters Treatments: Treatment A, Treatment

More information

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last two weeks: Sample, population and sampling

More information

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 22s:152 Applied Linear Regression Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) We now consider an analysis with only categorical predictors (i.e. all predictors are

More information

Factorial Analysis of Variance

Factorial Analysis of Variance Factorial Analysis of Variance Conceptual Example A repeated-measures t-test is more likely to lead to rejection of the null hypothesis if a) *Subjects show considerable variability in their change scores.

More information

Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.

Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6. Chapter 7 Reading 7.1, 7.2 Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.112 Introduction In Chapter 5 and 6, we emphasized

More information

Analysis of Variance (ANOVA)

Analysis of Variance (ANOVA) Analysis of Variance (ANOVA) Two types of ANOVA tests: Independent measures and Repeated measures Comparing 2 means: X 1 = 20 t - test X 2 = 30 How can we Compare 3 means?: X 1 = 20 X 2 = 30 X 3 = 35 ANOVA

More information

STA Module 11 Inferences for Two Population Means

STA Module 11 Inferences for Two Population Means STA 2023 Module 11 Inferences for Two Population Means Learning Objectives Upon completing this module, you should be able to: 1. Perform inferences based on independent simple random samples to compare

More information

STA Rev. F Learning Objectives. Two Population Means. Module 11 Inferences for Two Population Means

STA Rev. F Learning Objectives. Two Population Means. Module 11 Inferences for Two Population Means STA 2023 Module 11 Inferences for Two Population Means Learning Objectives Upon completing this module, you should be able to: 1. Perform inferences based on independent simple random samples to compare

More information

Introduction to Estimation. Martina Litschmannová K210

Introduction to Estimation. Martina Litschmannová K210 Introduction to Estimation Martina Litschmannová martina.litschmannova@vsb.cz K210 Populations vs. Sample A population includes each element from the set of observations that can be made. A sample consists

More information

Confidence intervals

Confidence intervals Confidence intervals We now want to take what we ve learned about sampling distributions and standard errors and construct confidence intervals. What are confidence intervals? Simply an interval for which

More information

Two-sample inference: Continuous data

Two-sample inference: Continuous data Two-sample inference: Continuous data Patrick Breheny April 6 Patrick Breheny University of Iowa to Biostatistics (BIOS 4120) 1 / 36 Our next several lectures will deal with two-sample inference for continuous

More information

HYPOTHESIS TESTING. Hypothesis Testing

HYPOTHESIS TESTING. Hypothesis Testing MBA 605 Business Analytics Don Conant, PhD. HYPOTHESIS TESTING Hypothesis testing involves making inferences about the nature of the population on the basis of observations of a sample drawn from the population.

More information

Battery Life. Factory

Battery Life. Factory Statistics 354 (Fall 2018) Analysis of Variance: Comparing Several Means Remark. These notes are from an elementary statistics class and introduce the Analysis of Variance technique for comparing several

More information