Two Factor ANOVA. March 2, 2017
|
|
- Flora Cox
- 6 years ago
- Views:
Transcription
1 Two Factor ANOVA March, 07 Contents Two Factor ANOVA Example : peanut butter and jelly Within-cell variance - the denominator of all three F-tests Main effects for rows: Does peanut butter affect taste ratings? Main effect for columns: does jelly affect taste ratings? The interaction: does the effect of peanut butter on taste ratings depend on jelly? The Summary Table Example : peanut butter and ham Example : A x ANOVA on study habits Questions Answers Two Factor ANOVA A two factor ANOVA (sometimes called two-way ANOVA) is a hypothesis test on means for which groups are placed in groups that vary along two factors instead of one. For example, we might want to measure BMI for subjects with different diets AND for different levels of exercise. Or, we might measure the duration of hospital stays for patients undergoing different levels of medication AND different types of physical therapy. In these cases, we re interest in not only how the means vary across each factor (called the main effects), but also how these two dimensions interact with each other (called interactions). Here s how to get to the -factor ANOVA with the flow chart: START HERE Test for ρ = 0 Ch 7. number of correlations correlation (r) measurement scale frequency number of variables χ test frequency Ch 9.5 Test for ρ = ρ Ch 7.4 Means χ test independence Ch 9.9 z-test Ch. Yes Do you know σ? number of means More than number of factors -factor ANOVA Ch No one sample t-test Ch.4 -factor ANOVA Ch 0 independent measures t-test Ch 5.6 Yes independent samples? No dependent measures t-test Ch 6.4
2 For a two-factor ANOVA, each of the two factors (like diet and exercise ) are comprised of different levels, (like different diets, or different levels of exercise). The complete design involves having subjects placed into all possible combinations of levels for the two factors. For example, if there are two diets and three levels of exercise, there will be x = 6 groups, or cells. Each cell should have the same number of subjects, which called a balanced design. There are ways to deal with unequal sample sizes, but we will not go there. Example : peanut butter and jelly Let s start with a simple example. Suppose you want to study the effects of adding peanut butter and jelly to bread to study how they affect taste preferences. You take 6 students and break them down into 4 groups of 9 in what s called a x design. One group will get just bread, another will get bread with peanut butter, another will get bread with jelly, and one group will get bread with both peanut butter and jelly. We ll run our test with α = Each student tastes their food and rates their preference on a scale from (gross) to 0 (excellent). The raw data looks like this: no jelly, no no jelly, peanut jelly, no peanut jelly and peanut peanut butter butter butter butter Here are the summary statistics from the experiment: Means no jelly jelly no peanut butter peanut butter SSwc no jelly jelly no peanut butter peanut butter Totals grand mean 6. SS total We ve organized statistics in a x matrix for which the factor peanut butter varies across the rows and the factor jelly varies across the columnns. We therefore call peanut butter the row factor and jelly the column factor. The table titled SSwc contains the sums of squared deviation of each score from the mean of for the cell that it belongs to. For example, for the no peanut butter, no jelly category, the mean for that cell is 4., so the SSwc for that cell is: (8 4.) + (6 4.) (5 4.) =.5556 SS total is the sums of squared deviation of each score from the grand mean. Before we run any statistical analyses, it s useful to plot the means with error bars representing the standard errors of the mean. Remember, standard errors can be calculated by first calculating standard deviations from SSwc:
3 sx = SS wc n And then dividing by by the square root of the sample size to get the standard error of the mean: s x = s x n For example, for the no peanut butter, no jelly category, sx = So =.986 s x = = 0.66 Check for yourself that the standard errors of the mean turn out to be: Standard errors no jelly jelly no peanut butter peanut butter Here s the plot: no peanut butter peanut butter Taste rating no jelly jelly Visually inspecting the graph, you can see that the green points are shifted above the red ones. This means that adding peanut butter helps the taste for both the plain bread, and for bread with jelly. This is called a main effect for rows, where the rows correspond the factor peanut butter. Similarly, adding jelly increases the taste rating both for plain bread, and for bread with peanut butter. This is a main effect for columns where columns refer to the factor jelly. Notice also that the lines are roughly parallel. This means that the increase in the taste rating by adding peanut butter is about the same with and without jelly. We therefore say that there is no interaction between the factors of peanut butter and jelly.
4 The -factor ANOVA quantifies the statistical significance of these three observations: the main main effect for rows, the main effect for columns, and the interaction. Within-cell variance - the denominator of all three F-tests The statistical significance of the main effect for rows, the main effect for columns and the interaction will all be done with F-tests. Just as with the -factor ANOVA, the F-test will be a ratio of variances which are both estimates of the population variance under the null hypothesis. Also, like the -factor ANOVA, the denominator will be an estimate of variance based on the variability within each group, and again we calculate it by summing up the sums of squared error between each score and its own cell s mean: SSwc = (X X cell ) I ve given you SSwc for each cell, so the total SSwc is just the sum of these four numbers: SSwc = = 7. Just as with the -way ANOVA, the degrees of freedom for SSwc is the total number of scores minus the total number of groups, so df = = The variance within cells is SSwc divided by its degrees of freedom: s wc = SS wc dfwc = 7. =.97 Main effects for rows: Does peanut butter affect taste ratings? To quantify the main effect for rows (peanut butter) we average the means across the two levels for columns (jelly). We can show these two means in a third column in the matrix of means: Means no jelly jelly Row Means no peanut butter peanut butter The difference between these two row means indicates the effect of adding peanut butter after averaging across the two levels of jelly. We ll use these two means to calculate the numerator of the F-test for the main effect for rows. We first calculate SS for rows much like we would if this were a -factor ANOVA. We calculate the sum of squared difference of the row means from the grand mean, scaled by the sample size for the number of scores for each mean. Since there are groups for each mean, the sample size for each row mean is ()(9) = 8. So, SSrow = (nrow)( Xrow X) = SS R = ()(9)( ) + ()(9)( ) = 5 The degrees of freedom for SS for rows is equal to the number of rows minus (df = - = ). The variance for rows is SS for rows (SSrows) divided by its degrees of freedom: s rows = SS rows dfrows = 5 = 5 The F-statistic for the main effect for rows is the variance for rows divided by the variance within cells: Frows = s rows s wc = 5.97 = 6.9 Under the null hypothesis, the variance for rows is another estimate of the population variance. So if the null hypothesis is true, this F statistic should be around, on average. A large value of F indicates that the row means vary more than expected under the null hypothesis. If F is large enough, we conclude that this is too unusual for the null hypothesis to be true, so we reject it and conclude that the population means for the two rows are not the same. 4
5 The critical value for this F can be found in table E, using the degrees of freedom and. dfw df b The critical value is 4.5. You can use the F-calculator to find that the p-value is For our example, since our observed value of F (6.9) is greater than the critical value of F (4.5), we reject the null hypothesis. We can conclude using APA format: There is a significant main effect of peanut butter on taste ratings, F(,) = 6.9, p = Main effect for columns: does jelly affect taste ratings? Conducting the F-test on the main effect for columns (jelly) is completely analogous to the main effect for rows. We first calculate the column means by calculating the mean for each of the two colummns. We can show this as a third row in the matrix of means: Means no jelly jelly no peanut butter peanut butter Column Means 5. 7 The difference between these two column means indicates the effect of adding jelly after averaging across the two levels of peanut butter. As for rows, we ll calculate SS for columns, which is the sum of squared difference of the column means from the grand mean, scaled by the sample size for the number of scores for each mean: SS col = (n col )( X col X) = (8)(5. 6.) + (8)(7 6.) = The degrees of freedom for SS for columns is equal to the number of columns minus (df = - = ). The variance for the main effect for columns is SS for columns divided by its degrees of freedom: s col = SS col = df col = The F-statistic for the main effect for columns is the variance for columns divided by the variance within cells: F col = s col s wc = = 7.6 The F-calculator shows that the p-value for the main effect for columns (the df s are and ) is For our example, since our observed value of F (7.6) is greater than the critical value of F (4.5), we reject the null hypothesis. We can conclude using APA format: There is a significant main effect of jelly on taste ratings, F(,) = 7.6, p = 0.07 The interaction: does the effect of peanut butter on taste ratings depend on jelly? This third F-test determines if there is a statistically significant interaction between our two factors on our dependent variable. For our example, an interaction would occur if the effect of adding peanut butter to the bread added a different amount to the taste rating depending upon whether or not there was jelly (and vice versa). Remember from the graph that the two lines 5
6 were roughly parallel, indicating that peanut butter had a similar increase in taste ratings with and without jelly. Parallel lines indicate no interaction between the two variables. To run this F-test, we need to calculate a third estimate of the population variance. The formula for the sums of squares for the interaction between rows and columns (SS RxC ) is messy and not very intuitive. Fortunately we don t need it because we can infer it by using the fact that the sums of squared deviation from the grand mean can be broken down into three components: SS total = SSrows + SS cols + SS RxC + SSwc I ve given you SS total and we just calculated SSrows, SS cols and SSwc. So SS RxC = SS total SSrows SS cols SSwc = = The degrees of freedom for the row by column interaction is the number of rows minus times the number of columns minus : (-)(-) =. The variance for the row by column interaction is the SS divided by df: s RxC = SS RxC df RxC = = The F-statistic is: F RxC = s RxC s wc =.97 = 0.5 The F-calculator shows that the p-value for this value of F (the df s are and ) is Since our p-value (0.605) is greater than alpha (0.05), we fail to reject the null hypothesis. We can conclude using APA format: There is not a significant interaction between the effects of peanut butter and jelly on taste ratings, F(,) = 0.5, p = The Summary Table Like for -factor ANOVAs, the results from -factor ANOVAs are often reported in a summary table like this: SS df MS F F crit p-value Rows Columns R X C wc Total Example : peanut butter and ham Now suppose you want to study the effects of adding peanut butter and ham to bread. The (x) design is like the previous example; you take 6 students and break them down into 4 groups. Again, we ll run our test with α = Here are the summary statistics from the experiment: Means no ham ham no peanut butter peanut butter
7 SSwc no ham ham no peanut butter peanut butter Totals grand mean 5. SS total Here s a plot the means with error bars representing the standard errors of the mean. 6. no peanut butter peanut butter Taste rating no ham ham The first thing you see is that the lines are not parallel. Looking at the red points (no peanut butter), you can see that adding ham to bread improves its taste. But, if you start with peanut butter on your bread, adding ham makes it taste worse (yuck). This is a classic X shaped interaction, and we ll see below that it s statistically significant. For a main effect for rows we d need the overall rating to change when we add peanut butter. But adding peanut butter increases the rating for bread with no ham but decreases the rating with ham. Overall, the greeen points aren t higher or lower than the red points. It therefore looks like there is no main effect for peanut butter. For a main main effect for columns (ham) we d need the overall rating to change when we add ham. But you can see that the mean of the data point on the left (no ham are no higher or lower than the mean of the data points on the right (with jelly. So we don t expect a main effect for the column factor, ham. Now we ll work through the problem and fill in the summary table at the end: First, the within-cell variance: SSwc = (X X cell ) SSwc = = df = = s wc = SS wc dfwc = =
8 For the main effects we can create a new table with both the row and column means: Means no ham ham Row Means no peanut butter peanut butter Column Means Note that the two row means don t differ by much, nor do the two column means. This shows a lack of main effects for both rows and columns. We ll now compute the SS, variance and F-statistic for the main effect for rows. SS R = ()(9)( ) + ()(9)( ) = 0. With df R = = s rows = SS rows dfrows = 0. = 0. Frows = s rows s wc = = 0.4 The F-calculator shows that the p-value is Since For our example, since our observed value of F (0.4) is not greater than the critical value of F (4.5), we fail to reject the null hypothesis. We can conclude using APA format: There is not a significant main effect of peanut butter on taste ratings, F(,) = 0.4, p = Now for the main effect for columns: SS col = (n col )( X col X) = (8)( ) + (8)( ) = 0. df = - = s col = SS col = df 0. col = 0. F col = s col s wc = = 0.4 The F-calculator shows that the p-value for the main effect for columns (the df s are and ) is Because our p-value (0.675) is greater than alpha (0.05), we conclude There is not a significant main effect of ham on taste ratings, F(,) = 0.4, p = Finally, for the interaction: SS RxC = SS total SSrows SS cols SSwc = = 7. df = (-)(-) =. s RxC = SS RxC = df 7. RxC = 7. F RxC = s RxC s wc = = 5.5 The F-calculator shows that the p-value for this value of F (the df s are and ) is Since our p-value (0.0004) is less than alpha (0.05) We conclude: There is a significant interaction between the effects of 8
9 peanut butter and ham on test scores, F(,) = 5.5, p = The final summary table looks like this: SS df MS F F crit p-value Rows Columns R X C wc Total 5 Example : A x ANOVA on study habits This third example is an two-factor ANOVA on study habbits. You manipulate two factors, music and study location. for the factor music, you have students either study in silence, or listen to the music of their choice. For the factor study location, you have groups of students study in one of places: coffee shop, home and library. You find 5 students for each of the 4 groups. You generate the following summary statistics: Means coffee shop home library no music music SSwc coffee shop home library no music music Totals grand mean SS total Here s a plot the means with error bars representing the standard errors of the mean. 9
10 80 no music music Exam score coffee shop home library study location You ll probably first notice that the red points are higher than the green points. This is a main effecct for the row factor, music, for which listening to music leads to lower scores. Next you ll notice that the scores vary across the column factor, study location, with the best scores for students studying at home. Since there are three locations of study, we re looking for a difference across means - just like a -factor ANOVA with three levels. Here, we see that there does look like there is a main effect for study location. Now we ll work through the problem and fill in the summary table at the end: The within-cell variance is calculated as in the examples above by adding up SSwc across the 6 cells: SSwc = (X X cell ) SSwc = = df = n total k = 50 6 = 44 s wc = SS wc dfwc = = For the main effects, again we ll create a new table with both the row and column means: Means coffee shop home library column means no music music row means Next we ll compute the F-statistic for the main effect of rows (music) Note that since there are three columns, there are three means contributing to each row mean, the sample size for each row mean is (5)() = 75. SSrow = (nrow)( Xrow X) = (75)( ) + (75)( ) =
11 With df = - =. s rows = SS rows dfrows = 75.9 = 75.9 Frows = s rows s wc = = 4.7 The F-calculator shows that the critical value is.9 and the p-value is 0.0. Our observed value of F (4.7) is greater than the critical value of F (.9), we reject the null hypothesis. We can conclude using APA format: There is a significant main effect of music on test scores, F(,44) = 4.7, p = 0.0 Now for the main effect for columns (study location): SS col = (n col )( X col X) = (50)( ) + (50)( ) + (50)( ) = df = - = s col = SS col = df col = 4.68 F col = s col s wc = =.84 The F-calculator shows that the p-value for the main effect for columns (the df s are and 44) is Because our p-value (0.07) is less than alpha (0.05), we conclude There is a significant main effect of study location on taste ratings, F(,44) = 4.7, p = 0.0 Finally, for the interaction: SS RxC = SS total SSrows SS cols SSwc = = df = (-)(-) =. s RxC = SS RxC = df RxC = F RxC = s RxC s wc = = 0.5 The F-calculator shows that the p-value for this value of F (the df s are and 44) is Since our p-value (0.705) is greater than alpha (0.05) We conclude: There is not a significant interaction between the effects of music and study location on taste ratings, F(,44) = 0.5, p = The final summary table looks like this: SS df MS F F crit p-value Rows Columns R X C wc Total Questions Your turn again. Here are 0 random practice questions followed by their answers.
12 ) For a 499 project you measure the effect of kinds of musical groups and 4 kinds of grad students on the warmth of republicans. You find republicans for each group and measure warmth for each of the x4=8 groups. You generate the following table of means: Means grad students: grad students: grad students: grad students: 4 musical groups: musical groups: and the following table for SSwc: SSwc grad students: grad students: grad students: grad students: 4 musical groups: musical groups: The grand mean is and SS total is Make a plot of the data with error bars representing standard errors of the mean. Using an alpha value of α=0.05, test for a main effect of musical groups, a main effect of grad students and an interaction on the warmth of republicans. ) Your advsor asks you to measure the effect of kinds of psych 5 students and 4 kinds of facial expressions on the recognition of exams. You find 9 exams for each group and measure recognition for each of the x4=8 groups. You generate the following table of means: expres- facial sions: Means expres- facial sions: expres- facial sions: expres- facial sions: 4 psych 5 students: psych 5 students: and the following table for SSwc: expres- facial sions: SSwc expres- facial sions: expres- facial sions: expres- facial sions: 4 psych 5 students: psych 5 students: The grand mean is 0.79 and SS total is Make a plot of the data with error bars representing standard errors of the mean. Using an alpha value of α=0.05, test for a main effect of psych 5 students, a main effect of facial expressions and an interaction on the recognition of exams. ) I go and measure the effect of kinds of poo and 4 kinds of monkeys on the smell of potatoes. You find 0 potatoes for each group and measure smell for each of the x4=8 groups. You generate the following table of means: Means monkeys: monkeys: monkeys: monkeys: 4 poo: poo:
13 and the following table for SSwc: SSwc monkeys: monkeys: monkeys: monkeys: 4 poo: poo: The grand mean is and SS total is Using an alpha value of α=0.05, test for a main effect of poo, a main effect of monkeys and an interaction on the smell of potatoes. 4) Your advsor asks you to measure the effect of kinds of teenagers and 4 kinds of examples on the farts of statistics problems. You find 8 statistics problems for each group and measure farts for each of the x4=8 groups. You generate the following table of means: Means examples: examples: examples: examples: 4 teenagers: teenagers: and the following table for SSwc: SSwc examples: examples: examples: examples: 4 teenagers: teenagers: The grand mean is 6.97 and SS total is Using an alpha value of α=0.05, test for a main effect of teenagers, a main effect of examples and an interaction on the farts of statistics problems. 5) I go and measure the effect of kinds of fingers and kinds of photoreceptors on the happiness of personalities. You find personalities for each group and measure happiness for each of the x=6 groups. You generate the following table of means: Means photoreceptors: photoreceptors: photoreceptors: fingers: fingers: and the following table for SSwc: SSwc photoreceptors: photoreceptors: photoreceptors: fingers: fingers: The grand mean is and SS total is Make a plot of the data with error bars representing standard errors of the mean. Using an alpha value of α=0.0, test for a main effect of fingers, a main effect of photoreceptors and an interaction on the happiness of personalities. 6) Let s measure the effect of kinds of sponges and kinds of elbows on the education of Americans. You find 9
14 Americans for each group and measure education for each of the x=6 groups. You generate the following table of means: Means elbows: elbows: elbows: sponges: sponges: and the following table for SSwc: SSwc elbows: elbows: elbows: sponges: sponges: The grand mean is and SS total is Make a plot of the data with error bars representing standard errors of the mean. Using an alpha value of α=0.0, test for a main effect of sponges, a main effect of elbows and an interaction on the education of Americans. 7) You measure the effect of kinds of eggs and kinds of monkeys on the health of teams. You find teams for each group and measure health for each of the x=4 groups. You generate the following table of means: Means monkeys: monkeys: eggs: eggs: and the following table for SSwc: SSwc monkeys: monkeys: eggs: eggs: The grand mean is and SS total is Using an alpha value of α=0.0, test for a main effect of eggs, a main effect of monkeys and an interaction on the health of teams. 8) You measure the effect of kinds of video games and kinds of otter pops on the taste of fathers. You find 7 fathers for each group and measure taste for each of the x=4 groups. You generate the following table of means: Means otter pops: otter pops: video games: video games: and the following table for SSwc: 4
15 SSwc otter pops: otter pops: video games: video games: The grand mean is and SS total is.89. Using an alpha value of α=0.05, test for a main effect of video games, a main effect of otter pops and an interaction on the taste of fathers. 9) For some reason you measure the effect of kinds of spaghetti and 4 kinds of dinosaurs on the warmth of grandmothers. You find 7 grandmothers for each group and measure warmth for each of the x4=8 groups. You generate the following table of means: Means dinosaurs: dinosaurs: dinosaurs: dinosaurs: 4 spaghetti: spaghetti: and the following table for SSwc: SSwc dinosaurs: dinosaurs: dinosaurs: dinosaurs: 4 spaghetti: spaghetti: The grand mean is and SS total is Make a plot of the data with error bars representing standard errors of the mean. Using an alpha value of α=0.05, test for a main effect of spaghetti, a main effect of dinosaurs and an interaction on the warmth of grandmothers. 0) You want to measure the effect of kinds of brains and kinds of cartoon characters on the clothing of beers. You find 0 beers for each group and measure clothing for each of the x=6 groups. You generate the following table of means: charac- cartoon ters: Means charac- cartoon ters: charac- cartoon ters: brains: brains: and the following table for SSwc: charac- cartoon ters: SSwc charac- cartoon ters: charac- cartoon ters: brains: brains: The grand mean is 4.98 and SS total is Using an alpha value of α=0.05, test for a main effect of brains, a main effect of cartoon characters and an interaction on the clothing of beers. 5
16 Answers 6
17 ) Within cell: SSwc = = dfwc = 04 ()(4) = 96 s wc = = 4. Rows: musical groups mean row = mean row = = 5.7 = 6.0 SS R = (4)()( ) + (4)()( ) = 4. df R = = s R = 4. = 4. F R (, 96) = = 0.4 Columns: grad students mean col : mean col : mean col : mean col 4 : = = = 4.46 = SS C = ()()( ) +()()( ) +()()( ) +()()( ) = 7.9 df C = 4 = s C = 7.9 = F C (, 96) = = Interaction: SS RXC = ( ) = 558. df RXC = ( )(4 ) = s RXC = = F RXC (, 96) = =
18 SS df MS F F crit p-value Rows Columns R X C wc Total There is not a significant main effect for musical groups (rows) on the warmth of republicans, F(,96) = 0.40, p = There is not a significant main effect for grad students (columns), F(,96) = 0.570, p = There is a significant interaction between musical groups and grad students, F(,96) = 4.40, p = warmth of republicans musical groups musical groups 4 grad students 8
19 ) Within cell: SSwc = = dfwc = 5 ()(4) = 44 s wc = = Rows: psych 5 students mean row = mean row = = =.09 SS R = (4)(9)( ) + (4)(9)( ) = 5.59 df R = = s R = 5.59 = 5.59 F R (, 44) = = 5. Columns: facial expressions mean col : mean col : mean col : mean col 4 : =.95 = 0.5 = = SS C = ()(9)( ) + ()(9)( ) + ()(9)( ) + ()(9)( ) = 9.4 df C = 4 = s C = 9.4 =.08 F C (, 44) = =.744 Interaction: SS RXC = ( ) = 75.9 df RXC = ( )(4 ) = s RXC = = F RXC (, 44) = =
20 SS df MS F F crit p-value Rows Columns R X C wc Total There is a significant main effect for psych 5 students (rows) on the recognition of exams, F(,44) = 5.00, p = There is not a significant main effect for facial expressions (columns), F(,44) =.740, p = 0.6. There is not a significant interaction between psych 5 students and facial expressions, F(,44) = 0.780, p = recognition of exams psych 5 students psych 5 students 5 4 facial expressions 0
21 ) Within cell: SSwc = = dfwc = 60 ()(4) = 5 s wc = = Rows: poo mean row = mean row = = 86.4 = SS R = (4)(0)( ) + (4)(0)( ) = df R = = s R = = F R (, 5) = = 4. Columns: monkeys mean col : mean col : mean col : mean col 4 : = = = = SS C = ()(0)( ) +()(0)( ) +()(0)( ) +()(0)( ) = df C = 4 = s C = = F C (, 5) = = 4.78 Interaction: SS RXC = 46.4 ( ) = df RXC = ( )(4 ) = s RXC = = F RXC (, 5) = =.69
22 SS df MS F F crit p-value Rows Columns R X C wc Total There is a significant main effect for poo (rows) on the smell of potatoes, F(,5) = 4.00, p = There is a significant main effect for monkeys (columns), F(,5) = 4.70, p = There is not a significant interaction between poo and monkeys, F(,5) =.40, p = 0.5.
23 4) Within cell: SSwc = = dfwc = 64 ()(4) = 56 s wc = =.959 Rows: teenagers mean row = mean row = = = SS R = (4)(8)( ) + (4)(8)( ) = df R = = s R = = F R (, 56) = = 0.4 Columns: examples mean col : mean col : mean col : mean col 4 : = = = 6.86 = 6.06 SS C = ()(8)( ) + ()(8)( ) + ()(8)( ) + ()(8)( ) = df C = 4 = s C = = F C (, 56) = =.5 Interaction: SS RXC = ( ) = 5.78 df RXC = ( )(4 ) = s RXC = 5.78 = F RXC (, 56) = =.86
24 SS df MS F F crit p-value Rows Columns R X C wc Total There is not a significant main effect for teenagers (rows) on the farts of statistics problems, F(,56) = 0.400, p = There is a significant main effect for examples (columns), F(,56) =.50, p = There is not a significant interaction between teenagers and examples, F(,56) =.80, p =
25 5) Within cell: SSwc = = dfwc = ()() = 6 s wc = = 6.85 Rows: fingers mean row = mean row = = = SS R = ()()( ) + ()()( ) = df R = = s R = = F R (, 6) = =.85 Columns: photoreceptors mean col : mean col : mean col : = = = SS C = ()()( ) + ()()( ) + ()()( ) = 7.9 df C = = s C = = F C (, 6) = = Interaction: SS RXC = 64.9 ( ) = df RXC = ( )( ) = s RXC = = F RXC (, 6) = =
26 SS df MS F F crit p-value Rows < Columns R X C wc Total There is a significant main effect for fingers (rows) on the happiness of personalities, F(,6) =.850, p = There is a significant main effect for photoreceptors (columns), F(,6) = 9.750, p = There is a significant interaction between fingers and photoreceptors, F(,6) = 5.90, p = happiness of personalities fingers fingers 60 photoreceptors 6
27 6) Within cell: SSwc = = dfwc = 4 ()() = 08 s wc = = Rows: sponges mean row = mean row = = = SS R = ()(9)( ) + ()(9)( ) = df R = = s R = = F R (, 08) = =.0 Columns: elbows mean col : mean col : mean col : = = = SS C = ()(9)( ) + ()(9)( ) + ()(9)( ) = df C = = s C = = F C (, 08) = = 0.7 Interaction: SS RXC = 547. ( ) = 05.9 df RXC = ( )( ) = s RXC = = F RXC (, 08) = =
28 SS df MS F F crit p-value Rows Columns R X C wc Total There is not a significant main effect for sponges (rows) on the education of Americans, F(,08) =.00, p = There is not a significant main effect for elbows (columns), F(,08) = 0.0, p = There is a significant interaction between sponges and elbows, F(,08) = 5.590, p = education of Americans sponges sponges 70 elbows 8
29 7) Within cell: SSwc = = dfwc = 9 ()() = 88 s wc = = Rows: eggs mean row = mean row = = = 55.5 SS R = ()()( ) + ()()( ) = df R = = s R = = F R (, 88) = = 0.99 Columns: monkeys mean col : mean col : = = SS C = ()()( ) + ()()( ) = df C = = s C = = F C (, 88) = =.5694 Interaction: SS RXC = 68.5 ( ) = df RXC = ( )( ) = s RXC = = F RXC (, 88) = = 0.64 SS df MS F F crit p-value Rows Columns R X C wc Total There is not a significant main effect for eggs (rows) on the health of teams, F(,88) = 0.990, p = 0.. 9
30 There is not a significant main effect for monkeys (columns), F(,88) =.570, p = 0.4. There is not a significant interaction between eggs and monkeys, F(,88) = 0.60, p =
31 8) Within cell: SSwc = = dfwc = 68 ()() = 64 s wc = = 0.99 Rows: video games mean row = mean row = = = SS R = ()(7)( ) + ()(7)( ) = df R = = s R = = F R (, 64) = = 0.7 Columns: otter pops mean col : mean col : = = SS C = ()(7)( ) + ()(7)( ) = df C = = s C = = F C (, 64) = = 4.76 Interaction: SS RXC =.89 ( ) = df RXC = ( )( ) = s RXC = = F RXC (, 64) = = SS df MS F F crit p-value Rows Columns < R X C < wc Total There is not a significant main effect for video games (rows) on the taste of fathers, F(,64) = 0.70, p =
32 There is a significant main effect for otter pops (columns), F(,64) = 4.760, p = There is a significant interaction between video games and otter pops, F(,64) = 84.0, p =
33 9) Within cell: SSwc = = dfwc = 56 ()(4) = 48 s wc = = Rows: spaghetti mean row = mean row = = = SS R = (4)(7)( ) + (4)(7)( ) = 9.85 df R = = s R = 9.85 = 9.85 F R (, 48) = = 0.44 Columns: dinosaurs mean col : mean col : mean col : mean col 4 : = 5.98 = = = SS C = ()(7)( ) + ()(7)( ) + ()(7)( ) + ()(7)( ) = 7.50 df C = 4 = s C = 7.50 = F C (, 48) = =.00 Interaction: SS RXC = ( ) = 55.7 df RXC = ( )(4 ) = s RXC = 55.7 = F RXC (, 48) = =.9408
34 SS df MS F F crit p-value Rows Columns R X C wc Total There is not a significant main effect for spaghetti (rows) on the warmth of grandmothers, F(,48) = 0.440, p = There is not a significant main effect for dinosaurs (columns), F(,48) =.000, p = There is not a significant interaction between spaghetti and dinosaurs, F(,48) =.940, p = 0.6. warmth of grandmothers spaghetti spaghetti 50 4 dinosaurs 4
35 0) Within cell: SSwc = = dfwc = 60 ()() = 54 s wc = = Rows: brains mean row = mean row = = 4.70 = 5.6 SS R = ()(0)( ) + ()(0)( ) = df R = = s R = = F R (, 54) = = 0.7 Columns: cartoon characters mean col : mean col : mean col : = 4.05 = = SS C = ()(0)( ) + ()(0)( ) + ()(0)( ) = 8.06 df C = = s C = 8.06 = 9.0 F C (, 54) = =.90 Interaction: SS RXC = ( ) = df RXC = ( )( ) = s RXC = = F RXC (, 54) = =
36 SS df MS F F crit p-value Rows Columns R X C wc Total There is not a significant main effect for brains (rows) on the clothing of beers, F(,54) = 0.70, p = There is not a significant main effect for cartoon characters (columns), F(,54) =.900, p = There is a significant interaction between brains and cartoon characters, F(,54) = 5.460, p =
February 5, 2018 START HERE. measurement scale. Means. number of means. independent measures t-test Ch dependent measures t-test Ch 16.
χ Test for Frequencies February 5, 018 Contents Chi squared (χ ) test for frequencies Example 1: left vs right handers in our class The χ distribution One or two tailed? Example : Birthdays by month Using
More informationχ 2 Test for Frequencies January 15, 2019 Contents
χ 2 Test for Frequencies January 15, 2019 Contents Chi squared (χ 2 ) test for frequencies Example 1: left vs right handers in our class The χ 2 distribution One or two tailed? Example 2: Birthdays by
More informationT-tests for 2 Independent Means
T-tests for 2 Independent Means February 22, 208 Contents t-test for 2 independent means Tutorial Example : one-tailed test for independent means, equal sample sizes Error Bars Example 2: two-tailed test
More informationT-tests for 2 Independent Means
T-tests for 2 Independent Means January 15, 2019 Contents t-test for 2 independent means Tutorial Example 1: one-tailed test for independent means, equal sample sizes Error Bars Example 2: two-tailed test
More informationSampling Distributions: Central Limit Theorem
Review for Exam 2 Sampling Distributions: Central Limit Theorem Conceptually, we can break up the theorem into three parts: 1. The mean (µ M ) of a population of sample means (M) is equal to the mean (µ)
More informationStudent s t-distribution. The t-distribution, t-tests, & Measures of Effect Size
Student s t-distribution The t-distribution, t-tests, & Measures of Effect Size Sampling Distributions Redux Chapter 7 opens with a return to the concept of sampling distributions from chapter 4 Sampling
More informationHYPOTHESIS TESTING. Hypothesis Testing
MBA 605 Business Analytics Don Conant, PhD. HYPOTHESIS TESTING Hypothesis testing involves making inferences about the nature of the population on the basis of observations of a sample drawn from the population.
More informationPsych 230. Psychological Measurement and Statistics
Psych 230 Psychological Measurement and Statistics Pedro Wolf December 9, 2009 This Time. Non-Parametric statistics Chi-Square test One-way Two-way Statistical Testing 1. Decide which test to use 2. State
More informationUnit 27 One-Way Analysis of Variance
Unit 27 One-Way Analysis of Variance Objectives: To perform the hypothesis test in a one-way analysis of variance for comparing more than two population means Recall that a two sample t test is applied
More information" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2
Notation and Equations for Final Exam Symbol Definition X The variable we measure in a scientific study n The size of the sample N The size of the population M The mean of the sample µ The mean of the
More informationOHSU OGI Class ECE-580-DOE :Design of Experiments Steve Brainerd
Why We Use Analysis of Variance to Compare Group Means and How it Works The question of how to compare the population means of more than two groups is an important one to researchers. Let us suppose that
More informationThe One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)
The One-Way Repeated-Measures ANOVA (For Within-Subjects Designs) Logic of the Repeated-Measures ANOVA The repeated-measures ANOVA extends the analysis of variance to research situations using repeated-measures
More informationPower. January 12, 2019
Power January 12, 2019 Contents Definition of power Z-test example If H 0 is true If H 0 is false The 2x2 matrix of All Things That Can Happen Pr(Type I error) = α Pr(Type II error) = β power = 1 β Things
More informationHypothesis testing: Steps
Review for Exam 2 Hypothesis testing: Steps Repeated-Measures ANOVA 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region
More informationCh. 11 Inference for Distributions of Categorical Data
Ch. 11 Inference for Distributions of Categorical Data CH. 11 2 INFERENCES FOR RELATIONSHIPS The two sample z procedures from Ch. 10 allowed us to compare proportions of successes in two populations or
More informationOne-Way Analysis of Variance (ANOVA) Paul K. Strode, Ph.D.
One-Way Analysis of Variance (ANOVA) Paul K. Strode, Ph.D. Purpose While the T-test is useful to compare the means of two samples, many biology experiments involve the parallel measurement of three or
More informationRama Nada. -Ensherah Mokheemer. 1 P a g e
- 9 - Rama Nada -Ensherah Mokheemer - 1 P a g e Quick revision: Remember from the last lecture that chi square is an example of nonparametric test, other examples include Kruskal Wallis, Mann Whitney and
More informationWritten Exam (2 hours)
M. Müller Applied Analysis of Variance and Experimental Design Summer 2015 Written Exam (2 hours) General remarks: Open book exam. Switch off your mobile phone! Do not stay too long on a part where you
More informationHypothesis testing: Steps
Review for Exam 2 Hypothesis testing: Steps Exam 2 Review 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region 3. Compute
More informationUsing SPSS for One Way Analysis of Variance
Using SPSS for One Way Analysis of Variance This tutorial will show you how to use SPSS version 12 to perform a one-way, between- subjects analysis of variance and related post-hoc tests. This tutorial
More informationAnalysis of Variance: Part 1
Analysis of Variance: Part 1 Oneway ANOVA When there are more than two means Each time two means are compared the probability (Type I error) =α. When there are more than two means Each time two means are
More informationAnalysis of Variance (ANOVA)
Analysis of Variance (ANOVA) Two types of ANOVA tests: Independent measures and Repeated measures Comparing 2 means: X 1 = 20 t - test X 2 = 30 How can we Compare 3 means?: X 1 = 20 X 2 = 30 X 3 = 35 ANOVA
More informationAn Analysis of College Algebra Exam Scores December 14, James D Jones Math Section 01
An Analysis of College Algebra Exam s December, 000 James D Jones Math - Section 0 An Analysis of College Algebra Exam s Introduction Students often complain about a test being too difficult. Are there
More informationOne-Way ANOVA. Some examples of when ANOVA would be appropriate include:
One-Way ANOVA 1. Purpose Analysis of variance (ANOVA) is used when one wishes to determine whether two or more groups (e.g., classes A, B, and C) differ on some outcome of interest (e.g., an achievement
More informationLecture 11: Two Way Analysis of Variance
Lecture 11: Two Way Analysis of Variance Review: Hypothesis Testing o ANOVA/F ratio: comparing variances o F = s variance between treatment effect + chance s variance within sampling error (chance effects)
More informationRegression With a Categorical Independent Variable
Regression With a Categorical Independent Variable Lecture 15 March 17, 2005 Applied Regression Analysis Lecture #15-3/17/2005 Slide 1 of 29 Today s Lecture» Today s Lecture» Midterm Note» Example Regression
More informationStatistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017
Statistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017 I. χ 2 or chi-square test Objectives: Compare how close an experimentally derived value agrees with an expected value. One method to
More informationUNIVERSITY OF TORONTO Faculty of Arts and Science
UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator
More informationCalculating Fobt for all possible combinations of variances for each sample Calculating the probability of (F) for each different value of Fobt
PSY 305 Module 5-A AVP Transcript During the past two modules, you have been introduced to inferential statistics. We have spent time on z-tests and the three types of t-tests. We are now ready to move
More informationONE FACTOR COMPLETELY RANDOMIZED ANOVA
MALLOY PSYCH 3000 1-ANOVA PAGE 1 ONE FACTOR COMPLETELY RANDOMIZED ANOVA Sampling Distribution of F F is a test statistic [ ][ ][ ][ ] Test Statistic: F = MALLOY PSYCH 3000 1-ANOVA PAGE 2 ONE WAY ANOVA
More informationLecture 9. Selected material from: Ch. 12 The analysis of categorical data and goodness of fit tests
Lecture 9 Selected material from: Ch. 12 The analysis of categorical data and goodness of fit tests Univariate categorical data Univariate categorical data are best summarized in a one way frequency table.
More informationMathematical Notation Math Introduction to Applied Statistics
Mathematical Notation Math 113 - Introduction to Applied Statistics Name : Use Word or WordPerfect to recreate the following documents. Each article is worth 10 points and should be emailed to the instructor
More informationHYPOTHESIS TESTING: THE CHI-SQUARE STATISTIC
1 HYPOTHESIS TESTING: THE CHI-SQUARE STATISTIC 7 steps of Hypothesis Testing 1. State the hypotheses 2. Identify level of significant 3. Identify the critical values 4. Calculate test statistics 5. Compare
More informationOne-way ANOVA. Experimental Design. One-way ANOVA
Method to compare more than two samples simultaneously without inflating Type I Error rate (α) Simplicity Few assumptions Adequate for highly complex hypothesis testing 09/30/12 1 Outline of this class
More informationChapter 9: How can I tell if scores differ between three or more groups? One-way independent measures ANOVA.
Chapter 9: How can I tell if scores differ between three or more groups? One-way independent measures ANOVA. Full answers to study questions 1. Familywise error as a result of conducting multiple t tests.
More informationREVIEW 8/2/2017 陈芳华东师大英语系
REVIEW Hypothesis testing starts with a null hypothesis and a null distribution. We compare what we have to the null distribution, if the result is too extreme to belong to the null distribution (p
More informationOne-way between-subjects ANOVA. Comparing three or more independent means
One-way between-subjects ANOVA Comparing three or more independent means Data files SpiderBG.sav Attractiveness.sav Homework: sourcesofself-esteem.sav ANOVA: A Framework Understand the basic principles
More informationOne-Way ANOVA Source Table J - 1 SS B / J - 1 MS B /MS W. Pairwise Post-Hoc Comparisons of Means
One-Way ANOVA Source Table ANOVA MODEL: ij = µ* + α j + ε ij H 0 : µ 1 = µ =... = µ j or H 0 : Σα j = 0 Source Sum of Squares df Mean Squares F Between Groups n j ( j - * ) J - 1 SS B / J - 1 MS B /MS
More informationThe One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)
The One-Way Independent-Samples ANOVA (For Between-Subjects Designs) Computations for the ANOVA In computing the terms required for the F-statistic, we won t explicitly compute any sample variances or
More informationKeppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares
Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares K&W introduce the notion of a simple experiment with two conditions. Note that the raw data (p. 16)
More informationSection 9.4. Notation. Requirements. Definition. Inferences About Two Means (Matched Pairs) Examples
Objective Section 9.4 Inferences About Two Means (Matched Pairs) Compare of two matched-paired means using two samples from each population. Hypothesis Tests and Confidence Intervals of two dependent means
More informationReview for Final. Chapter 1 Type of studies: anecdotal, observational, experimental Random sampling
Review for Final For a detailed review of Chapters 1 7, please see the review sheets for exam 1 and. The following only briefly covers these sections. The final exam could contain problems that are included
More informationAcknowledge error Smaller samples, less spread
Hypothesis Testing with t Tests Al Arlo Clark-Foos kf Using Samples to Estimate Population Parameters Acknowledge error Smaller samples, less spread s = Σ ( X M N 1 ) 2 The t Statistic Indicates the distance
More informationWhile you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1
While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1 Chapter 12 Analysis of Variance McGraw-Hill, Bluman, 7th ed., Chapter 12 2
More informationOne-way Analysis of Variance. Major Points. T-test. Ψ320 Ainsworth
One-way Analysis of Variance Ψ30 Ainsworth Major Points Problem with t-tests and multiple groups The logic behind ANOVA Calculations Multiple comparisons Assumptions of analysis of variance Effect Size
More informationHypothesis Testing. We normally talk about two types of hypothesis: the null hypothesis and the research or alternative hypothesis.
Hypothesis Testing Today, we are going to begin talking about the idea of hypothesis testing how we can use statistics to show that our causal models are valid or invalid. We normally talk about two types
More informationPLSC PRACTICE TEST ONE
PLSC 724 - PRACTICE TEST ONE 1. Discuss briefly the relationship between the shape of the normal curve and the variance. 2. What is the relationship between a statistic and a parameter? 3. How is the α
More informationLecture 5: ANOVA and Correlation
Lecture 5: ANOVA and Correlation Ani Manichaikul amanicha@jhsph.edu 23 April 2007 1 / 62 Comparing Multiple Groups Continous data: comparing means Analysis of variance Binary data: comparing proportions
More informationCOMPARING SEVERAL MEANS: ANOVA
LAST UPDATED: November 15, 2012 COMPARING SEVERAL MEANS: ANOVA Objectives 2 Basic principles of ANOVA Equations underlying one-way ANOVA Doing a one-way ANOVA in R Following up an ANOVA: Planned contrasts/comparisons
More informationQuestions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.
Chapter 7 Reading 7.1, 7.2 Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.112 Introduction In Chapter 5 and 6, we emphasized
More informationQuestion. Hypothesis testing. Example. Answer: hypothesis. Test: true or not? Question. Average is not the mean! μ average. Random deviation or not?
Hypothesis testing Question Very frequently: what is the possible value of μ? Sample: we know only the average! μ average. Random deviation or not? Standard error: the measure of the random deviation.
More informationPSYC 331 STATISTICS FOR PSYCHOLOGISTS
PSYC 331 STATISTICS FOR PSYCHOLOGISTS Session 4 A PARAMETRIC STATISTICAL TEST FOR MORE THAN TWO POPULATIONS Lecturer: Dr. Paul Narh Doku, Dept of Psychology, UG Contact Information: pndoku@ug.edu.gh College
More information16.400/453J Human Factors Engineering. Design of Experiments II
J Human Factors Engineering Design of Experiments II Review Experiment Design and Descriptive Statistics Research question, independent and dependent variables, histograms, box plots, etc. Inferential
More informationAn Old Research Question
ANOVA An Old Research Question The impact of TV on high-school grade Watch or not watch Two groups The impact of TV hours on high-school grade Exactly how much TV watching would make difference Multiple
More informationStudy Guide #3: OneWay ANALYSIS OF VARIANCE (ANOVA)
Study Guide #3: OneWay ANALYSIS OF VARIANCE (ANOVA) About the ANOVA Test In educational research, we are most often involved finding out whether there are differences between groups. For example, is there
More informationContingency Tables. Safety equipment in use Fatal Non-fatal Total. None 1, , ,128 Seat belt , ,878
Contingency Tables I. Definition & Examples. A) Contingency tables are tables where we are looking at two (or more - but we won t cover three or more way tables, it s way too complicated) factors, each
More informationOne-way between-subjects ANOVA. Comparing three or more independent means
One-way between-subjects ANOVA Comparing three or more independent means ANOVA: A Framework Understand the basic principles of ANOVA Why it is done? What it tells us? Theory of one-way between-subjects
More informationCHI SQUARE ANALYSIS 8/18/2011 HYPOTHESIS TESTS SO FAR PARAMETRIC VS. NON-PARAMETRIC
CHI SQUARE ANALYSIS I N T R O D U C T I O N T O N O N - P A R A M E T R I C A N A L Y S E S HYPOTHESIS TESTS SO FAR We ve discussed One-sample t-test Dependent Sample t-tests Independent Samples t-tests
More information2 and F Distributions. Barrow, Statistics for Economics, Accounting and Business Studies, 4 th edition Pearson Education Limited 2006
and F Distributions Lecture 9 Distribution The distribution is used to: construct confidence intervals for a variance compare a set of actual frequencies with expected frequencies test for association
More informationCHAPTER 10 ONE-WAY ANALYSIS OF VARIANCE. It would be very unusual for all the research one might conduct to be restricted to
CHAPTER 10 ONE-WAY ANALYSIS OF VARIANCE It would be very unusual for all the research one might conduct to be restricted to comparisons of only two samples. Respondents and various groups are seldom divided
More informationIntroduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs
Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique
More informationSampling Distributions
Sampling Distributions Sampling Distribution of the Mean & Hypothesis Testing Remember sampling? Sampling Part 1 of definition Selecting a subset of the population to create a sample Generally random sampling
More informationArea1 Scaled Score (NAPLEX) .535 ** **.000 N. Sig. (2-tailed)
Institutional Assessment Report Texas Southern University College of Pharmacy and Health Sciences "An Analysis of 2013 NAPLEX, P4-Comp. Exams and P3 courses The following analysis illustrates relationships
More informationContingency Tables. Contingency tables are used when we want to looking at two (or more) factors. Each factor might have two more or levels.
Contingency Tables Definition & Examples. Contingency tables are used when we want to looking at two (or more) factors. Each factor might have two more or levels. (Using more than two factors gets complicated,
More informationLecture 18: Analysis of variance: ANOVA
Lecture 18: Announcements: Exam has been graded. See website for results. Lecture 18: Announcements: Exam has been graded. See website for results. Reading: Vasilj pp. 83-97. Lecture 18: Announcements:
More informationOne-factor analysis of variance (ANOVA)
One-factor analysis of variance (ANOVA) March 1, 2017 psych10.stanford.edu Announcements / Action Items Schedule update: final R lab moved to Week 10 Optional Survey 5 coming soon, due on Saturday Last
More informationIntroduction to Business Statistics QM 220 Chapter 12
Department of Quantitative Methods & Information Systems Introduction to Business Statistics QM 220 Chapter 12 Dr. Mohammad Zainal 12.1 The F distribution We already covered this topic in Ch. 10 QM-220,
More informationLab #12: Exam 3 Review Key
Psychological Statistics Practice Lab#1 Dr. M. Plonsky Page 1 of 7 Lab #1: Exam 3 Review Key 1) a. Probability - Refers to the likelihood that an event will occur. Ranges from 0 to 1. b. Sampling Distribution
More informationBlack White Total Observed Expected χ 2 = (f observed f expected ) 2 f expected (83 126) 2 ( )2 126
Psychology 60 Fall 2013 Practice Final Actual Exam: This Wednesday. Good luck! Name: To view the solutions, check the link at the end of the document. This practice final should supplement your studying;
More informationModule 7 Practice problem and Homework answers
Module 7 Practice problem and Homework answers Practice problem, page 1 Is the research hypothesis one-tailed or two-tailed? Answer: one tailed In the set up for the problem, we predicted a specific outcome
More informationAnalysis of variance
Analysis of variance 1 Method If the null hypothesis is true, then the populations are the same: they are normal, and they have the same mean and the same variance. We will estimate the numerical value
More informationReview. One-way ANOVA, I. What s coming up. Multiple comparisons
Review One-way ANOVA, I 9.07 /15/00 Earlier in this class, we talked about twosample z- and t-tests for the difference between two conditions of an independent variable Does a trial drug work better than
More informationHypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -
Hypothesis Testing with One-Way ANOVA Statistics Arlo Clark-Foos Conceptual Refresher 1. Standardized z distribution of scores and of means can be represented as percentile rankings. 2. t distribution
More informationSalt Lake Community College MATH 1040 Final Exam Fall Semester 2011 Form E
Salt Lake Community College MATH 1040 Final Exam Fall Semester 011 Form E Name Instructor Time Limit: 10 minutes Any hand-held calculator may be used. Computers, cell phones, or other communication devices
More informationChapter 7 Comparison of two independent samples
Chapter 7 Comparison of two independent samples 7.1 Introduction Population 1 µ σ 1 1 N 1 Sample 1 y s 1 1 n 1 Population µ σ N Sample y s n 1, : population means 1, : population standard deviations N
More informationModule 03 Lecture 14 Inferential Statistics ANOVA and TOI
Introduction of Data Analytics Prof. Nandan Sudarsanam and Prof. B Ravindran Department of Management Studies and Department of Computer Science and Engineering Indian Institute of Technology, Madras Module
More information21.0 Two-Factor Designs
21.0 Two-Factor Designs Answer Questions 1 RCBD Concrete Example Two-Way ANOVA Popcorn Example 21.4 RCBD 2 The Randomized Complete Block Design is also known as the two-way ANOVA without interaction. A
More informationStatistical Analysis How do we know if it works? Group workbook: Cartoon from XKCD.com. Subscribe!
Statistical Analysis How do we know if it works? Group workbook: Cartoon from XKCD.com. Subscribe! http://www.xkcd.com/552/ Significant Concepts We structure the presentation and processing of data to
More informationBusiness Statistics. Lecture 10: Course Review
Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,
More informationEconometrics. 4) Statistical inference
30C00200 Econometrics 4) Statistical inference Timo Kuosmanen Professor, Ph.D. http://nomepre.net/index.php/timokuosmanen Today s topics Confidence intervals of parameter estimates Student s t-distribution
More informationCHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)
FROM: PAGANO, R. R. (007) I. INTRODUCTION: DISTINCTION BETWEEN PARAMETRIC AND NON-PARAMETRIC TESTS Statistical inference tests are often classified as to whether they are parametric or nonparametric Parameter
More informationChapter 26: Comparing Counts (Chi Square)
Chapter 6: Comparing Counts (Chi Square) We ve seen that you can turn a qualitative variable into a quantitative one (by counting the number of successes and failures), but that s a compromise it forces
More informationSummary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)
Summary of Chapter 7 (Sections 7.2-7.5) and Chapter 8 (Section 8.1) Chapter 7. Tests of Statistical Hypotheses 7.2. Tests about One Mean (1) Test about One Mean Case 1: σ is known. Assume that X N(µ, σ
More informationFactorial Independent Samples ANOVA
Factorial Independent Samples ANOVA Liljenquist, Zhong and Galinsky (2010) found that people were more charitable when they were in a clean smelling room than in a neutral smelling room. Based on that
More informationCHAPTER 4 Analysis of Variance. One-way ANOVA Two-way ANOVA i) Two way ANOVA without replication ii) Two way ANOVA with replication
CHAPTER 4 Analysis of Variance One-way ANOVA Two-way ANOVA i) Two way ANOVA without replication ii) Two way ANOVA with replication 1 Introduction In this chapter, expand the idea of hypothesis tests. We
More informationRegression, part II. I. What does it all mean? A) Notice that so far all we ve done is math.
Regression, part II I. What does it all mean? A) Notice that so far all we ve done is math. 1) One can calculate the Least Squares Regression Line for anything, regardless of any assumptions. 2) But, if
More informationSTAT 135 Lab 11 Tests for Categorical Data (Fisher s Exact test, χ 2 tests for Homogeneity and Independence) and Linear Regression
STAT 135 Lab 11 Tests for Categorical Data (Fisher s Exact test, χ 2 tests for Homogeneity and Independence) and Linear Regression Rebecca Barter April 20, 2015 Fisher s Exact Test Fisher s Exact Test
More informationInferences for Regression
Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In
More informationStatistics: revision
NST 1B Experimental Psychology Statistics practical 5 Statistics: revision Rudolf Cardinal & Mike Aitken 29 / 30 April 2004 Department of Experimental Psychology University of Cambridge Handouts: Answers
More information8/23/2018. One-Way ANOVA F-test. 1. Situation/hypotheses. 2. Test statistic. 3.Distribution. 4. Assumptions
PSY 5101: Advanced Statistics for Psychological and Behavioral Research 1 1. Situation/hypotheses 2. Test statistic One-Way ANOVA F-test One factor J>2 independent samples H o :µ 1 µ 2 µ J F 3.Distribution
More informationFactorial Analysis of Variance
Factorial Analysis of Variance Conceptual Example A repeated-measures t-test is more likely to lead to rejection of the null hypothesis if a) *Subjects show considerable variability in their change scores.
More information10/31/2012. One-Way ANOVA F-test
PSY 511: Advanced Statistics for Psychological and Behavioral Research 1 1. Situation/hypotheses 2. Test statistic 3.Distribution 4. Assumptions One-Way ANOVA F-test One factor J>2 independent samples
More informationLecture 14. Analysis of Variance * Correlation and Regression. The McGraw-Hill Companies, Inc., 2000
Lecture 14 Analysis of Variance * Correlation and Regression Outline Analysis of Variance (ANOVA) 11-1 Introduction 11-2 Scatter Plots 11-3 Correlation 11-4 Regression Outline 11-5 Coefficient of Determination
More informationLecture 14. Outline. Outline. Analysis of Variance * Correlation and Regression Analysis of Variance (ANOVA)
Outline Lecture 14 Analysis of Variance * Correlation and Regression Analysis of Variance (ANOVA) 11-1 Introduction 11- Scatter Plots 11-3 Correlation 11-4 Regression Outline 11-5 Coefficient of Determination
More informationMarketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12)
Marketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12) Remember: Z.05 = 1.645, Z.01 = 2.33 We will only cover one-sided hypothesis testing (cases 12.3, 12.4.2, 12.5.2,
More informationData Analysis and Statistical Methods Statistics 651
Data Analysis and Statistical Methods Statistics 65 http://www.stat.tamu.edu/~suhasini/teaching.html Suhasini Subba Rao Comparing populations Suppose I want to compare the heights of males and females
More informationLecture 3: Analysis of Variance II
Lecture 3: Analysis of Variance II http://www.stats.ox.ac.uk/ winkel/phs.html Dr Matthias Winkel 1 Outline I. A second introduction to two-way ANOVA II. Repeated measures design III. Independent versus
More informationDepartment of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance ECON 509. Dr.
Department of Economics Business Statistics Chapter 1 Chi-square test of independence & Analysis of Variance ECON 509 Dr. Mohammad Zainal Chapter Goals After completing this chapter, you should be able
More informationSurvey on Population Mean
MATH 203 Survey on Population Mean Dr. Neal, Spring 2009 The first part of this project is on the analysis of a population mean. You will obtain data on a specific measurement X by performing a random
More informationAnalysis of Variance
Analysis of Variance Math 36b May 7, 2009 Contents 2 ANOVA: Analysis of Variance 16 2.1 Basic ANOVA........................... 16 2.1.1 the model......................... 17 2.1.2 treatment sum of squares.................
More information