Chapter 13 Section D. F versus Q: Different Approaches to Controlling Type I Errors with Multiple Comparisons

Size: px
Start display at page:

Download "Chapter 13 Section D. F versus Q: Different Approaches to Controlling Type I Errors with Multiple Comparisons"

Transcription

1 Explaining Psychological Statistics (2 nd Ed.) by Barry H. Cohen Chapter 13 Section D F versus Q: Different Approaches to Controlling Type I Errors with Multiple Comparisons In section B of this chapter, I mentioned that when using Tukey s HSD it is unusual to find a pair of means that are significantly different when the overall ANOVA is not significant (p. 375; EPS, 2e). However, it is also possible, though unusual, to find no significant differences with the HSD procedure when the ANOVA is significant. Tukey s HSD is based on a different distribution from the ANOVA, and though it always yields similar results, there is some room for discrepancies with respect to statistical significance, as I will demonstrate in the following examples. ANOVA Not Significant, but HSD Finds Significance Suppose that your experiment consists of five groups of ten participants each, and each group is subjected to a different form of stress (e.g., speeded arithmetic; mild electric shock). For each group the average heart rate of the participants is calculated, yielding the following set of means (in beats per minute): 65, 71, 72, 73, 79. The MS bet for these means is just 10 times their unbiased variance, so MS bet equals 250. For simplicity, let s say that MS W happens to be exactly 100, so that F for the ANOVA equals 2.5. Given that the critical value for a.05 test is about 2.58, the ANOVA has fallen just short of conventional significance, and therefore followup t tests using the LSD formula would not be justified. However, the Type I error protection of Tukey s HSD does not depend on the significance of the ANOVA, so we can go right ahead and calculate HSD for this example. The appropriate q for this example is about 4.02; when multiplied by the square root of MS W over n, we obtain 12.7 as our value for HSD. The difference between the two extreme means (79-65 = 14) is larger than HSD, so use of Tukey s procedure can legitimately declare those two means to differ significantly at the.05 level, even though the omnibus ANOVA is not significant at that level. As I said, this is an unusual but not an impossible combination of events. Note that the ANOVA was indeed close to being significant; if for instance, MS W were as large as 125, making F as small as 2.0, HSD would become slightly larger than 14, the largest difference of the means. Next, using a similar example, I will show how the ANOVA can be significant without HSD finding any pair of means that differs significantly. ANOVA Is Significant, but HSD Does Not Find Significance The pattern of the five means in the preceding example is the kind that favors Tukey s HSD over the ANOVA: two of the means are relatively far apart while the rest of the means are clustered together in the middle. That clustering tends to reduce the variance of the means and therefore the ANOVA F. The following set of means follows the opposite pattern: 65, 66, 72, 78, 79; note that the largest difference of means is still 14, but the variance of the means is now 42.5, so MS bet equals 425. If MS W is still only 100, F will increase to an easily significant In fact, given the increase in MS bet, MS W can increase to 125, and the ANOVA would still be significant (425 / 125 = 3.4 > 2.58). However, increasing MS W would also produce an increase in HSD, even though there has been no change in the largest difference of means. As mentioned above, with MS W at 125, HSD would come out larger than the difference of 65 and 79 (HSD = 14.l2), so despite the significance of the ANOVA, HSD would no longer indicate any significantly different pairs of means. As you can see, the pattern that favors the ANOVA over Tukey s HSD is one in which the means are not clustered centrally, but tend instead to be clustered at the two extremes. When dealing with only three groups, perform LSD-type followup tests, but only if your ANOVA is significant. With

2 2 more than three groups, you are allowed to use Tukey s HSD without looking at the ANOVA, but tradition strongly dictates that you report the results of the ANOVA before reporting any conclusions from a post hoc test like HSD. Whereas it is unlikely to find significance with HSD but not the ANOVA, it is even less likely to find such results published in the literature. Fisher-Hayter Test Tukey s HSD is easy to use and understand but it is more conservative than necessary. Computer simulations have shown that under typical data analytic conditions this procedure tends to keep experiment-wise alpha between about.02 and.03 when you think you are setting the overall alpha at.05. Hayter (1986) devised a modification of HSD that employs the two-step process originated by Fisher (so it is often called the Fisher-Hayter or the modified LSD test) to squeeze more power out of HSD without allowing a EW to rise above.05. The first step of the Fisher-Hayter test is to evaluate the significance of the one-way ANOVA. If (and only if) the ANOVA is significant, you can proceed to the second step, which involves the calculation of HSD, but with an important twist : the critical value for HSD (i.e., q) is determined by setting the number of groups to k - 1, rather than k. The Fisher-Hayter (F-H) test is easily illustrated in terms of the immediately preceding example. Because the ANOVA was significant, we can proceed to calculate HSD as though the number of groups were four instead of five. In this case, q is only about 3.77, instead of 4.02, and HSD comes out to Now, the largest difference of means is larger than HSD, so we can identify a significantly different pair of means to follow up on our significant ANOVA. Note that the Fisher-Hayter test comes in two varieties: the F version, just described, and a Q version, whose first step is testing the largest difference of means with HSD based on k groups, in order to decide whether to proceed. For the preceding example, the F-H Q test would not have gotten through the first step. Similarly, in the first example (ANOVA not significant), the F-H F test could not proceed, but F-H Q would have gone to the second step. HSD comes out to for that second step, but beyond the largest difference of means, F-H Q would not find any additional significant pairs of means. Simultaneous versus Sequential Tests Tukey s HSD is a good example of a simultaneous post hoc comparison test. That is, the value of HSD is determined at the outset and is used for all pairwise comparisons; there is no part of the test that changes based on the results of another part. Simultaneous tests have an advantage in terms of the ease with which confidence intervals (CI s) can be found, but sequential tests often have greater power without sacrificing Type I error control. Fisher s protected (or LSD) test is the simplest example of a sequential test. Depending on the results of the first step (testing the ANOVA), there may or may not be a second step in the sequence. If the HSD you have calculated is based on a q value for the.05 level, it is very easy to find the 95% CI for any difference of population means; it is just the difference of the corresponding sample means plus or minus HSD. If, for instance, the average heart rate is 69 for the physically stressed group and 73 for the mentally stressed group (in the five-group experiment alluded to above), your point estimate for the difference of the two conditions if applied to the entire population would be73-69, which equals 4. If your.05 HSD turned out to be 4.5 for this experiment, your 95% CI for the physical/mental stress difference would range from -.5 to 8.5. That zero falls within the 95% CI tells you that this difference is not significant at the.05 level. Of course, you could also tell that from the fact that the sample difference of 4 is less than HSD (4.5).

3 One problem with basing CI s on LSD, instead of HSD, even for only three groups, is that you cannot begin with calculating CI s and then determine significance by noting whether zero is contained within a particular CI or not. Zero may be outside the range of a particular CI, but if the ANOVA is not significant, the population difference being estimated cannot be declared significant either. Although they do not lend themselves to the calculation of CI s, the increased power advantages of sequential tests are hard to ignore. The F-H test described above is a good example of a powerful sequential test that maintains adequate control over experimentwise error. You may notice that other multiple comparison tests are available in F and Q versions (e.g., the REGW test). This tells you that the test is a sequential one, and that the first step requires the significance of either the ordinary ANOVA (F version), or the largest pairwise difference according to the usual HSD test (the Q version) in order to proceed with further testing. Sharper Bonferroni Tests The ordinary Bonferroni adjustment, as applied to all possible pairs of means following an ANOVA is, as I mentioned in section A, much too conservative for routine use. For instance, there are ten possible pairs of means that could be tested in the five-group stress example, so an alpha of.05 / 10 =.005 would have to be used for each test. By way of contrast, q was 4.02, which is equivalent to a critical t of 4.02 / sqrt (2) = 2.84, which would correspond to using an alpha of.0068 better than.005, but still on the conservative side. The second stage of the F-H test involving five groups of ten participants involves a q of 3.77, and therefore an alpha per comparison of about.011 (if you get that far). Of course, if you can plan to test only five of the ten possible pairwise differences, you can do even better (alpha per comparison =.01), but there are ways to make the Bonferroni adjustment somewhat less conservative, even without planning any comparisons, as I will show next. Sidak s Test The Bonferroni test is based on an inequality that the overall (i.e., experimentwise) alpha will be less than or equal to the number of tests conducted (j) times the alpha used for each comparison (a pc ). So, in the case of a seven-group experiment, in which all of the 21 possible pairwise comparisons are to be tested (each at the.05 level), Bonferroni tells us that a EW will be less than or equal to j*a pc = 21*.05 = Of course, we already knew that just by the way we define probability. As j gets larger, the Bonferroni inequality becomes progressively less informative. However, if we can assume that all of the tests we are conducting are mutually independent, we can use a sharper (i.e., more accurate) inequality, based on Formula Solving that formula for alpha, we obtain an adjustment for a pc that is somewhat less severe than the Bonferroni adjustment (as expressed in Formula 13.8): Formula When you are performing all possible pairwise comparisons among your samples, your tests are not all mutually independent, but Sidak (1967) showed that even with this lack of independence the use of Formula keeps a EW below the level chosen, while providing more power than the traditional Bonferroni correction. Therefore, the use of the preceding formula is often referred to as Sidak s test. [Note: Formula can be derived from the unnumbered formula Sidak (1967) presents on page 629, just before section 4 of his article.] A couple of examples (based on a EW =.05) will help to illustrate the difference between the Bonferroni and Sidak adjustments.

4 First, let us consider the three-group case, in which there are only three different pairings that can be tested. According to the Sidak test, the alpha that should be used for each of the three tests is: a pc = 1 - (1 - a EW ) 1/3 = 1 - (.95) 1/3. The fractional exponent means that rather than cubing.95, we are to take the cube root of.95, which is about.98305, so a pc = = This is only slightly larger than the Bonferroni a pc (.05 / 3), which rounds off to = There isn t much difference between the two adjustments, but the Sidak alpha is always the larger of the two. For another example, consider the seven-group experiment. There are 21 possible pairwise comparisons, so Sidak s adjusted alpha for each comparison comes to: 1 - (.95) 1/21 = =.00244, which is a bit larger than.05 / 21, which equals Because Sidak s test does not make a large difference, and is computationally more complex than the Bonferroni test, it was not popular before readily available statistical packages (e.g., SPSS) started to include it. Various tests that incorporate the Bonferroni adjustment can be made a bit more powerful by using instead the Sidak adjustment. Another way to add power to the Bonferroni adjustment is to turn it into a sequential test, as shown next. Sequential Bonferroni Tests A Step-Down Bonferroni Test: Holm (1979) demonstrated that you could add power to your Bonferroni test with the following sequentially rejective (step-down) procedure. The first step is to check whether any of your comparisons are significant with the ordinary Bonferroni adjustment; if not even one is significant by the unmodified Bonferroni criterion testing stops, and no significant comparisons are found. If you are testing all possible pairs of five means, as in the previous example, you would determine the p values for each of the ten pairwise comparisons, and then see if your smallest p value is less than.005. If it is, you declare it to be significant at the.05 level, and then compare the next smallest p with.05 / 9 = In terms of a general formula, to be significant, a comparison must have a p i < a EW / (m - i + 1), where p i goes from p 1 (the smallest p) to p m (the largest p), and m is the total number of comparisons in the set. However, you must start with the smallest p, and go up step by step (following the proper sequence), and stop as soon as you hit a p that is not significant according to the preceding formula. For example, suppose that the five-group stress study had large enough samples so that the p values for the ten possible pairwise comparisons are as follows (ordered from smallest to largest):.002,.0054,.007,.008,.009,.0094,.012,.015,.028,.067. Because the smallest p is less than.005, it is, according to Holm s procedure, significant. The second smallest,.0054, does not have to be smaller than.005, it only has to be less than.0056, and it is. Next,.007 is compared to.05 / 8 =.00625; because p 3 is not less than this value, it is not declared significant, and testing stops at this point. It does not matter that p 6 (.0094) happens to be less than.05 / 5 =.01; once you hit a nonsignificant result in the sequence you cannot test any larger p values without ruining the Type I error control. You can see the extra power in this approach in that the p of.0054 would not have been significant according to the ordinary Bonferroni correction, but it is significant according to Holm s step-down test. It is worth noting that more than one modification of the Bonferroni procedure can be applied within the same test. For instance, Holland and Copenhaver (1988) showed that the power of Holm s test can be improved slightly by basing it on Sidak s (sharper) inequality, rather than the usual Bonferroni adjustment. However, Olejnik, Li, Supattathum, and Huberty (1997) pointed out that Holland and Copenhaver s test requires an assumption that makes it somewhat less generally applicable than Holm s test, which makes no such assumptions. 4

5 A Step-Up Bonferroni Test: More recently, Hochberg (1986) demonstrated that a step-up procedure could have more power than Holm s step-down test. His step-up test begins with a test of the largest p value in the set. If this p is less than.05, then all of the comparisons are declared significant (e.g., if all 10 of the p s are between.02 and.04, they are declared significant at the.05 level, even though none of them would be significant with the usual Bonferroni adjustment). You use the same formula as Holm s test, but you apply it in the reverse order. To illustrate I will again use the set of ten p values above. First, p m,the largest p value, is compared to a EW / (m - i + 1) =.05 / (m - m +1) =.05 / 1 =.05. Because.067 is not less than.05, this comparison is not declared significant but testing does not stop. The second largest p is then compared to.05 / ( ) =.05 / (1+1) =.025. Because.028 is not less than.025, this result is not significant either, and the procedure continues. The next p, p 3, is compared to.05 / 3 =.0167; it (.015) is less than this value, so this comparison is significant, and therefore testing stops. All of the p values smaller than p 3 are automatically declared significant without further testing. It is easy to see that Holm s test is not as powerful as Hochberg s test. The former test found only the p s of.002 and.0051 to be significant, whereas the latter found all of the p values from.007 through.015 to be significant, as well. Of course, I deliberately devised an example to emphasize the difference between these two procedures. In most real-life cases, the conclusions from the two methods will rarely differ. Considering that Hochberg s procedure rests on the assumption that the tests are independent of one another (Olejnik, Li, Supattathum, & Huberty, 1997), and Holm s test does not, it seems that the conservative way to apply the Bonferroni test is by the use of either Holm s procedure, or Sidak s adjustment (Formula 13.16). Of course, you can still use the simpler, unmodified Bonferroni alpha correction, but if you are analyzing your data with statistical software, there is no excuse for throwing away even a small increase in power. Adjusted p Values When you request a Bonferroni test from SPSS under post hoc comparisons, what you get for each pair of means is a p value that is adjusted so that it can be compared directly to.05, assuming that that is your desired experimentwise alpha. For instance, for a three-group experiment, a pairwise comparison (i.e., a t test) that yields a p value of.016 would be considered significant at the.05 level, because.016 < (.05 / 3). Instead of giving you the actual two-tailed p value, SPSS adjusts the p value by multiplying it by 3, in this case, and gives you a Bonferroni p of.048 (.016*3), which you can see immediately is just under.05, and therefore significant by the Bonferroni test. Quite simply, SPSS adjusts the actual p value by applying the Bonferroni correction backwards. In the general case, without SPSS, you would divide.05 by the total number of possible pairwise comparisons (if conservative enough to use Bonferroni as a post hoc test), and then compare each of your p values to that shrunken value. SPSS performs the opposite operation, and multiplies each of your actual p values by the total number of possible pairs, so each can be compared to.05 (or whatever value you want a EW to be less than). To express the above operation as a formula, solve Formula 13.8 for a EW, and then change the term a EW to the Bonferroni adjusted p, and change a pc in that formula to your actual p value, like this: 5

6 Of course, you don t need to see a formula in this simple case, but the formula approach makes it easier to understand more complex adjustments, such as the one SPSS refers to as the Sidak adjustment. To obtain the formula for the Sidak-adjusted p, you have to solve Formula above for a EW but that just brings you back to Formula Changing a EW in Formula 3.2 to Sidak p, and plain alpha to your actual p value, you get the formula that SPSS uses to turn your p values into Sidak-adjusted p values: 6 Formula where j is the total number of possible comparisons. In the three-group Bonferroni example above, a p value of.016 was adjusted to.048. The corresponding Sidak p is = =.047. Not a big improvement, but if the computer is doing the work, why not get all the power you can (without sacrificing Type I error control)? References Hayter, A. J. (1986). The maximum familywise error rate of Fisher's least significant difference test. Journal of the American Statistical Association, 81, Hochberg, Y. (1988). A sharper Bonferroni procedure for multiple tests of significance. Biometrika, 75, Holland, B. S., & Copenhaver M. D. (1988). Improved Bonferroni-type multiple testing procedures. Psychological Bulletin, 104, Holm, S. (1979). A simple sequentially rejective multiple test procedure. Scandinavian Journal of Statistics, 6, Olejnik, S., Li, J., Supattathum, S., & Huberty, C. J. (1997). Multiple testing and statistical power with modified Bonferroni procedures. Journal of Educational and Behavioral Statistics, 22, Sidak, Z. (1967). Rectangular confidence regions for the means of multivariate normal distributions. Journal of the American Statistical Association, 62,

Multiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600

Multiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600 Multiple Comparison Procedures Cohen Chapter 13 For EDUC/PSY 6600 1 We have to go to the deductions and the inferences, said Lestrade, winking at me. I find it hard enough to tackle facts, Holmes, without

More information

Multiple Comparisons

Multiple Comparisons Multiple Comparisons Error Rates, A Priori Tests, and Post-Hoc Tests Multiple Comparisons: A Rationale Multiple comparison tests function to tease apart differences between the groups within our IV when

More information

COMPARING SEVERAL MEANS: ANOVA

COMPARING SEVERAL MEANS: ANOVA LAST UPDATED: November 15, 2012 COMPARING SEVERAL MEANS: ANOVA Objectives 2 Basic principles of ANOVA Equations underlying one-way ANOVA Doing a one-way ANOVA in R Following up an ANOVA: Planned contrasts/comparisons

More information

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions Introduction to Analysis of Variance 1 Experiments with More than 2 Conditions Often the research that psychologists perform has more conditions than just the control and experimental conditions You might

More information

Chapter Seven: Multi-Sample Methods 1/52

Chapter Seven: Multi-Sample Methods 1/52 Chapter Seven: Multi-Sample Methods 1/52 7.1 Introduction 2/52 Introduction The independent samples t test and the independent samples Z test for a difference between proportions are designed to analyze

More information

H0: Tested by k-grp ANOVA

H0: Tested by k-grp ANOVA Analyses of K-Group Designs : Omnibus F, Pairwise Comparisons & Trend Analyses ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures

More information

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table Your schedule of coming weeks One-way ANOVA, II 9.07 //00 Today: One-way ANOVA, part II Next week: Two-way ANOVA, parts I and II. One-way ANOVA HW due Thursday Week of May Teacher out of town all week

More information

H0: Tested by k-grp ANOVA

H0: Tested by k-grp ANOVA Pairwise Comparisons ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures Alpha estimation reconsidered H0: Tested by k-grp ANOVA Regardless

More information

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs) The One-Way Independent-Samples ANOVA (For Between-Subjects Designs) Computations for the ANOVA In computing the terms required for the F-statistic, we won t explicitly compute any sample variances or

More information

Contrasts and Multiple Comparisons Supplement for Pages

Contrasts and Multiple Comparisons Supplement for Pages Contrasts and Multiple Comparisons Supplement for Pages 302-323 Brian Habing University of South Carolina Last Updated: July 20, 2001 The F-test from the ANOVA table allows us to test the null hypothesis

More information

Specific Differences. Lukas Meier, Seminar für Statistik

Specific Differences. Lukas Meier, Seminar für Statistik Specific Differences Lukas Meier, Seminar für Statistik Problem with Global F-test Problem: Global F-test (aka omnibus F-test) is very unspecific. Typically: Want a more precise answer (or have a more

More information

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique

More information

Keppel, G. & Wickens, T. D. Design and Analysis Chapter 12: Detailed Analyses of Main Effects and Simple Effects

Keppel, G. & Wickens, T. D. Design and Analysis Chapter 12: Detailed Analyses of Main Effects and Simple Effects Keppel, G. & Wickens, T. D. Design and Analysis Chapter 1: Detailed Analyses of Main Effects and Simple Effects If the interaction is significant, then less attention is paid to the two main effects, and

More information

STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5)

STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5) STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons Ch. 4-5) Recall CRD means and effects models: Y ij = µ i + ϵ ij = µ + α i + ϵ ij i = 1,..., g ; j = 1,..., n ; ϵ ij s iid N0, σ 2 ) If we reject

More information

Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model

Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model Biostatistics 250 ANOVA Multiple Comparisons 1 ORIGIN 1 Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model When the omnibus F-Test for ANOVA rejects the null hypothesis that

More information

DETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics

DETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics DETAILED CONTENTS About the Author Preface to the Instructor To the Student How to Use SPSS With This Book PART I INTRODUCTION AND DESCRIPTIVE STATISTICS 1. Introduction to Statistics 1.1 Descriptive and

More information

Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means

Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means 4.1 The Need for Analytical Comparisons...the between-groups sum of squares averages the differences

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

Week 5 Video 1 Relationship Mining Correlation Mining

Week 5 Video 1 Relationship Mining Correlation Mining Week 5 Video 1 Relationship Mining Correlation Mining Relationship Mining Discover relationships between variables in a data set with many variables Many types of relationship mining Correlation Mining

More information

Multiple Testing. Gary W. Oehlert. January 28, School of Statistics University of Minnesota

Multiple Testing. Gary W. Oehlert. January 28, School of Statistics University of Minnesota Multiple Testing Gary W. Oehlert School of Statistics University of Minnesota January 28, 2016 Background Suppose that you had a 20-sided die. Nineteen of the sides are labeled 0 and one of the sides is

More information

Analysis of Variance II Bios 662

Analysis of Variance II Bios 662 Analysis of Variance II Bios 662 Michael G. Hudgens, Ph.D. mhudgens@bios.unc.edu http://www.bios.unc.edu/ mhudgens 2008-10-24 17:21 BIOS 662 1 ANOVA II Outline Multiple Comparisons Scheffe Tukey Bonferroni

More information

Orthogonal, Planned and Unplanned Comparisons

Orthogonal, Planned and Unplanned Comparisons This is a chapter excerpt from Guilford Publications. Data Analysis for Experimental Design, by Richard Gonzalez Copyright 2008. 8 Orthogonal, Planned and Unplanned Comparisons 8.1 Introduction In this

More information

13: Additional ANOVA Topics. Post hoc Comparisons

13: Additional ANOVA Topics. Post hoc Comparisons 13: Additional ANOVA Topics Post hoc Comparisons ANOVA Assumptions Assessing Group Variances When Distributional Assumptions are Severely Violated Post hoc Comparisons In the prior chapter we used ANOVA

More information

Three Factor Completely Randomized Design with One Continuous Factor: Using SPSS GLM UNIVARIATE R. C. Gardner Department of Psychology

Three Factor Completely Randomized Design with One Continuous Factor: Using SPSS GLM UNIVARIATE R. C. Gardner Department of Psychology Data_Analysis.calm Three Factor Completely Randomized Design with One Continuous Factor: Using SPSS GLM UNIVARIATE R. C. Gardner Department of Psychology This article considers a three factor completely

More information

B. Weaver (18-Oct-2006) MC Procedures Chapter 1: Multiple Comparison Procedures ) C (1.1)

B. Weaver (18-Oct-2006) MC Procedures Chapter 1: Multiple Comparison Procedures ) C (1.1) B. Weaver (18-Oct-2006) MC Procedures... 1 Chapter 1: Multiple Comparison Procedures 1.1 Introduction The omnibus F-test in a one-way ANOVA is a test of the null hypothesis that the population means of

More information

Comparisons among means (or, the analysis of factor effects)

Comparisons among means (or, the analysis of factor effects) Comparisons among means (or, the analysis of factor effects) In carrying out our usual test that μ 1 = = μ r, we might be content to just reject this omnibus hypothesis but typically more is required:

More information

DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective

DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective Second Edition Scott E. Maxwell Uniuersity of Notre Dame Harold D. Delaney Uniuersity of New Mexico J,t{,.?; LAWRENCE ERLBAUM ASSOCIATES,

More information

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 22s:152 Applied Linear Regression Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) We now consider an analysis with only categorical predictors (i.e. all predictors are

More information

These are all actually contrasts (the coef sum to zero). What are these contrasts representing? What would make them large?

These are all actually contrasts (the coef sum to zero). What are these contrasts representing? What would make them large? Lecture 12 Comparing treatment effects Orthogonal Contrasts What use are contrasts? Recall the Cotton data In this case, the treatment levels have an ordering to them this is not always the case) Consider

More information

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES BIOL 458 - Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES PART 1: INTRODUCTION TO ANOVA Purpose of ANOVA Analysis of Variance (ANOVA) is an extremely useful statistical method

More information

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018 Math 403 - P. & S. III - Dr. McLoughlin - 1 2018 2 Hand-out 2 Dr. M. P. M. M. M c Loughlin Revised 2018 3. Fundamentals 3.1. Preliminaries. Suppose we can produce a random sample of weights of 10 year-olds

More information

Multiple Testing. Tim Hanson. January, Modified from originals by Gary W. Oehlert. Department of Statistics University of South Carolina

Multiple Testing. Tim Hanson. January, Modified from originals by Gary W. Oehlert. Department of Statistics University of South Carolina Multiple Testing Tim Hanson Department of Statistics University of South Carolina January, 2017 Modified from originals by Gary W. Oehlert Type I error A Type I error is to wrongly reject the null hypothesis

More information

Outline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013

Outline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013 Topic 19 - Inference - Fall 2013 Outline Inference for Means Differences in cell means Contrasts Multiplicity Topic 19 2 The Cell Means Model Expressed numerically Y ij = µ i + ε ij where µ i is the theoretical

More information

22s:152 Applied Linear Regression. 1-way ANOVA visual:

22s:152 Applied Linear Regression. 1-way ANOVA visual: 22s:152 Applied Linear Regression 1-way ANOVA visual: Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Y We now consider an analysis

More information

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI Introduction of Data Analytics Prof. Nandan Sudarsanam and Prof. B Ravindran Department of Management Studies and Department of Computer Science and Engineering Indian Institute of Technology, Madras Module

More information

Multiple Comparison Procedures, Trimmed Means and Transformed Statistics. Rhonda K. Kowalchuk Southern Illinois University Carbondale

Multiple Comparison Procedures, Trimmed Means and Transformed Statistics. Rhonda K. Kowalchuk Southern Illinois University Carbondale Multiple Comparison Procedures 1 Multiple Comparison Procedures, Trimmed Means and Transformed Statistics Rhonda K. Kowalchuk Southern Illinois University Carbondale H. J. Keselman University of Manitoba

More information

Regression Part II. One- factor ANOVA Another dummy variable coding scheme Contrasts Mul?ple comparisons Interac?ons

Regression Part II. One- factor ANOVA Another dummy variable coding scheme Contrasts Mul?ple comparisons Interac?ons Regression Part II One- factor ANOVA Another dummy variable coding scheme Contrasts Mul?ple comparisons Interac?ons One- factor Analysis of variance Categorical Explanatory variable Quan?ta?ve Response

More information

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n = Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,

More information

STAT22200 Spring 2014 Chapter 5

STAT22200 Spring 2014 Chapter 5 STAT22200 Spring 2014 Chapter 5 Yibi Huang April 29, 2014 Chapter 5 Multiple Comparisons Chapter 5-1 Chapter 5 Multiple Comparisons Note the t-tests and C.I. s are constructed assuming we only do one test,

More information

Multiple Comparison Methods for Means

Multiple Comparison Methods for Means SIAM REVIEW Vol. 44, No. 2, pp. 259 278 c 2002 Society for Industrial and Applied Mathematics Multiple Comparison Methods for Means John A. Rafter Martha L. Abell James P. Braselton Abstract. Multiple

More information

Hypothesis testing: Steps

Hypothesis testing: Steps Review for Exam 2 Hypothesis testing: Steps Repeated-Measures ANOVA 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region

More information

Independent Samples ANOVA

Independent Samples ANOVA Independent Samples ANOVA In this example students were randomly assigned to one of three mnemonics (techniques for improving memory) rehearsal (the control group; simply repeat the words), visual imagery

More information

4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES

4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES 4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES FOR SINGLE FACTOR BETWEEN-S DESIGNS Planned or A Priori Comparisons We previously showed various ways to test all possible pairwise comparisons for

More information

Analysis of Variance (ANOVA)

Analysis of Variance (ANOVA) Analysis of Variance (ANOVA) Two types of ANOVA tests: Independent measures and Repeated measures Comparing 2 means: X 1 = 20 t - test X 2 = 30 How can we Compare 3 means?: X 1 = 20 X 2 = 30 X 3 = 35 ANOVA

More information

Hypothesis testing: Steps

Hypothesis testing: Steps Review for Exam 2 Hypothesis testing: Steps Exam 2 Review 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region 3. Compute

More information

Chapter 14: Repeated-measures designs

Chapter 14: Repeated-measures designs Chapter 14: Repeated-measures designs Oliver Twisted Please, Sir, can I have some more sphericity? The following article is adapted from: Field, A. P. (1998). A bluffer s guide to sphericity. Newsletter

More information

ANOVA. Testing more than 2 conditions

ANOVA. Testing more than 2 conditions ANOVA Testing more than 2 conditions ANOVA Today s goal: Teach you about ANOVA, the test used to measure the difference between more than two conditions Outline: - Why anova? - Contrasts and post-hoc tests

More information

ON STEPWISE CONTROL OF THE GENERALIZED FAMILYWISE ERROR RATE. By Wenge Guo and M. Bhaskara Rao

ON STEPWISE CONTROL OF THE GENERALIZED FAMILYWISE ERROR RATE. By Wenge Guo and M. Bhaskara Rao ON STEPWISE CONTROL OF THE GENERALIZED FAMILYWISE ERROR RATE By Wenge Guo and M. Bhaskara Rao National Institute of Environmental Health Sciences and University of Cincinnati A classical approach for dealing

More information

A posteriori multiple comparison tests

A posteriori multiple comparison tests A posteriori multiple comparison tests 11/15/16 1 Recall the Lakes experiment Source of variation SS DF MS F P Lakes 58.000 2 29.400 8.243 0.006 Error 42.800 12 3.567 Total 101.600 14 The ANOVA tells us

More information

9 One-Way Analysis of Variance

9 One-Way Analysis of Variance 9 One-Way Analysis of Variance SW Chapter 11 - all sections except 6. The one-way analysis of variance (ANOVA) is a generalization of the two sample t test to k 2 groups. Assume that the populations of

More information

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests: One sided tests So far all of our tests have been two sided. While this may be a bit easier to understand, this is often not the best way to do a hypothesis test. One simple thing that we can do to get

More information

Chapter 3 ALGEBRA. Overview. Algebra. 3.1 Linear Equations and Applications 3.2 More Linear Equations 3.3 Equations with Exponents. Section 3.

Chapter 3 ALGEBRA. Overview. Algebra. 3.1 Linear Equations and Applications 3.2 More Linear Equations 3.3 Equations with Exponents. Section 3. 4 Chapter 3 ALGEBRA Overview Algebra 3.1 Linear Equations and Applications 3.2 More Linear Equations 3.3 Equations with Exponents 5 LinearEquations 3+ what = 7? If you have come through arithmetic, the

More information

Chapter 6 Planned Contrasts and Post-hoc Tests for one-way ANOVA

Chapter 6 Planned Contrasts and Post-hoc Tests for one-way ANOVA Chapter 6 Planned Contrasts and Post-hoc Tests for one-way NOV Page. The Problem of Multiple Comparisons 6-. Types of Type Error Rates 6-. Planned contrasts vs. Post hoc Contrasts 6-7 4. Planned Contrasts

More information

10. Alternative case influence statistics

10. Alternative case influence statistics 10. Alternative case influence statistics a. Alternative to D i : dffits i (and others) b. Alternative to studres i : externally-studentized residual c. Suggestion: use whatever is convenient with the

More information

Using SPSS for One Way Analysis of Variance

Using SPSS for One Way Analysis of Variance Using SPSS for One Way Analysis of Variance This tutorial will show you how to use SPSS version 12 to perform a one-way, between- subjects analysis of variance and related post-hoc tests. This tutorial

More information

ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS

ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS Ravinder Malhotra and Vipul Sharma National Dairy Research Institute, Karnal-132001 The most common use of statistics in dairy science is testing

More information

An inferential procedure to use sample data to understand a population Procedures

An inferential procedure to use sample data to understand a population Procedures Hypothesis Test An inferential procedure to use sample data to understand a population Procedures Hypotheses, the alpha value, the critical region (z-scores), statistics, conclusion Two types of errors

More information

Analysis of Variance (ANOVA)

Analysis of Variance (ANOVA) Analysis of Variance (ANOVA) Used for comparing or more means an extension of the t test Independent Variable (factor) = categorical (qualita5ve) predictor should have at least levels, but can have many

More information

N J SS W /df W N - 1

N J SS W /df W N - 1 One-Way ANOVA Source Table ANOVA MODEL: ij = µ* + α j + ε ij H 0 : µ = µ =... = µ j or H 0 : Σα j = 0 Source Sum of Squares df Mean Squares F J Between Groups nj( j * ) J - SS B /(J ) MS B /MS W = ( N

More information

y response variable x 1, x 2,, x k -- a set of explanatory variables

y response variable x 1, x 2,, x k -- a set of explanatory variables 11. Multiple Regression and Correlation y response variable x 1, x 2,, x k -- a set of explanatory variables In this chapter, all variables are assumed to be quantitative. Chapters 12-14 show how to incorporate

More information

One-Way ANOVA Source Table J - 1 SS B / J - 1 MS B /MS W. Pairwise Post-Hoc Comparisons of Means

One-Way ANOVA Source Table J - 1 SS B / J - 1 MS B /MS W. Pairwise Post-Hoc Comparisons of Means One-Way ANOVA Source Table ANOVA MODEL: ij = µ* + α j + ε ij H 0 : µ 1 = µ =... = µ j or H 0 : Σα j = 0 Source Sum of Squares df Mean Squares F Between Groups n j ( j - * ) J - 1 SS B / J - 1 MS B /MS

More information

What Does the F-Ratio Tell Us?

What Does the F-Ratio Tell Us? Planned Comparisons What Does the F-Ratio Tell Us? The F-ratio (called an omnibus or overall F) provides a test of whether or not there a treatment effects in an experiment A significant F-ratio suggests

More information

Linear Combinations. Comparison of treatment means. Bruce A Craig. Department of Statistics Purdue University. STAT 514 Topic 6 1

Linear Combinations. Comparison of treatment means. Bruce A Craig. Department of Statistics Purdue University. STAT 514 Topic 6 1 Linear Combinations Comparison of treatment means Bruce A Craig Department of Statistics Purdue University STAT 514 Topic 6 1 Linear Combinations of Means y ij = µ + τ i + ǫ ij = µ i + ǫ ij Often study

More information

Introduction to the Analysis of Variance (ANOVA)

Introduction to the Analysis of Variance (ANOVA) Introduction to the Analysis of Variance (ANOVA) The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique for testing for differences between the means of multiple (more

More information

Finding Limits Graphically and Numerically

Finding Limits Graphically and Numerically Finding Limits Graphically and Numerically 1. Welcome to finding limits graphically and numerically. My name is Tuesday Johnson and I m a lecturer at the University of Texas El Paso. 2. With each lecture

More information

Introduction to Analysis of Variance. Chapter 11

Introduction to Analysis of Variance. Chapter 11 Introduction to Analysis of Variance Chapter 11 Review t-tests Single-sample t-test Independent samples t-test Related or paired-samples t-test s m M t ) ( 1 1 ) ( m m s M M t M D D D s M t n s s M 1 )

More information

Comparing Several Means: ANOVA

Comparing Several Means: ANOVA Comparing Several Means: ANOVA Understand the basic principles of ANOVA Why it is done? What it tells us? Theory of one way independent ANOVA Following up an ANOVA: Planned contrasts/comparisons Choosing

More information

2 Prediction and Analysis of Variance

2 Prediction and Analysis of Variance 2 Prediction and Analysis of Variance Reading: Chapters and 2 of Kennedy A Guide to Econometrics Achen, Christopher H. Interpreting and Using Regression (London: Sage, 982). Chapter 4 of Andy Field, Discovering

More information

ANOVA Analysis of Variance

ANOVA Analysis of Variance ANOVA Analysis of Variance ANOVA Analysis of Variance Extends independent samples t test ANOVA Analysis of Variance Extends independent samples t test Compares the means of groups of independent observations

More information

UNIT 3 CONCEPT OF DISPERSION

UNIT 3 CONCEPT OF DISPERSION UNIT 3 CONCEPT OF DISPERSION Structure 3.0 Introduction 3.1 Objectives 3.2 Concept of Dispersion 3.2.1 Functions of Dispersion 3.2.2 Measures of Dispersion 3.2.3 Meaning of Dispersion 3.2.4 Absolute Dispersion

More information

ANOVA continued. Chapter 10

ANOVA continued. Chapter 10 ANOVA continued Chapter 10 Zettergren (003) School adjustment in adolescence for previously rejected, average, and popular children. Effect of peer reputation on academic performance and school adjustment

More information

Chapter 6: Linear combinations and multiple comparisons

Chapter 6: Linear combinations and multiple comparisons Chapter 6: Linear combinations and multiple comparisons The Centers for Disease Control estimates that the prevalence of HIV infection among adults and adolescents to be 447.8 per 100,000 (.448%) in the

More information

Contrasts (in general)

Contrasts (in general) 10/1/015 6-09/749 Experimental Design for Behavioral and Social Sciences Contrasts (in general) Context: An ANOVA rejects the overall null hypothesis that all k means of some factor are not equal, i.e.,

More information

8. TRANSFORMING TOOL #1 (the Addition Property of Equality)

8. TRANSFORMING TOOL #1 (the Addition Property of Equality) 8 TRANSFORMING TOOL #1 (the Addition Property of Equality) sentences that look different, but always have the same truth values What can you DO to a sentence that will make it LOOK different, but not change

More information

PSYC 331 STATISTICS FOR PSYCHOLOGISTS

PSYC 331 STATISTICS FOR PSYCHOLOGISTS PSYC 331 STATISTICS FOR PSYCHOLOGISTS Session 4 A PARAMETRIC STATISTICAL TEST FOR MORE THAN TWO POPULATIONS Lecturer: Dr. Paul Narh Doku, Dept of Psychology, UG Contact Information: pndoku@ug.edu.gh College

More information

Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay. Lecture - 15 Momentum Energy Four Vector

Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay. Lecture - 15 Momentum Energy Four Vector Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay Lecture - 15 Momentum Energy Four Vector We had started discussing the concept of four vectors.

More information

Hypothesis testing, part 2. With some material from Howard Seltman, Blase Ur, Bilge Mutlu, Vibha Sazawal

Hypothesis testing, part 2. With some material from Howard Seltman, Blase Ur, Bilge Mutlu, Vibha Sazawal Hypothesis testing, part 2 With some material from Howard Seltman, Blase Ur, Bilge Mutlu, Vibha Sazawal 1 CATEGORICAL IV, NUMERIC DV 2 Independent samples, one IV # Conditions Normal/Parametric Non-parametric

More information

The entire data set consists of n = 32 widgets, 8 of which were made from each of q = 4 different materials.

The entire data set consists of n = 32 widgets, 8 of which were made from each of q = 4 different materials. One-Way ANOVA Summary The One-Way ANOVA procedure is designed to construct a statistical model describing the impact of a single categorical factor X on a dependent variable Y. Tests are run to determine

More information

In ANOVA the response variable is numerical and the explanatory variables are categorical.

In ANOVA the response variable is numerical and the explanatory variables are categorical. 1 ANOVA ANOVA means ANalysis Of VAriance. The ANOVA is a tool for studying the influence of one or more qualitative variables on the mean of a numerical variable in a population. In ANOVA the response

More information

The Pennsylvania State University The Graduate School Eberly College of Science GENERALIZED STEPWISE PROCEDURES FOR

The Pennsylvania State University The Graduate School Eberly College of Science GENERALIZED STEPWISE PROCEDURES FOR The Pennsylvania State University The Graduate School Eberly College of Science GENERALIZED STEPWISE PROCEDURES FOR CONTROLLING THE FALSE DISCOVERY RATE A Dissertation in Statistics by Scott Roths c 2011

More information

ANOVA continued. Chapter 11

ANOVA continued. Chapter 11 ANOVA continued Chapter 11 Zettergren (003) School adjustment in adolescence for previously rejected, average, and popular children. Effect of peer reputation on academic performance and school adjustment

More information

Volume vs. Diameter. Teacher Lab Discussion. Overview. Picture, Data Table, and Graph

Volume vs. Diameter. Teacher Lab Discussion. Overview. Picture, Data Table, and Graph 5 6 7 Middle olume Length/olume vs. Diameter, Investigation page 1 of olume vs. Diameter Teacher Lab Discussion Overview Figure 1 In this experiment we investigate the relationship between the diameter

More information

26:010:557 / 26:620:557 Social Science Research Methods

26:010:557 / 26:620:557 Social Science Research Methods 26:010:557 / 26:620:557 Social Science Research Methods Dr. Peter R. Gillett Associate Professor Department of Accounting & Information Systems Rutgers Business School Newark & New Brunswick 1 Overview

More information

ANOVA continued. Chapter 10

ANOVA continued. Chapter 10 ANOVA continued Chapter 10 Zettergren (003) School adjustment in adolescence for previously rejected, average, and popular children. Effect of peer reputation on academic performance and school adjustment

More information

MORE ON SIMPLE REGRESSION: OVERVIEW

MORE ON SIMPLE REGRESSION: OVERVIEW FI=NOT0106 NOTICE. Unless otherwise indicated, all materials on this page and linked pages at the blue.temple.edu address and at the astro.temple.edu address are the sole property of Ralph B. Taylor and

More information

ANOVA: Analysis of Variance

ANOVA: Analysis of Variance ANOVA: Analysis of Variance Marc H. Mehlman marcmehlman@yahoo.com University of New Haven The analysis of variance is (not a mathematical theorem but) a simple method of arranging arithmetical facts so

More information

arxiv: v1 [stat.me] 14 Jan 2019

arxiv: v1 [stat.me] 14 Jan 2019 arxiv:1901.04443v1 [stat.me] 14 Jan 2019 An Approach to Statistical Process Control that is New, Nonparametric, Simple, and Powerful W.J. Conover, Texas Tech University, Lubbock, Texas V. G. Tercero-Gómez,Tecnológico

More information

Comparing Several Means

Comparing Several Means Comparing Several Means Some slides from R. Pruim STA303/STA1002: Methods of Data Analysis II, Summer 2016 Michael Guerzhoy The Dating World of Swordtail Fish In some species of swordtail fish, males develop

More information

Probability Year 9. Terminology

Probability Year 9. Terminology Probability Year 9 Terminology Probability measures the chance something happens. Formally, we say it measures how likely is the outcome of an event. We write P(result) as a shorthand. An event is some

More information

Confidence Intervals. Confidence interval for sample mean. Confidence interval for sample mean. Confidence interval for sample mean

Confidence Intervals. Confidence interval for sample mean. Confidence interval for sample mean. Confidence interval for sample mean Confidence Intervals Confidence interval for sample mean The CLT tells us: as the sample size n increases, the sample mean is approximately Normal with mean and standard deviation Thus, we have a standard

More information

Multiple Testing of General Contrasts: Truncated Closure and the Extended Shaffer-Royen Method

Multiple Testing of General Contrasts: Truncated Closure and the Extended Shaffer-Royen Method Multiple Testing of General Contrasts: Truncated Closure and the Extended Shaffer-Royen Method Peter H. Westfall, Texas Tech University Randall D. Tobias, SAS Institute Pairwise Comparisons ANOVA, g =10groups,

More information

Examining Multiple Comparison Procedures According to Error Rate, Power Type and False Discovery Rate

Examining Multiple Comparison Procedures According to Error Rate, Power Type and False Discovery Rate Journal of Modern Applied Statistical Methods Volume 11 Issue 2 Article 7 11-1-2012 Examining Multiple Comparison Procedures According to Error Rate, Power Type and False Discovery Rate Guven Ozkaya Uludag

More information

Numbers. The aim of this lesson is to enable you to: describe and use the number system. use positive and negative numbers

Numbers. The aim of this lesson is to enable you to: describe and use the number system. use positive and negative numbers Module One: Lesson One Aims The aim of this lesson is to enable you to: describe and use the number system use positive and negative numbers work with squares and square roots use the sign rule master

More information

GLM Repeated-measures designs: One within-subjects factor

GLM Repeated-measures designs: One within-subjects factor GLM Repeated-measures designs: One within-subjects factor Reading: SPSS dvanced Models 9.0: 2. Repeated Measures Homework: Sums of Squares for Within-Subject Effects Download: glm_withn1.sav (Download

More information

Sequences and infinite series

Sequences and infinite series Sequences and infinite series D. DeTurck University of Pennsylvania March 29, 208 D. DeTurck Math 04 002 208A: Sequence and series / 54 Sequences The lists of numbers you generate using a numerical method

More information

Pre-Algebra Notes Unit Three: Multi-Step Equations and Inequalities (optional)

Pre-Algebra Notes Unit Three: Multi-Step Equations and Inequalities (optional) Pre-Algebra Notes Unit Three: Multi-Step Equations and Inequalities (optional) CCSD Teachers note: CCSD syllabus objectives (2.8)The student will solve multi-step inequalities and (2.9)The student will

More information

Error Correcting Codes Prof. Dr. P. Vijay Kumar Department of Electrical Communication Engineering Indian Institute of Science, Bangalore

Error Correcting Codes Prof. Dr. P. Vijay Kumar Department of Electrical Communication Engineering Indian Institute of Science, Bangalore (Refer Slide Time: 00:15) Error Correcting Codes Prof. Dr. P. Vijay Kumar Department of Electrical Communication Engineering Indian Institute of Science, Bangalore Lecture No. # 03 Mathematical Preliminaries:

More information

Design of Engineering Experiments Part 2 Basic Statistical Concepts Simple comparative experiments

Design of Engineering Experiments Part 2 Basic Statistical Concepts Simple comparative experiments Design of Engineering Experiments Part 2 Basic Statistical Concepts Simple comparative experiments The hypothesis testing framework The two-sample t-test Checking assumptions, validity Comparing more that

More information

MITOCW ocw f99-lec05_300k

MITOCW ocw f99-lec05_300k MITOCW ocw-18.06-f99-lec05_300k This is lecture five in linear algebra. And, it will complete this chapter of the book. So the last section of this chapter is two point seven that talks about permutations,

More information

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Lecture No. # 36 Sampling Distribution and Parameter Estimation

More information