Multiple Comparisons
|
|
- Louise Dawson
- 6 years ago
- Views:
Transcription
1 Multiple Comparisons Error Rates, A Priori Tests, and Post-Hoc Tests Multiple Comparisons: A Rationale Multiple comparison tests function to tease apart differences between the groups within our IV when we conduct ANOVAs Our ANOVA provides a single F statistic assessing the suitability of H o NOT how groups differ Omnibus test Rejecting H 0 simply tells us that: H 0 :µ 1 = µ 2 = µ 3 Is not an accurate representation of the data 1
2 Multiple Comparisons: A Rationale Any of the following could be true: H 1 :µ 1 µ 2 µ 3 H 1 :µ 1 µ 2 = µ 3 H 1 :µ 1 µ 3 = µ 2 H 1 :µ 1 = µ 2 µ 3 H 1 :µ 1 = µ 3 µ 2 All are potentially valid null hypotheses Which is most accurate? Multiple Comparisons: Example Going back to our AN treatment example: Which treatment is best for AN? H 0 :µ = µ = µ H 1 :µ µ µ H 1 :µ µ = µ H 1 :µ µ = µ H 1 :µ = µ µ H 1 :µ = µ µ 2
3 Multiple Comparisons: Example We could simply compare differences between groups by conducting a series of individual groups t-tests vs. vs. vs. From these three analyses, we can answer which of the alternate hypotheses best fits the data Why don t we do it this way? Multiple Comparisons: The Limitation Too much work (3 more analyses!) Inflation of Type I error rate 3
4 Error Associated With Multiple Comparisons When we want to consider error rates for analyses, there are two ways to calculate error: Per-comparison Error Associated With Multiple Comparisons The probability associated with making an error for each analysis Could be Type I or Type II but generally we think only about Type I error Per-Comparison (PC) error rate = α Thus, for every analysis we conduct, there is a fixed amount of error we have to live with 4
5 Familywise Error Associated With Multiple Comparisons The probability associated with making an error for a set of comparisons Familywise error rate = 1 (1 α) c α = α c = the number of comparisons made Error Associated With Multiple Comparisons Thus, for our AN treatment example: α = 1 (1.05) 3 α = 1 (.95) 3 α = 1 (.8574) α =.14 Thus, we see that the familywise error rate is slightly less than the sum of the α values for all 3 analyses 5
6 Error Associated With Multiple Comparisons When conducting multiple comparisons, we run the risk of inflating the error in our analysis The more analyses, the greater the chance of making a Type I error 3 t-tests =.14 Multiple Comparisons: Conclusion Most multiple comparison procedures seek to minimize or eliminate the impact of familywise error This is the reason we use multiple comparison tests to evaluate group differences in ANOVA 6
7 Types of Multiple Comparison Tests A Priori Tests based on hypotheses you have BEFORE collecting and analyzing data Driven by your theory Planned without seeing the results of the analyses However, because of this theoretical grounding, a priori tests must be a subset of all possible comparisons A Priori Tests: FAQ Why can t I conduct a priori tests after collecting data? Seeing the data (group means) may bias the hypotheses you generate Why can t I explore all possible comparisons when conducting a priori tests? Exploring all possible comparisons is the process of conducting post hoc tests Mining the data 7
8 A Priori Tests: FAQ Can I conduct a priori tests without an ANOVA? Yes! If you can narrow down your hypotheses to begin with, it is possible to conduct a priori tests without an F test however, you should have a good reason for doing so. Types of Multiple Comparison Tests Post Hoc Tests planned after the data is collected and the experimenter has examined the group means These tests are subject to experimenter bias and may be influenced by expectancy effects 8
9 Post-hoc Tests: FAQ SPSS isn t printing out post-hoc results for me. It s giving me an error. Remember: multiple comparison tests require at least 3 groups. Otherwise, you can simply interpret the means. You likely have a grouping variable with 2 categories. Which MC is best? Given the grounding of a priori tests in theory, they are generally preferred to post hoc tests More difficult to plan and conduct Unexpected results? 9
10 A Priori Tests: Multiple Comparison t-tests The easiest a priori test is the multiple comparison t-test This test will NOT control for familywise error rate important to choose comparisons carefully. A Priori Tests: Multiple Comparison t-tests For homogeneous variances: t = MS n x Error x MS + n 1 2 Error Note: Variance is calculated through MS Error 10
11 A Priori Tests: Multiple Comparison t-tests SPSS Output: Descriptive Statistics Descriptive Statistics Treatment Group Change in weight at Post-Intervention Valid N (listwise) Change in weight at Post-Intervention Valid N (listwise) Change in weight at Post-Intervention Valid N (listwise) Std. N Minimum Maximum Mean Deviation Variance Skewness Kurtosis Statistic Statistic Statistic Statistic Statistic Statistic Statistic Statistic A Priori Tests: Multiple Comparison t-tests SPSS Output: t-tests Independent Samples Test vs. Levene's Test for Equality of Variances t-test for Equality of Means Change in weight at Post-Intervention Equal variances assumed Equal variances not assumed F Sig. t df Sig. (2-tailed) Mean Difference Std. Error Difference vs. Levene's Test for Equality of Variances Independent Samples Test t-test for Equality of Means Change in weight at Post-Intervention Equal variances assumed Equal variances not assumed Std. Error F Sig. t df Sig. (2-tailed) Mean Difference Difference
12 A Priori Tests: Multiple Comparison t-tests SPSS Output: t-tests Independent Samples Test vs. Levene's Test for Equality of Variances t-test for Equality of Means Change in weight at Post-Intervention Equal variances assumed Equal variances not assumed Std. Error F Sig. t df Sig. (2-tailed) Mean Difference Difference Obviously, we would NOT conduct all 3 of these analyses for a priori comparisons this example is for educational purposes only A Priori Tests: Bonferroni Correction A way of correcting the α level to correct for the number of analyses Divide the α used in the analysis by the number of comparisons conducted Thus, the available Type I error rate is split amongst the comparisons Test is more conservative α = α / c α =.05 / 3 α =
13 A Priori Tests: Bonferroni Correction Where, oh where, is the α =.017 table? It doesn t exist Bonferroni correction to α is useful ONLY when you can calculate the exact probability associated with a given finding Computerized statistical packages like SPSS Most commonly used correction for familywise error A Priori Tests: Multiple Comparison t-tests SPSS Output: t-tests α =.017 Independent Samples Test vs. Levene's Test for Equality of Variances t-test for Equality of Means Change in weight at Post-Intervention Equal variances assumed Equal variances not assumed Std. Error F Sig. t df Sig. (2-tailed) Mean Difference Difference vs. Levene's Test for Equality of Variances Independent Samples Test t-test for Equality of Means Change in weight at Post-Intervention Equal variances assumed Equal variances not assumed Std. Error F Sig. t df Sig. (2-tailed) Mean Difference Difference
14 A Priori Tests: Multiple Comparison t-tests SPSS Output: t-tests α =.017 vs. Levene's Test for Equality of Variances Independent Samples Test t-test for Equality of Means Change in weight at Post-Intervention Equal variances assumed Equal variances not assumed F Sig. t df Sig. (2-tailed) Mean Difference Std. Error Difference Obviously, we would NOT conduct all 3 of these analyses for a priori comparisons this example is for educational purposes only Before Post-Hoc Tests One-Way ANOVA results ANOVA Change in weight at Post-Intervention Sum of Squares df Mean Square F Sig. Between Groups Within Groups Total
15 Post Hoc Tests: Fisher s Least Significant Difference Test Also known as Fisher s LSD (no, not that one ) Same as a priori t-tests we explored earlier HOWEVER: Requires a significant F test When H 1 completely true: α = α H 1 :µ 1 µ 3 µ 2 If H 1 NOT completely true: α α H 1 :µ 1 = µ 3 µ 2 Since we know that H 1 is not always completely true, AVOID FISHER S LSD Post Hoc Tests: Fisher s Least Significant Difference Test Dependent Variable: Change in weight at Post-Intervention Multiple Comparisons LSD (I) Treatment Group (J) Treatment Group 95% Confidence Interval Mean Difference (I-J) Std. Error Sig. Lower Bound Upper Bound * * * * * * *. The mean difference is significant at the.05 level. 15
16 Post Hoc Tests: The Studentized Range Statistic A statistic reflecting the difference between the largest and smallest means The studentized range statistic, because we look only at the largest and smallest means, systematically underestimates the Type I error rate However, the studentized range statistic is a step towards calculating other post hoc tests Post Hoc Tests: The Studentized Range Statistic First, rank order the means ().32 () 3.75 () Second, plug into equation q = x l arg est smallest MS n Error l x MS + n 2 Error s 16
17 Post Hoc Tests: The Studentized Range Statistic A more useful form of the studentized range equation allows us to calculate the minimum difference between means that would be statistically significant MSError + (, ) nl xl xs = q.05 groups dferror 2 MS n Error s Post Hoc Tests: Tukey Highly Significant Difference Test Derived from the Studentized range statistic Conservatively controls for the number of steps between comparison groups One of the most common Post-hoc tests in use 17
18 Post Hoc Tests: Tukey s Highly Significant Difference Test Dependent Variable: Change in weight at Post-Intervention Multiple Comparisons Tukey HSD (I) Treatment Group (J) Treatment Group 95% Confidence Interval Mean Difference (I-J) Std. Error Sig. Lower Bound Upper Bound * * * * * * *. The mean difference is significant at the.05 level. Post Hoc Tests: Scheffé Test Another derivation of the Studentized range statistic and Tukey s HSD One of the most conservative post-hoc tests in use 18
19 Post Hoc Tests: Scheffé Test Dependent Variable: Change in weight at Post-Intervention Multiple Comparisons (I) Treatment Group (J) Treatment Group 95% Confidence Interval Mean Difference (I-J) Std. Error Sig. Lower Bound Upper Bound Scheffe * * * * * * *. The mean difference is significant at the.05 level. Dependent Variable: Change in weight at Post-Intervention Multiple Comparisons Tukey HSD Scheffe LSD B f i (I) Treatment Group C t l (J) Treatment Group C t l 95% Confidence Interval Mean Difference (I-J) Std. Error Sig. Lower Bound Upper Bound * * * * * * * * * * * * * * * * * *
20 Post-Hoc Test Comparisons p LSD < p Tukey < p Scheffé Scheffé is most conservative LSD is least conservative (& potentially wrong) Which post-hoc test should you use? Purpose of the analysis Effect size 20
H0: Tested by k-grp ANOVA
Analyses of K-Group Designs : Omnibus F, Pairwise Comparisons & Trend Analyses ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures
More informationH0: Tested by k-grp ANOVA
Pairwise Comparisons ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures Alpha estimation reconsidered H0: Tested by k-grp ANOVA Regardless
More informationMultiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions
Introduction to Analysis of Variance 1 Experiments with More than 2 Conditions Often the research that psychologists perform has more conditions than just the control and experimental conditions You might
More informationA posteriori multiple comparison tests
A posteriori multiple comparison tests 11/15/16 1 Recall the Lakes experiment Source of variation SS DF MS F P Lakes 58.000 2 29.400 8.243 0.006 Error 42.800 12 3.567 Total 101.600 14 The ANOVA tells us
More informationMultiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600
Multiple Comparison Procedures Cohen Chapter 13 For EDUC/PSY 6600 1 We have to go to the deductions and the inferences, said Lestrade, winking at me. I find it hard enough to tackle facts, Holmes, without
More informationCOMPARING SEVERAL MEANS: ANOVA
LAST UPDATED: November 15, 2012 COMPARING SEVERAL MEANS: ANOVA Objectives 2 Basic principles of ANOVA Equations underlying one-way ANOVA Doing a one-way ANOVA in R Following up an ANOVA: Planned contrasts/comparisons
More informationSEVERAL μs AND MEDIANS: MORE ISSUES. Business Statistics
SEVERAL μs AND MEDIANS: MORE ISSUES Business Statistics CONTENTS Post-hoc analysis ANOVA for 2 groups The equal variances assumption The Kruskal-Wallis test Old exam question Further study POST-HOC ANALYSIS
More informationWhat Does the F-Ratio Tell Us?
Planned Comparisons What Does the F-Ratio Tell Us? The F-ratio (called an omnibus or overall F) provides a test of whether or not there a treatment effects in an experiment A significant F-ratio suggests
More informationHypothesis testing: Steps
Review for Exam 2 Hypothesis testing: Steps Repeated-Measures ANOVA 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region
More information13: Additional ANOVA Topics. Post hoc Comparisons
13: Additional ANOVA Topics Post hoc Comparisons ANOVA Assumptions Assessing Group Variances When Distributional Assumptions are Severely Violated Post hoc Comparisons In the prior chapter we used ANOVA
More informationOne-Way ANOVA Source Table J - 1 SS B / J - 1 MS B /MS W. Pairwise Post-Hoc Comparisons of Means
One-Way ANOVA Source Table ANOVA MODEL: ij = µ* + α j + ε ij H 0 : µ 1 = µ =... = µ j or H 0 : Σα j = 0 Source Sum of Squares df Mean Squares F Between Groups n j ( j - * ) J - 1 SS B / J - 1 MS B /MS
More informationIntroduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs
Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique
More informationMANOVA is an extension of the univariate ANOVA as it involves more than one Dependent Variable (DV). The following are assumptions for using MANOVA:
MULTIVARIATE ANALYSIS OF VARIANCE MANOVA is an extension of the univariate ANOVA as it involves more than one Dependent Variable (DV). The following are assumptions for using MANOVA: 1. Cell sizes : o
More informationComparing Several Means: ANOVA
Comparing Several Means: ANOVA Understand the basic principles of ANOVA Why it is done? What it tells us? Theory of one way independent ANOVA Following up an ANOVA: Planned contrasts/comparisons Choosing
More informationB. Weaver (18-Oct-2006) MC Procedures Chapter 1: Multiple Comparison Procedures ) C (1.1)
B. Weaver (18-Oct-2006) MC Procedures... 1 Chapter 1: Multiple Comparison Procedures 1.1 Introduction The omnibus F-test in a one-way ANOVA is a test of the null hypothesis that the population means of
More informationDegrees of freedom df=1. Limitations OR in SPSS LIM: Knowing σ and µ is unlikely in large
Z Test Comparing a group mean to a hypothesis T test (about 1 mean) T test (about 2 means) Comparing mean to sample mean. Similar means = will have same response to treatment Two unknown means are different
More informationContrasts (in general)
10/1/015 6-09/749 Experimental Design for Behavioral and Social Sciences Contrasts (in general) Context: An ANOVA rejects the overall null hypothesis that all k means of some factor are not equal, i.e.,
More informationChapter 13 Section D. F versus Q: Different Approaches to Controlling Type I Errors with Multiple Comparisons
Explaining Psychological Statistics (2 nd Ed.) by Barry H. Cohen Chapter 13 Section D F versus Q: Different Approaches to Controlling Type I Errors with Multiple Comparisons In section B of this chapter,
More informationHypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -
Hypothesis Testing with One-Way ANOVA Statistics Arlo Clark-Foos Conceptual Refresher 1. Standardized z distribution of scores and of means can be represented as percentile rankings. 2. t distribution
More informationHypothesis testing: Steps
Review for Exam 2 Hypothesis testing: Steps Exam 2 Review 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region 3. Compute
More informationIntroduction to Analysis of Variance (ANOVA) Part 2
Introduction to Analysis of Variance (ANOVA) Part 2 Single factor Serpulid recruitment and biofilms Effect of biofilm type on number of recruiting serpulid worms in Port Phillip Bay Response variable:
More informationContrasts and Multiple Comparisons Supplement for Pages
Contrasts and Multiple Comparisons Supplement for Pages 302-323 Brian Habing University of South Carolina Last Updated: July 20, 2001 The F-test from the ANOVA table allows us to test the null hypothesis
More informationAnalysis of variance
Analysis of variance 1 Method If the null hypothesis is true, then the populations are the same: they are normal, and they have the same mean and the same variance. We will estimate the numerical value
More informationIntroduction to the Analysis of Variance (ANOVA)
Introduction to the Analysis of Variance (ANOVA) The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique for testing for differences between the means of multiple (more
More informationChapter 6 Planned Contrasts and Post-hoc Tests for one-way ANOVA
Chapter 6 Planned Contrasts and Post-hoc Tests for one-way NOV Page. The Problem of Multiple Comparisons 6-. Types of Type Error Rates 6-. Planned contrasts vs. Post hoc Contrasts 6-7 4. Planned Contrasts
More informationN J SS W /df W N - 1
One-Way ANOVA Source Table ANOVA MODEL: ij = µ* + α j + ε ij H 0 : µ = µ =... = µ j or H 0 : Σα j = 0 Source Sum of Squares df Mean Squares F J Between Groups nj( j * ) J - SS B /(J ) MS B /MS W = ( N
More informationThe entire data set consists of n = 32 widgets, 8 of which were made from each of q = 4 different materials.
One-Way ANOVA Summary The One-Way ANOVA procedure is designed to construct a statistical model describing the impact of a single categorical factor X on a dependent variable Y. Tests are run to determine
More informationIndependent Samples ANOVA
Independent Samples ANOVA In this example students were randomly assigned to one of three mnemonics (techniques for improving memory) rehearsal (the control group; simply repeat the words), visual imagery
More informationAnalysis of Variance (ANOVA)
Analysis of Variance (ANOVA) Two types of ANOVA tests: Independent measures and Repeated measures Comparing 2 means: X 1 = 20 t - test X 2 = 30 How can we Compare 3 means?: X 1 = 20 X 2 = 30 X 3 = 35 ANOVA
More information4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES
4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES FOR SINGLE FACTOR BETWEEN-S DESIGNS Planned or A Priori Comparisons We previously showed various ways to test all possible pairwise comparisons for
More informationAn Analysis of College Algebra Exam Scores December 14, James D Jones Math Section 01
An Analysis of College Algebra Exam s December, 000 James D Jones Math - Section 0 An Analysis of College Algebra Exam s Introduction Students often complain about a test being too difficult. Are there
More informationOne-way between-subjects ANOVA. Comparing three or more independent means
One-way between-subjects ANOVA Comparing three or more independent means Data files SpiderBG.sav Attractiveness.sav Homework: sourcesofself-esteem.sav ANOVA: A Framework Understand the basic principles
More informationSTAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5)
STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons Ch. 4-5) Recall CRD means and effects models: Y ij = µ i + ϵ ij = µ + α i + ϵ ij i = 1,..., g ; j = 1,..., n ; ϵ ij s iid N0, σ 2 ) If we reject
More informationOne-way between-subjects ANOVA. Comparing three or more independent means
One-way between-subjects ANOVA Comparing three or more independent means ANOVA: A Framework Understand the basic principles of ANOVA Why it is done? What it tells us? Theory of one-way between-subjects
More informationANOVA Analysis of Variance
ANOVA Analysis of Variance ANOVA Analysis of Variance Extends independent samples t test ANOVA Analysis of Variance Extends independent samples t test Compares the means of groups of independent observations
More informationTopic 1. Definitions
S Topic. Definitions. Scalar A scalar is a number. 2. Vector A vector is a column of numbers. 3. Linear combination A scalar times a vector plus a scalar times a vector, plus a scalar times a vector...
More informationKeppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means
Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means 4.1 The Need for Analytical Comparisons...the between-groups sum of squares averages the differences
More informationANOVA continued. Chapter 11
ANOVA continued Chapter 11 Zettergren (003) School adjustment in adolescence for previously rejected, average, and popular children. Effect of peer reputation on academic performance and school adjustment
More informationANOVA continued. Chapter 10
ANOVA continued Chapter 10 Zettergren (003) School adjustment in adolescence for previously rejected, average, and popular children. Effect of peer reputation on academic performance and school adjustment
More informationThe One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)
The One-Way Repeated-Measures ANOVA (For Within-Subjects Designs) Logic of the Repeated-Measures ANOVA The repeated-measures ANOVA extends the analysis of variance to research situations using repeated-measures
More informationBIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES
BIOL 458 - Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES PART 1: INTRODUCTION TO ANOVA Purpose of ANOVA Analysis of Variance (ANOVA) is an extremely useful statistical method
More informationThe One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)
The One-Way Independent-Samples ANOVA (For Between-Subjects Designs) Computations for the ANOVA In computing the terms required for the F-statistic, we won t explicitly compute any sample variances or
More informationDifference in two or more average scores in different groups
ANOVAs Analysis of Variance (ANOVA) Difference in two or more average scores in different groups Each participant tested once Same outcome tested in each group Simplest is one-way ANOVA (one variable as
More informationANOVA continued. Chapter 10
ANOVA continued Chapter 10 Zettergren (003) School adjustment in adolescence for previously rejected, average, and popular children. Effect of peer reputation on academic performance and school adjustment
More informationParametric Analysis of Variance
Parametric Analysis of Variance The same principles used in the two-sample means test can be applied to more than two sample means. In these cases we call the test analysis of variance. The null hypothesis
More informationAn Old Research Question
ANOVA An Old Research Question The impact of TV on high-school grade Watch or not watch Two groups The impact of TV hours on high-school grade Exactly how much TV watching would make difference Multiple
More information2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018
Math 403 - P. & S. III - Dr. McLoughlin - 1 2018 2 Hand-out 2 Dr. M. P. M. M. M c Loughlin Revised 2018 3. Fundamentals 3.1. Preliminaries. Suppose we can produce a random sample of weights of 10 year-olds
More informationLinear Combinations. Comparison of treatment means. Bruce A Craig. Department of Statistics Purdue University. STAT 514 Topic 6 1
Linear Combinations Comparison of treatment means Bruce A Craig Department of Statistics Purdue University STAT 514 Topic 6 1 Linear Combinations of Means y ij = µ + τ i + ǫ ij = µ i + ǫ ij Often study
More informationUsing SPSS for One Way Analysis of Variance
Using SPSS for One Way Analysis of Variance This tutorial will show you how to use SPSS version 12 to perform a one-way, between- subjects analysis of variance and related post-hoc tests. This tutorial
More informationLaboratory Topics 4 & 5
PLS205 Lab 3 January 23, 2014 Orthogonal contrasts Class comparisons in SAS Trend analysis in SAS Multiple mean comparisons Laboratory Topics 4 & 5 Orthogonal contrasts Planned, single degree-of-freedom
More informationSTAT22200 Spring 2014 Chapter 5
STAT22200 Spring 2014 Chapter 5 Yibi Huang April 29, 2014 Chapter 5 Multiple Comparisons Chapter 5-1 Chapter 5 Multiple Comparisons Note the t-tests and C.I. s are constructed assuming we only do one test,
More informationMultiple Testing. Gary W. Oehlert. January 28, School of Statistics University of Minnesota
Multiple Testing Gary W. Oehlert School of Statistics University of Minnesota January 28, 2016 Background Suppose that you had a 20-sided die. Nineteen of the sides are labeled 0 and one of the sides is
More informationSPSS Guide For MMI 409
SPSS Guide For MMI 409 by John Wong March 2012 Preface Hopefully, this document can provide some guidance to MMI 409 students on how to use SPSS to solve many of the problems covered in the D Agostino
More informationMultiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model
Biostatistics 250 ANOVA Multiple Comparisons 1 ORIGIN 1 Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model When the omnibus F-Test for ANOVA rejects the null hypothesis that
More informationComparisons among means (or, the analysis of factor effects)
Comparisons among means (or, the analysis of factor effects) In carrying out our usual test that μ 1 = = μ r, we might be content to just reject this omnibus hypothesis but typically more is required:
More informationLec 1: An Introduction to ANOVA
Ying Li Stockholm University October 31, 2011 Three end-aisle displays Which is the best? Design of the Experiment Identify the stores of the similar size and type. The displays are randomly assigned to
More information13: Additional ANOVA Topics
13: Additional ANOVA Topics Post hoc comparisons Least squared difference The multiple comparisons problem Bonferroni ANOVA assumptions Assessing equal variance When assumptions are severely violated Kruskal-Wallis
More informationDETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics
DETAILED CONTENTS About the Author Preface to the Instructor To the Student How to Use SPSS With This Book PART I INTRODUCTION AND DESCRIPTIVE STATISTICS 1. Introduction to Statistics 1.1 Descriptive and
More informationM A N O V A. Multivariate ANOVA. Data
M A N O V A Multivariate ANOVA V. Čekanavičius, G. Murauskas 1 Data k groups; Each respondent has m measurements; Observations are from the multivariate normal distribution. No outliers. Covariance matrices
More informationAnalysis of Variance: Repeated measures
Repeated-Measures ANOVA: Analysis of Variance: Repeated measures Each subject participates in all conditions in the experiment (which is why it is called repeated measures). A repeated-measures ANOVA is
More informationAnalysis of Variance (ANOVA)
Analysis of Variance (ANOVA) Used for comparing or more means an extension of the t test Independent Variable (factor) = categorical (qualita5ve) predictor should have at least levels, but can have many
More informationChap The McGraw-Hill Companies, Inc. All rights reserved.
11 pter11 Chap Analysis of Variance Overview of ANOVA Multiple Comparisons Tests for Homogeneity of Variances Two-Factor ANOVA Without Replication General Linear Model Experimental Design: An Overview
More informationOutline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013
Topic 19 - Inference - Fall 2013 Outline Inference for Means Differences in cell means Contrasts Multiplicity Topic 19 2 The Cell Means Model Expressed numerically Y ij = µ i + ε ij where µ i is the theoretical
More informationANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS
ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS Ravinder Malhotra and Vipul Sharma National Dairy Research Institute, Karnal-132001 The most common use of statistics in dairy science is testing
More informationLinear Combinations of Group Means
Linear Combinations of Group Means Look at the handicap example on p. 150 of the text. proc means data=mth567.disability; class handicap; var score; proc sort data=mth567.disability; by handicap; proc
More informationHypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =
Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,
More informationFactorial Independent Samples ANOVA
Factorial Independent Samples ANOVA Liljenquist, Zhong and Galinsky (2010) found that people were more charitable when they were in a clean smelling room than in a neutral smelling room. Based on that
More informationPLSC PRACTICE TEST ONE
PLSC 724 - PRACTICE TEST ONE 1. Discuss briefly the relationship between the shape of the normal curve and the variance. 2. What is the relationship between a statistic and a parameter? 3. How is the α
More informationYour schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table
Your schedule of coming weeks One-way ANOVA, II 9.07 //00 Today: One-way ANOVA, part II Next week: Two-way ANOVA, parts I and II. One-way ANOVA HW due Thursday Week of May Teacher out of town all week
More informationSampling Distributions: Central Limit Theorem
Review for Exam 2 Sampling Distributions: Central Limit Theorem Conceptually, we can break up the theorem into three parts: 1. The mean (µ M ) of a population of sample means (M) is equal to the mean (µ)
More informationGroup comparison test for independent samples
Group comparison test for independent samples The purpose of the Analysis of Variance (ANOVA) is to test for significant differences between means. Supposing that: samples come from normal populations
More informationSpecific Differences. Lukas Meier, Seminar für Statistik
Specific Differences Lukas Meier, Seminar für Statistik Problem with Global F-test Problem: Global F-test (aka omnibus F-test) is very unspecific. Typically: Want a more precise answer (or have a more
More informationSelf-Assessment Weeks 6 and 7: Multiple Regression with a Qualitative Predictor; Multiple Comparisons
Self-Assessment Weeks 6 and 7: Multiple Regression with a Qualitative Predictor; Multiple Comparisons 1. Suppose we wish to assess the impact of five treatments on an outcome Y. How would these five treatments
More informationIntroduction. Chapter 8
Chapter 8 Introduction In general, a researcher wants to compare one treatment against another. The analysis of variance (ANOVA) is a general test for comparing treatment means. When the null hypothesis
More informationPreview from Notesale.co.uk Page 3 of 63
Stem-and-leaf diagram - vertical numbers on far left represent the 10s, numbers right of the line represent the 1s The mean should not be used if there are extreme scores, or for ranks and categories Unbiased
More informationTwo-Way ANOVA. Chapter 15
Two-Way ANOVA Chapter 15 Interaction Defined An interaction is present when the effects of one IV depend upon a second IV Interaction effect : The effect of each IV across the levels of the other IV When
More informationPSY 216. Assignment 12 Answers. Explain why the F-ratio is expected to be near 1.00 when the null hypothesis is true.
PSY 21 Assignment 12 Answers 1. Problem 1 from the text Explain why the F-ratio is expected to be near 1.00 when the null hypothesis is true. When H0 is true, the treatment had no systematic effect. In
More informationhttp://www.statsoft.it/out.php?loc=http://www.statsoft.com/textbook/ Group comparison test for independent samples The purpose of the Analysis of Variance (ANOVA) is to test for significant differences
More informationIn ANOVA the response variable is numerical and the explanatory variables are categorical.
1 ANOVA ANOVA means ANalysis Of VAriance. The ANOVA is a tool for studying the influence of one or more qualitative variables on the mean of a numerical variable in a population. In ANOVA the response
More informationWorkshop Research Methods and Statistical Analysis
Workshop Research Methods and Statistical Analysis Session 2 Data Analysis Sandra Poeschl 08.04.2013 Page 1 Research process Research Question State of Research / Theoretical Background Design Data Collection
More informationLecture Notes #3: Contrasts and Post Hoc Tests 3-1
Lecture Notes #3: Contrasts and Post Hoc Tests 3-1 Richard Gonzalez Psych 613 Version 2.6 (9/2016) LECTURE NOTES #3: Contrasts and Post Hoc tests Reading assignment Read MD chs 4, 5, & 6 Read G chs 7 &
More informationTukey Complete Pairwise Post-Hoc Comparison
Tukey Complete Pairwise Post-Hoc Comparison Engineering Statistics II Section 10.2 Josh Engwer TTU 2018 Josh Engwer (TTU) Tukey Complete Pairwise Post-Hoc Comparison 2018 1 / 23 PART I PART I: Gosset s
More informationSelf-Assessment Weeks 8: Multiple Regression with Qualitative Predictors; Multiple Comparisons
Self-Assessment Weeks 8: Multiple Regression with Qualitative Predictors; Multiple Comparisons 1. Suppose we wish to assess the impact of five treatments while blocking for study participant race (Black,
More informationSTAT 115:Experimental Designs
STAT 115:Experimental Designs Josefina V. Almeda 2013 Multisample inference: Analysis of Variance 1 Learning Objectives 1. Describe Analysis of Variance (ANOVA) 2. Explain the Rationale of ANOVA 3. Compare
More informationHypothesis Testing. Hypothesis: conjecture, proposition or statement based on published literature, data, or a theory that may or may not be true
Hypothesis esting Hypothesis: conjecture, proposition or statement based on published literature, data, or a theory that may or may not be true Statistical Hypothesis: conjecture about a population parameter
More informationKeppel, G. & Wickens, T. D. Design and Analysis Chapter 12: Detailed Analyses of Main Effects and Simple Effects
Keppel, G. & Wickens, T. D. Design and Analysis Chapter 1: Detailed Analyses of Main Effects and Simple Effects If the interaction is significant, then less attention is paid to the two main effects, and
More information1 One-way Analysis of Variance
1 One-way Analysis of Variance Suppose that a random sample of q individuals receives treatment T i, i = 1,,... p. Let Y ij be the response from the jth individual to be treated with the ith treatment
More informationAdvanced Experimental Design
Advanced Experimental Design Topic Four Hypothesis testing (z and t tests) & Power Agenda Hypothesis testing Sampling distributions/central limit theorem z test (σ known) One sample z & Confidence intervals
More informationSPSS Output. ANOVA a b Residual Coefficients a Standardized Coefficients
SPSS Output Homework 1-1e ANOVA a Sum of Squares df Mean Square F Sig. 1 Regression 351.056 1 351.056 11.295.002 b Residual 932.412 30 31.080 Total 1283.469 31 a. Dependent Variable: Sexual Harassment
More information9 One-Way Analysis of Variance
9 One-Way Analysis of Variance SW Chapter 11 - all sections except 6. The one-way analysis of variance (ANOVA) is a generalization of the two sample t test to k 2 groups. Assume that the populations of
More informationBasic Statistical Analysis
indexerrt.qxd 8/21/2002 9:47 AM Page 1 Corrected index pages for Sprinthall Basic Statistical Analysis Seventh Edition indexerrt.qxd 8/21/2002 9:47 AM Page 656 Index Abscissa, 24 AB-STAT, vii ADD-OR rule,
More informationCheat Sheet: ANOVA. Scenario. Power analysis. Plotting a line plot and a box plot. Pre-testing assumptions
Cheat Sheet: ANOVA Measurement and Evaluation of HCC Systems Scenario Use ANOVA if you want to test the difference in continuous outcome variable vary between multiple levels (A, B, C, ) of a nominal variable
More informationIntroduction to Statistics with GraphPad Prism 7
Introduction to Statistics with GraphPad Prism 7 Outline of the course Power analysis with G*Power Basic structure of a GraphPad Prism project Analysis of qualitative data Chi-square test Analysis of quantitative
More informationAnalysis of Variance
Statistical Techniques II EXST7015 Analysis of Variance 15a_ANOVA_Introduction 1 Design The simplest model for Analysis of Variance (ANOVA) is the CRD, the Completely Randomized Design This model is also
More informationSampling Distributions
Sampling Distributions Sampling Distribution of the Mean & Hypothesis Testing Remember sampling? Sampling Part 1 of definition Selecting a subset of the population to create a sample Generally random sampling
More informationRegression Part II. One- factor ANOVA Another dummy variable coding scheme Contrasts Mul?ple comparisons Interac?ons
Regression Part II One- factor ANOVA Another dummy variable coding scheme Contrasts Mul?ple comparisons Interac?ons One- factor Analysis of variance Categorical Explanatory variable Quan?ta?ve Response
More informationChapter 13 Correlation
Chapter Correlation Page. Pearson correlation coefficient -. Inferential tests on correlation coefficients -9. Correlational assumptions -. on-parametric measures of correlation -5 5. correlational example
More informationMultiple Testing. Tim Hanson. January, Modified from originals by Gary W. Oehlert. Department of Statistics University of South Carolina
Multiple Testing Tim Hanson Department of Statistics University of South Carolina January, 2017 Modified from originals by Gary W. Oehlert Type I error A Type I error is to wrongly reject the null hypothesis
More informationANOVA. Testing more than 2 conditions
ANOVA Testing more than 2 conditions ANOVA Today s goal: Teach you about ANOVA, the test used to measure the difference between more than two conditions Outline: - Why anova? - Contrasts and post-hoc tests
More informationChapter 10. Design of Experiments and Analysis of Variance
Chapter 10 Design of Experiments and Analysis of Variance Elements of a Designed Experiment Response variable Also called the dependent variable Factors (quantitative and qualitative) Also called the independent
More information