The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization.

Similar documents
Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

DESAIN EKSPERIMEN Analysis of Variances (ANOVA) Semester Genap 2017/2018 Jurusan Teknik Industri Universitas Brawijaya

The One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

Chapter 10. Design of Experiments and Analysis of Variance

Notes for Week 13 Analysis of Variance (ANOVA) continued WEEK 13 page 1

One-Way Analysis of Variance. With regression, we related two quantitative, typically continuous variables.

Design & Analysis of Experiments 7E 2009 Montgomery

Lec 1: An Introduction to ANOVA

Lecture 13 Extra Sums of Squares

20.0 Experimental Design

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)

Multiple comparisons - subsequent inferences for two-way ANOVA

Regression With a Categorical Independent Variable: Mean Comparisons

Econ 3790: Business and Economic Statistics. Instructor: Yogesh Uppal

Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is

EXST Regression Techniques Page 1. We can also test the hypothesis H :" œ 0 versus H :"

Analysis of Variance

Much of the material we will be covering for a while has to do with designing an experimental study that concerns some phenomenon of interest.

Factorial designs. Experiments

Difference in two or more average scores in different groups

Lecture 3: Inference in SLR

Sociology 6Z03 Review II

4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants.

ANOVA Analysis of Variance

16.3 One-Way ANOVA: The Procedure

1 Introduction to One-way ANOVA

ANALYTICAL COMPARISONS AMONG TREATMENT MEANS (CHAPTER 4)

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

Lecture 5: Comparing Treatment Means Montgomery: Section 3-5

One-way ANOVA. Experimental Design. One-way ANOVA

COMPARING SEVERAL MEANS: ANOVA

CH.9 Tests of Hypotheses for a Single Sample

Analysis Of Variance Compiled by T.O. Antwi-Asare, U.G

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

Chapter 10: Analysis of variance (ANOVA)

OHSU OGI Class ECE-580-DOE :Design of Experiments Steve Brainerd

What If There Are More Than. Two Factor Levels?

Introduction to the Analysis of Variance (ANOVA)

Chap The McGraw-Hill Companies, Inc. All rights reserved.

STAT 115:Experimental Designs

Orthogonal, Planned and Unplanned Comparisons

Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /1/2016 1/46

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018

What Is ANOVA? Comparing Groups. One-way ANOVA. One way ANOVA (the F ratio test)

TWO OR MORE RANDOM EFFECTS. The two-way complete model for two random effects:

Lecture 10 Multiple Linear Regression


One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups

PSYC 331 STATISTICS FOR PSYCHOLOGISTS

Lecture 5: Hypothesis tests for more than one sample

Variance Estimates and the F Ratio. ERSH 8310 Lecture 3 September 2, 2009

3. Design Experiments and Variance Analysis

Introduction. Chapter 8

An Old Research Question

Chapter Seven: Multi-Sample Methods 1/52

One-way between-subjects ANOVA. Comparing three or more independent means

ANOVA CIVL 7012/8012

STAT Chapter 10: Analysis of Variance

Linear Combinations of Group Means

Orthogonal and Non-orthogonal Polynomial Constrasts

Comparing Several Means: ANOVA

Analysis of Variance (ANOVA)

Incomplete Block Designs

Lecture 7: Hypothesis Testing and ANOVA

These are all actually contrasts (the coef sum to zero). What are these contrasts representing? What would make them large?

Contrasts and Multiple Comparisons Supplement for Pages

Chapter 11 - Lecture 1 Single Factor ANOVA

Chapter 11 - Lecture 1 Single Factor ANOVA

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

Analysis of Variance

Unit 27 One-Way Analysis of Variance

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

Topic 17 - Single Factor Analysis of Variance. Outline. One-way ANOVA. The Data / Notation. One way ANOVA Cell means model Factor effects model

Stats fest Analysis of variance. Single factor ANOVA. Aims. Single factor ANOVA. Data

21.0 Two-Factor Designs

Battery Life. Factory

Analytical Comparisons Among Treatment Means (Chapter 4) Analysis of Trend (Chapter 5) ERSH 8310 Fall 2009

Power & Sample Size Calculation

RCB - Example. STA305 week 10 1

Introduction to Analysis of Variance (ANOVA) Part 2

Analysis of Variance. Read Chapter 14 and Sections to review one-way ANOVA.

Unit 12: Analysis of Single Factor Experiments

Using SPSS for One Way Analysis of Variance

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

Topic 6. Two-way designs: Randomized Complete Block Design [ST&D Chapter 9 sections 9.1 to 9.7 (except 9.6) and section 15.8]

One-Way Analysis of Variance (ANOVA)

One-Way Analysis of Variance (ANOVA) Paul K. Strode, Ph.D.

Inferences for Regression

Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests

COMPLETELY RANDOM DESIGN (CRD) -Design can be used when experimental units are essentially homogeneous.

W&M CSCI 688: Design of Experiments Homework 2. Megan Rose Bryant

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

Tukey Complete Pairwise Post-Hoc Comparison

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

Variance Decomposition and Goodness of Fit

Statistics for Managers Using Microsoft Excel Chapter 10 ANOVA and Other C-Sample Tests With Numerical Data

Design of Engineering Experiments Part 2 Basic Statistical Concepts Simple comparative experiments

Formal Statement of Simple Linear Regression Model

Transcription:

1 Chapter 1: Research Design Principles The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization. 2 Chapter 2: Completely Randomized Design 2.1 The cell means model The first model we consider for a set of treatments is the cell means model: y ij = µ i + e ij, i = 1, 2,..., t; j = 1, 2,..., r This is considered a full model (representing an alternative hypothesis such as H a : µ i µ k for some i k), and to detect treatment effects we can compare it to a reduced model (representing the null hypothesis H 0 : µ 1 = µ 2 = = µ t ) in which the mean is the same for all groups: y ij = µ + e ij, i = 1, 2,..., t; j = 1, 2,..., r Both of these models (as well as the treatment effects model to be introduced later) are special cases of the general linear model, as discussed on pages 45-47. Least squares Estimation of model parameters. 2.2 The treatment effects model An alternative model can be used by considering the group means under H a and their differences from the overall mean µ. = t µ i /t : which can be rewritten as: y ij = µ. + (µ i µ. ) + (y ij µ i ), y ij = µ + τ i + e ij The hypothesis of no group effect for this (full) model is H 0 : τ 1 = τ 2 = = τ t = 0, and the reduced model is the same as above for the cell means model. 2.3 Analysis of Variance The key idea behind an analysis of variance involves a decomposition of the total sum of squares. For a one-way ANOVA (possibly arising from a completely randomized design) the decomposition is SS Total = SS Treatment + SS Error. If y ij is the jth observation in group i, then y ij y.. = (y ij y i. ) + (y i. y.. ), which leads to: r r r (y ij y.. ) 2 = (y ij y i. ) 2 + (y i. y.. ) 2, or SS Total = SS Error + SS Treatment. j=1 j=1 j=1 1

As an example, consider three groups with the following data: Group 1 has y 1j values of 1, 2, and 3, Group 2 has y 2j values of 5, 3, and 4, and Group 3 has y 3j values of 6, 7, and 5. The overall sample mean is y.. = ( t r j=1 y ij)/tr = 36/9 = 4. Then SS Total is j=1 r (y ij y.. ) 2 = (1 4) 2 + (2 4) 2 +... + (5 4) 2 = 30. The group means are y 1. = 2, y 2. = 4, and y 3. = 6, so SS Error and SS Treatment are r SS Error = (y ij y i. ) 2 = (1 2) 2 + (2 2) 2 + (3 2) 2 + (5 4) 2 +... + (5 6) 2 = 6, and j=1 SS Treatment = j=1 r (y i. y.. ) 2 = r(y i. y.. ) 2 = 3(2 4) 2 + 3(4 4) 2 + 3(6 4) 2 = 24. Thus SS Total = SS Error +SS Treatment or 30 = 6 + 24 partitions the total sum of squares about the overall mean into two parts, one within groups (due to error, or effects not accounted by the model) and one between groups (measuring the difference between sample means). Since each group here has r observations, each group contributes r 1 degrees of freedom for the within group sum of squares, for a total of t(r 1) degrees of freedom for SS Error. SS Treatment is calculating the sum of squares of t sample means about their (overall) mean, so it has t 1 degrees of freedom. For the example data above, t(r 1) = 3(2) = 6, and t 1 = 3-1 = 2. We can summarize this information in an analysis of variance table: Source SS df MS F Treatments 24 2 12 12 Error 6 6 1 Total sum of squares 30 8 As seen in the ANOVA table above, we divide sums of squares by their degrees of freedom to obtain mean squares for both treatments and error. The F statistic is then the mean square for treatments (MST) divided by the mean square for error (MSE). To test the null hypothesis H 0 : µ 1 = µ 2 = µ 3 against the alternative hypothesis H a : µ i µ k for some i k, we compare the F statistic to an F distribution with numerator df = t 1 = 3-1 = 2, and denominator df = t(r 1) = 3(2) = 6. When group sample sizes are unequal we replace r in the expressions by r i, which is the sample size in the ith group. As is shown at the end of Chapter 2, E(MSE) = σ 2 and E(MST ) = σ 2 + rθt 2 where τ θt 2 2 = i (t 1), so when the null hypothesis is true, E(MSE) = E(MST ) and the F statistic value should be close to 1. 2.4 A general principle for GLM tests As noted in the text, a general way to conduct tests for linear models (and even more generally) is to obtain full and reduced models as illustrated above. An F test is obtained by measuring the reduction in sum of 2

squared error in the two models compared to the sum of squared error for the full model, with both terms divided by their degrees of freedom to create mean squares: F = (SSE r SSE f )/(t 1) SSE f /(N t) Under the standard ANOVA assumptions, this F statistics follows an F distribution with t 1 and N t degrees of freedom. Calculation of standard errors for group means are also discussed in the text. 2.5 Power and sample size for completely randomized designs Power and sample size analyses are important tools for assessing the ability of a statistical test to detect when a null hypothesis is false, and for deciding what sample size is required for having a reasonable chance to reject a false null hypothesis. Recall that for a test of a statistical hypothesis, the Type I error (α) is the probability of rejecting the null hypothesis when it is true. The Type II error (β ) is the probability of not rejecting the null hypothesis when it is false. The power of the test equals 1 β, and is the probability of rejecting the null hypothesis when it is false. The power will depend on the alternative hypothesis, and we would like to have high power to detect alternative hypotheses of interest. For the completely randomized design with one-way treatment structure, when the null hypothesis is true, the F statistic has an F distribution with t 1 and t(r 1) degrees of freedom. When the null hypothesis is false, the F statistic follows a non-central F distribution with t 1 and t(r 1) degrees of freedom, and noncentrality parameter: λ = r τ 2 i σ 2. The power of the F test is a monotonically increasing function of the parameter λ. Notice that when the null hypothesis is true, λ = 0, so that the usual (or central) F distribution is just a special case of the non-central F distribution. It should make sense intuitively that the power of the F test increases as r increases, as τ 2 i increases, and as σ 2 decreases, as predicted by the noncentrality parameter λ. 3 Chapter 3: Treatment Comparisons Usually we have more specific research questions than just rejecting the ANOVA null hypothesis that all groups have the same mean (H 0 : µ 1 = µ 2 =... = µ t ). Common questions: 1) While planning an experiment, you wish to test certain hypotheses that are a subset of the global ANOVA H 0. (A priori tests) 2) After a significant ANOVA, which groups differ? (Post hoc tests) For example, the text discusses several questions that arise after finding that the ANOVA global H 0 : µ 1 = µ 2 = µ 3 = µ 4 is rejected with the meat storage data. 3

3.1 Terminology When describing treatment comparisons, it is useful to introduce some terminology. The term contrast is used to describe a comparison of means. Specifically, a contrast is a linear combination of the population means, C = k i µ i that also satisfies k i = 0. We require that the coeffi cients k i sum to zero so that the comparison is meaningful (we would not be interested in µ 1 3µ 2 for example). For example, if we had 6 treatment groups and we wished to test whether the the first two group means differed, we can express the null hypothesis as H 0 : 1µ 1 1µ 2 = µ 1 µ 2 = 0. Here k 1 =1 and k 2 = -1, so k 1 + k 2 = 0 as required. This is an example of a pairwise contrast, which is defined as a contrast involving only two groups. An example of a non-pairwise contrast would be if we wished to test if the mean of the first two groups differed from the mean of the other four groups. We can express this null hypothesis as H 0 : (µ 1 + µ 2 )/2 (µ 3 + µ 4 + µ 5 + µ 6 )/4 = 0. Here k 1 =1/2, k 2 = 1/2, k 3 = -1/4, k 4 = -1/4, k 5 = -1/4 and k 6 = -1/4, so again k 1 + k 2 + k 3 + k 4 + k 5 + k 6 = 0, as required by the definition of a contrast. One other property of a set of contrasts, called orthogonality, is useful when considering a priori tests. Two contrasts C 1 = k 1i µ i and C 2 = k 2i µ i are orthogonal if k 1i k 2i n i = 0. If the group sample sizes are equal then this is equivalent to k 1i k 2i = 0. In the examples above, if we identify the first contrast as C 1 and the second contrast as C 2, then k 11 =1, k 12 = -1, k 13 = k 14 = k 15 = k 16 = 0, are the coeffi cients for C 1 and k 21 =1/2, k 22 = 1/2, k 23 = -1/4, k 24 = -1/4, k 25 = -1/4 and k 26 = -1/4 are the coeffi cients for C 2. Then if the sample sizes are equal, k 1i k 2i = (1)(1/2) + ( 1)(1/2) + (0)( 1/4) + (0)( 1/4) + (0)( 1/4) + (0)( 1/4) = 0, so C 1 and C 2 are orthogonal. Orthogonal contrasts are statistically independent, so that the outcome of testing one contrast is independent of the outcome of testing the other contrast. A set of more than two contrasts is mutually orthogonal if each pair of contrasts in the set is orthogonal to each other. The concept of a contrast or a set of contrasts at first seems somewhat esoteric, but in fact it is essential to understand these concepts to fully understand ANOVA, particularly in complicated situations. The contrasts that you use should depend on the questions of scientific interest from your research. 4 Inference for contrasts We can estimate the contrast with the observed sample contrast: C = k i µ i 4

c = k i y i. The text presents formulas for the variance of a sample contrast that can be used to obtain a t test for the null hypothesis H 0 : C = 0. 4.1 Orthogonal contrasts can form a partition of SS Treatment As noted in the text, we can form as many orthogonal contrasts as we have degrees of freedom for the treatment factor. These orthogonal contrasts partition the SS Treatment into SSC i terms that allow us to separate the total sum of squares for treatments into parts attributable to different contrasts. Especially when you have equal sample sizes, this can be a powerful tool for understanding treatment effects. One common example of this is when we use orthogonal polynomial contrasts to partition quantitative treatment effects into parts attributable to linear trend, quadratic trend, and higher-order trends. For equally-spaced dosage levels with equal-sample-size groups, the coeffi cients for orthogonal polynomial contrasts are given in Table XI on page 623 of our text. The SAS code for this lecture illustrates the use of orthogonal contrasts for the meat storage data, and also the use of the ORPOL function to obtain orthogonal polynomial coeffi cients for quantitative treatment levels. The code also shows the use of Proc Reg for obtaining predictions from response curves and standard errors of the predictions. 5