. Example: For 3 factors, sse = (y ijkt. " y ijk

Similar documents
TWO OR MORE RANDOM EFFECTS. The two-way complete model for two random effects:

Chap The McGraw-Hill Companies, Inc. All rights reserved.

Inference for the Regression Coefficient

SIZE = Vehicle size: 1 small, 2 medium, 3 large. SIDE : 1 right side of car, 2 left side of car

Lec 5: Factorial Experiment

Confidence Interval for the mean response

ST505/S697R: Fall Homework 2 Solution.

Factorial designs. Experiments

Chapter 5 Introduction to Factorial Designs Solutions

Analysis of Variance and Design of Experiments-I

1 Use of indicator random variables. (Chapter 8)

NESTED FACTORS (Chapter 18)

Factorial and Unbalanced Analysis of Variance

Power & Sample Size Calculation

Reference: Chapter 13 of Montgomery (8e)

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

Fractional Factorial Designs

Multiple comparisons - subsequent inferences for two-way ANOVA

Contents. TAMS38 - Lecture 6 Factorial design, Latin Square Design. Lecturer: Zhenxia Liu. Factorial design 3. Complete three factor design 4

Inference for Regression

THREE-WAY ANOVA MODELS (CHAPTER 7) Example: Pollution noise data ( ltersdat.

Chapter 10. Design of Experiments and Analysis of Variance

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

MS&E 226: Small Data

Notes for Week 13 Analysis of Variance (ANOVA) continued WEEK 13 page 1

Inferences for Regression

Lecture 18: Simple Linear Regression

PROBLEM TWO (ALKALOID CONCENTRATIONS IN TEA) 1. Statistical Design

Answer Keys to Homework#10

MEMORIAL UNIVERSITY OF NEWFOUNDLAND DEPARTMENT OF MATHEMATICS AND STATISTICS FINAL EXAM - STATISTICS FALL 1999

Assignment 9 Answer Keys

ANOVA: Analysis of Variation

Formal Statement of Simple Linear Regression Model

One-Way Analysis of Variance (ANOVA)

T-test: means of Spock's judge versus all other judges 1 12:10 Wednesday, January 5, judge1 N Mean Std Dev Std Err Minimum Maximum

1 Introduction to One-way ANOVA

Multiple Regression Examples

Contents. TAMS38 - Lecture 8 2 k p fractional factorial design. Lecturer: Zhenxia Liu. Example 0 - continued 4. Example 0 - Glazing ceramic 3

Multiple Comparisons. The Interaction Effects of more than two factors in an analysis of variance experiment. Submitted by: Anna Pashley

Orthogonal contrasts for a 2x2 factorial design Example p130

Simple Linear Regression. Material from Devore s book (Ed 8), and Cengagebrain.com

Chapter 7, continued: MANOVA

Ch 2: Simple Linear Regression

Section 9.4. Notation. Requirements. Definition. Inferences About Two Means (Matched Pairs) Examples

Factorial Treatment Structure: Part I. Lukas Meier, Seminar für Statistik

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants.

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College

Unit 7: Random Effects, Subsampling, Nested and Crossed Factor Designs

Outline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013

Two-Way Analysis of Variance - no interaction

2.830 Homework #6. April 2, 2009

Random and Mixed Effects Models - Part III

Inference for Regression Simple Linear Regression

Topic 29: Three-Way ANOVA

The simple linear regression model discussed in Chapter 13 was written as

y = µj n + β 1 b β b b b + α 1 t α a t a + e

STAT 525 Fall Final exam. Tuesday December 14, 2010

Design of Engineering Experiments Part 2 Basic Statistical Concepts Simple comparative experiments

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

Lec 1: An Introduction to ANOVA

Key Features: More than one type of experimental unit and more than one randomization.

STA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

9 One-Way Analysis of Variance

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

Residual Analysis for two-way ANOVA The twoway model with K replicates, including interaction,

Analysis of Variance

Inference for Regression Inference about the Regression Model and Using the Regression Line

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA

1 Statistical inference for a population mean

Chapter 13 Experiments with Random Factors Solutions

Econ 3790: Business and Economic Statistics. Instructor: Yogesh Uppal

Sleep data, two drugs Ch13.xls

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

W&M CSCI 688: Design of Experiments Homework 2. Megan Rose Bryant

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018

Multiple Regression Methods

What Is ANOVA? Comparing Groups. One-way ANOVA. One way ANOVA (the F ratio test)

Lecture 6 Multiple Linear Regression, cont.

General Linear Models (GLM) for Fixed Factors

Variance Decomposition and Goodness of Fit

Lecture 10 Multiple Linear Regression

Chapter 14 Simple Linear Regression (A)

Lecture 9: Factorial Design Montgomery: chapter 5

Allow the investigation of the effects of a number of variables on some response

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Econ 3790: Business and Economics Statistics. Instructor: Yogesh Uppal

Correlation Analysis

Stat 217 Final Exam. Name: May 1, 2002

The Random Effects Model Introduction

STAT22200 Spring 2014 Chapter 13B

Increasing precision by partitioning the error sum of squares: Blocking: SSE (CRD) à SSB + SSE (RCBD) Contrasts: SST à (t 1) orthogonal contrasts

Simple linear regression

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

3.4. A computer ANOVA output is shown below. Fill in the blanks. You may give bounds on the P-value.

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company

Factorial ANOVA. Psychology 3256

INFERENCE FOR REGRESSION

Multiple Linear Regression

Hypothesis Testing hypothesis testing approach

Transcription:

ANALYSIS OF BALANCED FACTORIAL DESIGNS Estimates of model parameters and contrasts can be obtained by the method of Least Squares. Additional constraints must be added to estimate non-estimable parameters. Example: The cell means are estimable with Least Squares estimates y ijk. From the cell-means model and one-way analysis of variance, we have the following: The error sum of squares sse is the sum of squared deviations from the fits y ijk. #### Example: For 3 factors, sse = (y ijkt " y ijk ) i j sse has associated degrees of freedom n v Example: For 3 factors, sse has abcr abc = abc(r) degrees of freedom. The mean square error mse is sse divided by its degrees of freedom. Example: For 3 factors, mse = sse/ abc(r) degrees of freedom. mse is an estimate for the variance σ. The standard error of the residuals y ijkt " y ijk is sse/(n) Before doing inference, we should use the fits and residuals to check the model assumptions of independent, normal errors with constant variance. Example: Evaporation experiment The experimenters wanted to study the evaporation rate of water under different conditions. They varied three factors: Factor Levels A: Container : cup : plate B: Salt concentration : tsp salt per cc water : 3 tsp salt per cc water C: Location : windowsill : floor in closet k t After 48 hours in the location, the amount of water remaining in the container was measured. The response was amount evaporated (= cc amount remaining). Two observations were taken for each treatment combination. Look at the data thinking about estimating standard deviations of errors for treatment combinations. What do you notice that needs thought? Use the cell-means model to obtain residuals and test for equal variance. What requires thought? Compare standard deviations of residuals between levels of each factor:

Factor Level Level Estimated variance ratio A.58.37.5 B.673.55 4.5 C.43.443. Should we proceed to further model checks or not? Simulation: Ten samples of eight each from a normal distribution with mean, standard deviation.4. Standard deviations of the samples:.458.33.386.4.36.386.4.3.39.594 These would give an estimated ratio of max to min variance of 6.3. So we will proceed, but with some caution. First, model checking plots: Standardized residuals versus levels of factors:. A.. B.. C. Does anything look problematical? If so, discuss.

3 Versus fits: FITS 5 Any problems? Normal probability plot: Normal Probability Plot.999.99.95.8.5..5.. - -.5 stres.5 Average: Std Dev: N of data: 6 Anderson-Darling Normality Test A-Squared:.34 p-value:.45 Any problems? Hypothesis Tests and Analysis of Variance Table: Submodels can be tested against larger models by F-tests, with F-statistic obtained as a ratio of mean squares. The error sum of squares of a model is the sum of the squared deviations from the fits from that model. Example: For three factors, the error sum of squares for a model is ####(y ijkt " y ˆ ), where y ˆ is the fit for the model. i j k t Special case I: Since the fit for the complete model is y ijk, the error sum of squares for the complete model is sse mentioned above. Special case II: The total sum of squares is the error sum of squares for the model y ijkt = µ + ε ijkt

4 Since the fit for this model is the overall mean, the total sum of squares can be described as the sum of the squared deviations from the overall mean. Example: For 3 factors, sstot = (y ijkt " y ) #### i j The total sum of squares has n associated degrees of freedom, where n = total number of observations (e.g., abcr for 3 factors). The sum of squares for a main effect or interaction is the difference between the error sum of squares for the model excluding terms involving this effect or interaction from the full model and the full model. Example: For 3 factors, ssbc = sse -sse a, where sse is the error sum of squares for the model excluding terms for BC and ABC interactions and sse a = sse is the error sum of squares for the complete 3-way model. The degrees of freedom for a sum of squares for a main effect or interaction is the product of one less than the number of levels for each factor involved in the effect or interaction. Examples: Sum of squares for Degrees of freedom A a AB (a)(b) BC (b)(c) ABC (a)(b)(c) The total sum of squares partitions into (is the sum of) the error sum of squares plus the sums of squares for all main effects and interactions in the model. The degrees of freedom add correspondingly. For example, for the complete 3-way model, this is just the algebraic identity (a)(b)(c) + (a)(b) + (a)(c ) + (b)(c) + (a) + (b) + (c)+ abcr abc = abcr Note: If we are dealing with another model, the error sum of squares for that model will have a different number of degrees of freedom. The degrees of freedom for error are usually most easily found by subtracting the other degrees of freedom from the total degrees of freedom. The mean square for a main effect or interaction is the sum of squares for that effect/interaction divided by its degrees of freedom. Example: msbc = ssbc/(b)(c) The F-statistic for a main effect or interaction is the mean square for the effect/interaction divided by the mean square error, with corresponding degrees of freedom in the numerator in denominator. Example: In the 3-way complete model, the null hypothesis for BC interaction is k t

5 H : [(βγ) ij - (βγ) iq ] - [(βγ) sj - (βγ) sq ] = for all i s, j q and an analogous condition on the αβγ terms The F-statistic for this test is F = msbc/mse, which has an F((b)(c), abc(r)) distribution if H is true. Example: Evaporation data. Check out degrees of freedom, etc. Conclusions? (Take into account overall significance level.) Contrasts Contrasts are defined and estimated as in the cell means model. There are various labels that are given to certain types of contrasts; see p. 99 for details. Contrasts can also be defined relative to submodels, in which case parameters not present in the submodel are omitted from the contrast. The various methods for simultaneous confidence intervals still apply. There is a modification that can be used for finding simultaneous confidence intervals for contrasts in the levels of a single factor: Replace v by the number of levels of the factor in question, and replace r by the number of observations on each level of the factor of interest. (See p. 5 for an example.) Example: Suppose that in the evaporation experiment, the researchers were interested in comparing the levels of salt for all combinations of levels of the other two factors. Thus they were considering the 4 contrasts τ ik - τ ik. (What does a positive contrast mean in terms of the context of the experiment?)if they wanted simultaneous 98% confidence intervals, they might consider the following methods:. The Tukey method for all contrasts, giving msd.69 (from Minitab output). The Bonferroni method, which would give individual 99.5% confidence intervals. The " msd for this method is t(6 8,.5) mse + % $ ' = 3.835.38 =.94. # & " 3. The Scheffe method, giving msd (8 ")F 8",6"8,. mse + % $ ' = # & [7(4.897)] /.38 = 3.35. Bonferroni is clearly the best in this situation. The estimates and resulting simultaneous 98% confidence intervals are: i,k Estimate for τ ik - τ ik CI 7.75-7.5 =.5 (.654,.54) 7.75-8.5 = -.5 (.44,.44) 7.75-6.5 =.5 (-.65, 3.54) 9.5 6. = 3.5 (.346, 5.54)

Thus, at an overall 98% confidence level, if the model assumptions indeed fit, we see that the higher concentration solution evaporates at a rate that was statistically significantly lower only when the container was the saucer (high surface area) and the location was the closet (low air movement). 6