Lecture 12: 2 k Factorial Design Montgomery: Chapter 6

Similar documents
Lecture 10: 2 k Factorial Design Montgomery: Chapter 6

Lecture 11: Blocking and Confounding in 2 k design

Lecture 12: 2 k p Fractional Factorial Design

Lecture 14: 2 k p Fractional Factorial Design

Lecture 9: Factorial Design Montgomery: chapter 5

Strategy of Experimentation II

Answer Keys to Homework#10

Assignment 9 Answer Keys

Lecture 3. Experiments with a Single Factor: ANOVA Montgomery 3.1 through 3.3

Lecture 3. Experiments with a Single Factor: ANOVA Montgomery 3-1 through 3-3

Single Factor Experiments

Lecture 4. Checking Model Adequacy

Comparison of a Population Means

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Lecture 7: Latin Square and Related Design

Statistics GIDP Ph.D. Qualifying Exam Methodology

Lecture 11: Simple Linear Regression

SAS Procedures Inference about the Line ffl model statement in proc reg has many options ffl To construct confidence intervals use alpha=, clm, cli, c

SAS Commands. General Plan. Output. Construct scatterplot / interaction plot. Run full model

The 2 k Factorial Design. Dr. Mohammad Abuhaiba 1

STOR 455 STATISTICAL METHODS I

Lecture 7: Latin Squares and Related Designs

Outline Topic 21 - Two Factor ANOVA

Topic 28: Unequal Replication in Two-Way ANOVA

19. Blocking & confounding

Two-factor studies. STAT 525 Chapter 19 and 20. Professor Olga Vitek

Topic 20: Single Factor Analysis of Variance

23. Fractional factorials - introduction

Statistics 512: Solution to Homework#11. Problems 1-3 refer to the soybean sausage dataset of Problem 20.8 (ch21pr08.dat).

Unreplicated 2 k Factorial Designs

Lecture 1 Linear Regression with One Predictor Variable.p2

Statistics GIDP Ph.D. Qualifying Exam Methodology January 10, 9:00am-1:00pm

Topic 17 - Single Factor Analysis of Variance. Outline. One-way ANOVA. The Data / Notation. One way ANOVA Cell means model Factor effects model

Suppose we needed four batches of formaldehyde, and coulddoonly4runsperbatch. Thisisthena2 4 factorial in 2 2 blocks.

Overview Scatter Plot Example

Chapter 1 Linear Regression with One Predictor

Fractional Factorial Designs

Topic 29: Three-Way ANOVA

Chapter 11. Analysis of Variance (One-Way)

One-Way Analysis of Variance (ANOVA) There are two key differences regarding the explanatory variable X.

Introduction to Design and Analysis of Experiments with the SAS System (Stat 7010 Lecture Notes)

Outline. Topic 20 - Diagnostics and Remedies. Residuals. Overview. Diagnostics Plots Residual checks Formal Tests. STAT Fall 2013

T-test: means of Spock's judge versus all other judges 1 12:10 Wednesday, January 5, judge1 N Mean Std Dev Std Err Minimum Maximum

Chapter 11: Factorial Designs

STATISTICS 479 Exam II (100 points)

Chapter 2 Inferences in Simple Linear Regression

Outline. Topic 22 - Interaction in Two Factor ANOVA. Interaction Not Significant. General Plan

Institutionen för matematik och matematisk statistik Umeå universitet November 7, Inlämningsuppgift 3. Mariam Shirdel

Analysis of Covariance

2 k, 2 k r and 2 k-p Factorial Designs

ST3232: Design and Analysis of Experiments

Fractional Factorial Designs

Design and Analysis of

6. Fractional Factorial Designs (Ch.8. Two-Level Fractional Factorial Designs)

Statistics for exp. medical researchers Regression and Correlation

Lecture 5: Comparing Treatment Means Montgomery: Section 3-5

Topic 32: Two-Way Mixed Effects Model

Advanced Digital Design with the Verilog HDL, Second Edition Michael D. Ciletti Prentice Hall, Pearson Education, 2011

Chapter 13 Experiments with Random Factors Solutions

Reference: Chapter 6 of Montgomery(8e) Maghsoodloo

EXST Regression Techniques Page 1. We can also test the hypothesis H :" œ 0 versus H :"

Lecture 4. Random Effects in Completely Randomized Design

BE640 Intermediate Biostatistics 2. Regression and Correlation. Simple Linear Regression Software: SAS. Emergency Calls to the New York Auto Club

3.4. A computer ANOVA output is shown below. Fill in the blanks. You may give bounds on the P-value.

Chapter 8 (More on Assumptions for the Simple Linear Regression)

5. Blocking and Confounding

Unit 9: Confounding and Fractional Factorial Designs

Analysis of Variance. Source DF Squares Square F Value Pr > F. Model <.0001 Error Corrected Total

STAT 3A03 Applied Regression Analysis With SAS Fall 2017

Response Surface Methodology

Design and Analysis of Multi-Factored Experiments

STA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007

Lecture 10: Experiments with Random Effects

General Linear Model (Chapter 4)

Residuals from regression on original data 1

Unit 6: Fractional Factorial Experiments at Three Levels

Chapter 6 The 2 k Factorial Design Solutions

Design of Engineering Experiments Part 5 The 2 k Factorial Design

Contents. TAMS38 - Lecture 8 2 k p fractional factorial design. Lecturer: Zhenxia Liu. Example 0 - continued 4. Example 0 - Glazing ceramic 3

unadjusted model for baseline cholesterol 22:31 Monday, April 19,

a. The least squares estimators of intercept and slope are (from JMP output): b 0 = 6.25 b 1 =

CSCI 688 Homework 6. Megan Rose Bryant Department of Mathematics William and Mary

Correlation and the Analysis of Variance Approach to Simple Linear Regression

ST Correlation and Regression

TWO-LEVEL FACTORIAL EXPERIMENTS: BLOCKING. Upper-case letters are associated with factors, or regressors of factorial effects, e.g.

data proc sort proc corr run proc reg run proc glm run proc glm run proc glm run proc reg CONMAIN CONINT run proc reg DUMMAIN DUMINT run proc reg

Lecture 2. The Simple Linear Regression Model: Matrix Approach

Topic 25 - One-Way Random Effects Models. Outline. Random Effects vs Fixed Effects. Data for One-way Random Effects Model. One-way Random effects

Unbalanced Data in Factorials Types I, II, III SS Part 1

STAT 350: Summer Semester Midterm 1: Solutions

COMPREHENSIVE WRITTEN EXAMINATION, PAPER III FRIDAY AUGUST 26, 2005, 9:00 A.M. 1:00 P.M. STATISTICS 174 QUESTION

SIMPLE LINEAR REGRESSION

STAT 3A03 Applied Regression With SAS Fall 2017

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression

Chapter 31 The GLMMOD Procedure. Chapter Table of Contents

Topic 18: Model Selection and Diagnostics

Math 3330: Solution to midterm Exam

20g g g Analyze the residuals from this experiment and comment on the model adequacy.

Statistics 512: Applied Linear Models. Topic 9

CS 5014: Research Methods in Computer Science

Transcription:

Lecture 12: 2 k Factorial Design Montgomery: Chapter 6 1 Lecture 12 Page 1

2 k Factorial Design Involvingk factors: each has two levels (often labeled+and ) Very useful design for preliminary study Can weed out unimportant factors Identify important factors and their interactions Each main/interaction effect (of any order) has ONE degree of freedom Factors need not be on numeric scale For general two-factor factorial model y ijk = µ+α i +β j +(αβ) ij +ǫ ijk There are 1+(a-1)+(b-1)+(a-1)(b-1) model parameters (excludingσ 2 ) 2 2 factorial design has only four parameters (excludingσ 2 ) α 1 = α 2 β 1 = β 2 (αβ) 11 = (αβ) 22 (αβ) 12 = (αβ) 22 (αβ) 21 = (αβ) 22 2 Lecture 12 Page 2

Example: 2 2 Factorial Design factor replicate A B treatment 1 2 3 mean (1) 28 25 27 80/3 + a 36 32 32 100/3 + b 18 19 23 60/3 + + ab 31 30 29 90/3 Reparametrization with treatment meanµ ij = µ+α i +β j +(αβ) ij Main effect of A:A = µ + µ = α + α = 2α + (zero-sum constraint) Main effect of B:B = µ + µ = β + β = 2β + (zero-sum constraint) Interaction effect: AB = {(µ ++ µ + ) (µ + µ )}/2 = {(αβ) ++ (αβ) + (αβ) + +(αβ) }/2 = 2(αβ) = 2(αβ) ++ 3 Lecture 12 Page 3

Main effect of A: Reparametrization via Constrasts A = µ + µ = (µ + +µ ++ )/2 (µ +µ + )/2 = ( µ +µ + µ + +µ ++ )/2 = C A /2 ContrastC A = µ +µ + µ + +µ ++ = ( 1,1, 1,1) ĈA = 16.67 = Â = 8.33 Main effect of B: B = µ + µ = (µ + +µ ++ )/2 (µ +µ + )/2 = ( µ µ + +µ + +µ ++ )/2 = C B /2 ContrastC B = µ µ + +µ + +µ ++ = ( 1, 1,1,1) ĈB = 10.00 = ˆB = 5.00 4 Lecture 12 Page 4

Interaction effect: AB = (µ µ + µ + +µ )/2 = C AB /2 ContrastC AB = µ µ + µ + +µ = (1, 1, 1,1) ĈAB = 3.33 = ÂB = 1.67 Notice that effects are defined in a different way than Chapter 5. But they are connected and equivalent. 5 Lecture 12 Page 5

Effects and Contrasts factor contrast of effect A B total mean I A B AB 80 80/3 1-1 -1 1 + 100 100/3 1 1-1 -1 + 60 60/3 1-1 1-1 + + 90 90/3 1 1 1 1 There is a one-to-one correspondence between effects and contrasts, and contrasts can be directly used to estimate the effects. For a effect corresponding to contrastc = (c 1,c 2,...) in2 2 design êffect = 1 2 c i ȳ i whereiis an index for treatments and the summation is over all treatments. i 6 Lecture 12 Page 6

Sum of Squares Due to Effect Because effects are defined using contrasts, their sum of squares can also be calculated through contrasts. Recall for contrastc = (c 1,c 2,...), its sum of squares is SS Contrast = ( c i ȳ i ) 2 c 2 i /n So SS A = ( ȳ +ȳ + ȳ + +ȳ ++ ) 2 4/n SS B = ( ȳ ȳ + +ȳ + +ȳ ++ ) 2 4/n SS AB = (ȳ ȳ + ȳ + +ȳ ++ ) 2 4/n = 208.33 = 75.00 = 8.33 7 Lecture 12 Page 7

Sum of Squares and ANOVA Total sum of squares: SS T = i,j,k y2 ijk y2... N Error sum of squares: SS E = SS T SS A SS B SS AB ANOVA Table Source of Sum of Degrees of Mean Variation Squares Freedom Square F 0 A SS A 1 MS A B SS B 1 MS B AB SS AB 1 MS AB Error SS E N 4 MS E Total SS T N 1 8 Lecture 12 Page 8

Using SAS data one; input A B resp; datalines; -1-1 28-1 -1 25-1 -1 27 1-1 36 1-1 32 1-1 32-1 1 18-1 1 19-1 1 23 1 1 31 1 1 30 1 1 29 ; proc glm data=one; class A B; model resp=a B; run; quit; Sum of Source DF Squares Mean Square F Value Pr > F Model 3 291.6666667 97.2222222 24.82 0.0002 Error 8 31.3333333 3.9166667 Cor Total 11 323.0000000 A 1 208.3333333 208.3333333 53.19 <.0001 B 1 75.0000000 75.0000000 19.15 0.0024 A*B 1 8.3333333 8.3333333 2.13 0.1828 9 Lecture 12 Page 9

Analyzing2 2 Experiment Using Regression Model Because every effect in2 2 design, or its sum of squares, has one degree of freedom, it can be equivalently represented by a numerical variable, and regression analysis can be directly used to analyze the data. The original factors are not necessarily continuous. Code the levels of factor A and B as follows Fit regression model A x1 B x2 - -1 - -1 + 1 + 1 y = β 0 +β 1 x 1 +β 2 x 2 +β 12 x 1 x 2 +ǫ The fitted model should be y = ȳ.. + Â 2 x 1 + ˆB 2 x 2 + ÂB 2 x 1x 2 i.e. the estimated coefficients are half of the effects, respectively. 10 Lecture 12 Page 10

SAS Code and Output data one; input x1 x2 resp; x1x2=x1*x2; datalines; -1-1 28-1 -1 25-1 -1 27... ; proc reg data=one; model resp=x1 x2 x1x2; run; quit; Analysis of Variance Sum of Mean Source DF Squares Square F Value Pr > F Model 3 291.66667 97.22222 24.82 0.0002 Error 8 31.33333 3.91667 Corrected Total 11 323.00000 Parameter Estimates Parameter Standard Variable DF Estimate Error t Value Pr > t Intercept 1 27.50000 0.57130 48.14 <.0001 x1 1 4.16667 0.57130 7.29 <.0001 x2 1-2.50000 0.57130-4.38 0.0024 x1x2 1 0.83333 0.57130 1.46 0.1828 11 Lecture 12 Page 11

Bottling Experiment: 2 3 Factorial Design factor response A B C treatment 1 2 total (1) -3-1 -4 + a 0 1 1 + b -1 0-1 + + ab 2 3 5 + c -1 0-1 + + ac 2 1 3 + + bc 1 1 2 + + + abc 6 5 11 12 Lecture 12 Page 12

Main effects A = µ + ȳ Factorial Effects and Contrasts = 1 4 ( µ + µ + µ + + µ ++ µ + + µ + + µ ++ + µ +++ ) A: contrastc A = ( 1,1, 1,1, 1,1, 1,1), = 3.00 B: contrastc B = ( 1, 1,1,1, 1, 1,1,1), ˆB = 2.25 C: contrastc C = ( 1, 1, 1, 1,1,1,1,1),Ĉ = 1.75 2-factor interactions AB: A B componentwise,âb =.75 AC: A C componentwise,âc =.25 BC: B C componentwise, BC =.50 3-factor interaction ABC = 1 (int(ab C+) int(ab C )) 2 = 1 4 ( µ + µ + + µ + µ ++ + µ + µ + + µ ++ + µ +++ ) ABC: C ABC = ( 1,1,1, 1,1, 1, 1,1) = A B C,ÂBC =.50. 13 Lecture 12 Page 13

Contrasts for Calculating Effects in2 3 Design factorial effects A B C treatment I A B AB C AC BC ABC (1) 1-1 -1 1-1 1 1-1 + a 1 1-1 -1-1 -1 1 1 + b 1-1 1-1 -1 1-1 1 + + ab 1 1 1 1-1 -1-1 -1 + c 1-1 -1 1 1-1 -1 1 + + ac 1 1-1 -1 1 1-1 -1 + + bc 1-1 1-1 1-1 1-1 + + + abc 1 1 1 1 1 1 1 1 14 Lecture 12 Page 14

Estimates grand mean : ȳi. 2 3 Contrast Sum of Squares effect : ci ȳ i. 2 3 1 SS effect = ( c i ȳ i. ) 2 2 3 /n = 2n(êffect) 2 Variance of Estimate Var(êffect) = σ2 n2 3 2 t-test for effects (confidence interval approach) êffect±t α/2,2 k (n 1)S.E.(êffect) 15 Lecture 12 Page 15

Regression Model Code the levels of factor A and B as follows Fit regression model A x1 B x2 C x3 - -1 - -1 - -1 + 1 + 1 + 1 y = β 0 +β 1 x 1 +β 2 x 2 +β 3 x 3 +β 12 x 1 x 2 +β 13 x 1 x 3 +β 23 x 2 x 3 +β 123 x 1 x 2 x 3 +ǫ The fitted model should be y = ȳ.. +Â 2 x 1+ ˆB 2 x 2+Ĉ 2 x3+âb 2 x 1x 2 +ÂC 2 x 1x 3 + BC 2 x 2x 3 +ÂBC x 1 x 2 x 3 2 That is, ˆβ = êffect, and 2 Var(ˆβ) = σ2 n2 k = σ2 n2 3 16 Lecture 12 Page 16

Bottling Experiment: Using SAS data bottle; input A B C devi; datalines; -1-1 -1-3 -1-1 -1-1 1-1 -1 0 1-1 -1 1-1 1-1 -1-1 1-1 0 1 1-1 2 1 1-1 3-1 -1 1-1 -1-1 1 0 1-1 1 2 1-1 1 1-1 1 1 1-1 1 1 1 1 1 1 6 1 1 1 5 ; proc glm data=bottle; class A B C; model devi=a B C; output out=botone r=res p=pred; run; proc univariate data=botone pctldef=4; var res; qqplot res/normal (L=1 mu=est sigma=est); histogram res/normal; run; proc gplot; plot res*pred/frame; run; data bottlenew; set bottle; x1=a; x2=b; x3=c; x1x2=x1*x2; x1x3=x1*x3; x2x3=x2*x3; x1x2x3=x1*x2*x3; drop A B C; proc reg data=bottlenew; model devi=x1 x2 x3 x1x2 x1x3 x2x3 x1x2x3; run; quit; 17 Lecture 12 Page 17

ANOVA Model: SAS Output for Bottling Experiment Dependent Variable: devi Sum of Source DF Squares Mean Square F Value Pr > F Model 7 73.00000000 10.42857143 16.69 0.0003 Error 8 5.00000000 0.62500000 CorTotal 15 78.00000000 A 1 36.00000000 36.00000000 57.60 <.0001 B 1 20.25000000 20.25000000 32.40 0.0005 A*B 1 2.25000000 2.25000000 3.60 0.0943 C 1 12.25000000 12.25000000 19.60 0.0022 A*C 1 0.25000000 0.25000000 0.40 0.5447 B*C 1 1.00000000 1.00000000 1.60 0.2415 A*B*C 1 1.00000000 1.00000000 1.60 0.2415 18 Lecture 12 Page 18

Regression Model: Parameter Standard Variable DF Estimate Error t Value Pr > t Intercept 1 1.00000 0.19764 5.06 0.0010 x1 1 1.50000 0.19764 7.59 <.0001 x2 1 1.12500 0.19764 5.69 0.0005 x3 1 0.87500 0.19764 4.43 0.0022 x1x2 1 0.37500 0.19764 1.90 0.0943 x1x3 1 0.12500 0.19764 0.63 0.5447 x2x3 1 0.25000 0.19764 1.26 0.2415 x1x2x3 1 0.25000 0.19764 1.26 0.2415 19 Lecture 12 Page 19

General2 k Design k factors: A,B,...,K each with 2 levels (+, ) consists of all possible level combinations (2 k treatments) each withnreplicates Classify factorial effects: type of effect label the number of effects main effects (of order 1) A,B,C,...,K k 2-factor interactions (of order 2) 3-factor interactions (of order 3) AB,AC,...,JK k 2 ABC,ABD,...,IJK k 3......... k-factor interaction (of orderk) ABC K k k 20 Lecture 12 Page 20

In total, how many effects? Each effect (main or interaction) has 1 degree of freedom full model (i.e. model consisting of all the effects) has2 k 1degrees of freedom. Error component has2 k (n 1) degrees of freedom (why?). One-to-one correspondence between effects and contrasts: For main effect: convert the level column of a factor using 1 and + 1 For interactions: multiply the contrasts of the main effects of the involved factors, componentwisely. 21 Lecture 12 Page 21

General2 k Design: Analysis Estimates: grand mean : ȳi. 2 k For effect with contrastc = (c 1,c 2,...,c 2 k), its estimate is êffect = ci ȳ i 2 k 1 Variance what is the standard error of the effect? Var(êffect) = σ2 n2 k 2 t-test forh 0 : effect=0. Using the confidence interval approach, êffect±t α/2,2 k (n 1) S.E.(êffect) 22 Lecture 12 Page 22

Using ANOVA model: Sum of Squares due to an effect, using its contrast, SS effect = ( c i ȳ i. ) 2 2 k /n = n2 k 2 (êffect) 2 SS T andss E can be calculated as before and a ANOVA table including SS due to the effects andss E can be constructed and the effects can be tested byf -tests. Using regression: Introducing variablesx 1,...,x k for main effects, their products are used for interactions, the following regression model can be fitted y = β 0 +β 1 x 1 +...+β k x k +...+β 12 k x 1 x 2 x k +ǫ The coefficients are estimated by half of effects they represent, that is, ˆβ = êffect 2 23 Lecture 12 Page 23

Unreplicated2 k Design: Filtration Rate Experiment factor A B C D filtration 45 + 71 + 48 + + 65 + 68 + + 60 + + 80 + + + 65 + 43 + + 100 + + 45 + + + 104 + + 75 + + + 86 + + + 70 + + + + 96 24 Lecture 12 Page 24

Unreplicated2 k Design No degree of freedom left for error component if full model is fitted. Formulas used for estimates and contrast sum of squares are given in Slides 22-23 with n=1 No error sum of squares available, cannot estimateσ 2 and test effects in both the ANOVA and Regression approaches. Approach 1: pooling high-order interactions Often assume 3 or higher interactions do not occur Pool estimates together for error Warning: may pool significant interaction 25 Lecture 12 Page 25

Unreplicated2 k Design Approach 2: Using the normal probability plot (QQ plot) to identify significant effects. Recall Var(êffect) = σ2 2 k 2 If the effect is not significant (=0), then the effect estimate follows N(0, σ 2 2 k 2 ) Assume all effects not significant, their estimates can be considered as a random sample fromn(0, σ 2 2 k 2 ) QQ plot of the estimates is expected to be a linear line ( ) Plot effect j v.s. Φ 1 rj 3/8 n+1/4 optionnormal=blom inproc RANK in SAS). wherer j is the rank of effect j (using Deviation from a linear line indicates significant effects 26 Lecture 12 Page 26

Statistics 514:2 k Factorial Design Summer 2018 Using SAS to Generate QQ Plot for Effects data filter; do D = -1 to 1 by 2; do C = -1 to 1 by 2; do B = -1 to 1 by 2; do A = -1 to 1 by 2; input y @@; output; end; end; end; end; datalines; 45 71 48 65 68 60 80 65 43 100 45 104 75 86 70 96 ; data inter; /* Define Interaction Terms */ set filter; AB=A*B; AC=A*C; AD=A*D; BC=B*C; BD=B*D; CD=C*D; ABC=AB*C; ABD=AB*D; ACD=AC*D; BCD=BC*D; ABCD=ABC*D; proc glm data=inter; /* GLM Proc to Obtain Effects */ class A B C D AB AC AD BC BD CD ABC ABD ACD BCD ABCD; model y=a B C D AB AC AD BC BD CD ABC ABD ACD BCD ABCD; estimate A A 1-1; estimate AC AC 1-1; run; 27 Lecture 12 Page 27

proc reg outest=effects data=inter; /* REG Proc to Obtain Effects */ model y=a B C D AB AC AD BC BD CD ABC ABD ACD BCD ABCD; data effect2; set effects; drop y intercept _RMSE_; proc transpose data=effect2 out=effect3; data effect4; set effect3; effect=col1*2; proc sort data=effect4; by effect; proc print data=effect4; /*Generate the QQ plot */ proc rank data=effect4 out=effect5 normal=blom; var effect; ranks neff; proc print data=effect5; symbol1 v=circle; proc gplot data=effect5; plot effect*neff=_name_; run; quit; 28 Lecture 12 Page 28

Ranked Effects Obs _NAME_ COL1 effect neff 1 AC -9.0625-18.125-1.73938 2 BCD -1.3125-2.625-1.24505 3 ACD -0.8125-1.625-0.94578 4 CD -0.5625-1.125-0.71370 5 BD -0.1875-0.375-0.51499 6 AB 0.0625 0.125-0.33489 7 ABCD 0.6875 1.375-0.16512 8 ABC 0.9375 1.875-0.00000 9 BC 1.1875 2.375 0.16512 10 B 1.5625 3.125 0.33489 11 ABD 2.0625 4.125 0.51499 12 C 4.9375 9.875 0.71370 13 D 7.3125 14.625 0.94578 14 AD 8.3125 16.625 1.24505 15 A 10.8125 21.625 1.73938 29 Lecture 12 Page 29

QQ plot 30 Lecture 12 Page 30

Filtration Experiment Analysis Fit a linear line based on small effects, identify the effects which are potentially significant, then use ANOVA or regression fit a sub-model with those effects. 1. Potentially significant effects: A, AD, C, D, AC. 2. Use main effect plot and interaction plot. 3. ANOVA model involving onlya,c,dand their interactions (projecting the original unreplicated2 4 experiment onto a replicated2 3 experiment). 4. regression model only involvinga,c,d,ac andad. 5. Diagnostics using residuals. 31 Lecture 12 Page 31

/* data step is the same */ proc sort; by A C; proc means noprint; var y; by A C; output out=ymeanac mean=mn; Interaction Plots forac symbol1 v=circle i=join; symbol2 v=square i=join; proc gplot data=ymeanac; plot mn*a=c; run; 32 Lecture 12 Page 32

/* data step is the same */ proc sort; by A D; proc means noprint; var y; by A D; output out=ymeanad mean=mn; Interaction Plots forad symbol1 v=circle i=join; symbol2 v=square i=join; proc gplot data=ymeanad; plot mn*a=d; run; 33 Lecture 12 Page 33

ANOVA witha,c andd and Their Interactions proc glm data=filter; class A C D; model y=a C D; run; quit; Source DF Sum of Squares Mean Square F Value Pr > F Model 7 5551.437500 793.062500 35.35 <.0001 Error 8 179.500000 22.437500 Cor Total 15 5730.937500 Source DF Type I SS Mean Square F Value Pr > F A 1 1870.562500 1870.562500 83.37 <.0001 C 1 390.062500 390.062500 17.38 0.0031 A*C 1 1314.062500 1314.062500 58.57 <.0001 D 1 855.562500 855.562500 38.13 0.0003 A*D 1 1105.562500 1105.562500 49.27 0.0001 C*D 1 5.062500 5.062500 0.23 0.6475 A*C*D 1 10.562500 10.562500 0.47 0.5120 ANOVA confirms that A, C, D, AC and AD are significant effects 34 Lecture 12 Page 34

Statistics 514:2 k Factorial Design Summer 2018 /* the same date step */ Regression Model data inter; set filter; AC=A*C; AD=A*D; proc reg data=inter; model y=a C D AC AD; output out=outres r=res p=pred; proc gplot data=outres; plot res*pred; run; quit; Analysis of Variance Sum of Mean Source DF Squares Square F Value Pr > F Model 5 5535.81250 1107.16250 56.74 <.0001 Error 10 195.12500 19.51250 Corrected Total 15 5730.93750 35 Lecture 12 Page 35

Root MSE 4.41730 R-Square 0.9660 Dependent Mean 70.06250 Adj R-Sq 0.9489 Coeff Var 6.30479 Parameter Estimates Parameter Standard Variable DF Estimate Error t Value Pr > t Intercept 1 70.06250 1.10432 63.44 <.0001 A 1 10.81250 1.10432 9.79 <.0001 C 1 4.93750 1.10432 4.47 0.0012 D 1 7.31250 1.10432 6.62 <.0001 AC 1-9.06250 1.10432-8.21 <.0001 AD 1 8.31250 1.10432 7.53 <.0001 36 Lecture 12 Page 36

Response Optimization / Best Setting Selection Usex 1,x 3,x 4 fora,c,d; andx 1 x 3,x 1 x 4 forac,ad respectively. The regression model gives the following function for the response (filtration rate): y = 70.06+10.81x 1 +4.94x 3 +7.31x 4 9.06x 1 x 3 +8.31x 1 x 4 Want to maximize the response. LetD be set at high level (x 4 = 1) y = 77.37+19.12x 1 +4.94x 3 9.06x 1 x 3 37 Lecture 12 Page 37

Contour Plot for Response GivenD: Using SAS data one; do x1 = -1 to 1 by.1; do x3 = -1 to 1 by.1; y=77.37+19.12*x1 +4.94*x3-9.06*x1*x3 ; output; end; end; proc gcontour data=one; plot x3*x1=y; run; quit; 38 Lecture 12 Page 38

Half normal plot for(x i ),i = 1,...,n: let x i be the absolute values ofx i sort the( x i ): x (1)... x (n) Some Other Issues calculateu i = Φ 1 ( n+i ),i = 1,...,n 2n+1 plot x (i) againstu i look for a straight line Half normal plot can also be used for identifying important factorial effects Alternatives: Hamada & Balakrishnan (1998) Analyzing unreplicated factorial experiments: a review with some new proposals, Statistica Sinica 8: 1-41. Detect dispersion effects Experiment with duplicate measurements for each treatment combination: n responses from duplicate measurements calculate meanȳ and standard deviations. Useȳ and treat the experiment as unreplicated in analysis 39 Lecture 12 Page 39