One Factor Experiments

Similar documents
Two Factor Full Factorial Design with Replications

2 k Factorial Designs Raj Jain

2 k Factorial Designs Raj Jain Washington University in Saint Louis Saint Louis, MO These slides are available on-line at:

Fractional Factorial Designs

CS 147: Computer Systems Performance Analysis

2 k, 2 k r and 2 k-p Factorial Designs

Two-Factor Full Factorial Design with Replications

Convolution Algorithm

EEC 686/785 Modeling & Performance Evaluation of Computer Systems. Lecture 19

Outline. Simulation of a Single-Server Queueing System. EEC 686/785 Modeling & Performance Evaluation of Computer Systems.

CS 5014: Research Methods in Computer Science

ANOVA: Analysis of Variation

CS 5014: Research Methods in Computer Science. Experimental Design. Potential Pitfalls. One-Factor (Again) Clifford A. Shaffer.

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants.

Chapter 10: Analysis of variance (ANOVA)

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

Statistics for Managers Using Microsoft Excel Chapter 10 ANOVA and Other C-Sample Tests With Numerical Data

Chap The McGraw-Hill Companies, Inc. All rights reserved.

CS 5014: Research Methods in Computer Science

Chapte The McGraw-Hill Companies, Inc. All rights reserved.

Number Theory. Raj Jain. Washington University in St. Louis

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College

Factorial designs. Experiments

Comparing Systems Using Sample Data

Analysing qpcr outcomes. Lecture Analysis of Variance by Dr Maartje Klapwijk

Analysis of Variance and Design of Experiments-I

Ch. 1: Data and Distributions

Unit 12: Analysis of Single Factor Experiments

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

Chapter 14 Simple Linear Regression (A)

Y it = µ + τ i + ε it ; ε it ~ N(0, σ 2 ) (1)

n i n T Note: You can use the fact that t(.975; 10) = 2.228, t(.95; 10) = 1.813, t(.975; 12) = 2.179, t(.95; 12) =

Lecture 9: Linear Regression

y response variable x 1, x 2,, x k -- a set of explanatory variables

SIMPLE REGRESSION ANALYSIS. Business Statistics

Simple Linear Regression. (Chs 12.1, 12.2, 12.4, 12.5)

Statistics For Economics & Business

Introduction to Queueing Theory

Tables Table A Table B Table C Table D Table E 675

Econ 3790: Business and Economic Statistics. Instructor: Yogesh Uppal

Two-Way Analysis of Variance - no interaction

Hypothesis Testing hypothesis testing approach

K. Model Diagnostics. residuals ˆɛ ij = Y ij ˆµ i N = Y ij Ȳ i semi-studentized residuals ω ij = ˆɛ ij. studentized deleted residuals ɛ ij =

Multiple Predictor Variables: ANOVA

Simple linear regression

1 Introduction to One-way ANOVA

Two-Way Factorial Designs

Sociology 6Z03 Review II

Inference for Regression Simple Linear Regression

10 One-way analysis of variance (ANOVA)

Correlation and the Analysis of Variance Approach to Simple Linear Regression

EEC 686/785 Modeling & Performance Evaluation of Computer Systems. Lecture 11

ANOVA Randomized Block Design

Example. Multiple Regression. Review of ANOVA & Simple Regression /749 Experimental Design for Behavioral and Social Sciences

Simple linear regression

Diagnostics and Remedial Measures

Linear models and their mathematical foundations: Simple linear regression

1 The Randomized Block Design

Simple Linear Regression

7.5 Operations with Matrices. Copyright Cengage Learning. All rights reserved.

From Practical Data Analysis with JMP, Second Edition. Full book available for purchase here. About This Book... xiii About The Author...

If we have many sets of populations, we may compare the means of populations in each set with one experiment.

Topic 21 Goodness of Fit

Inference for Regression Inference about the Regression Model and Using the Regression Line

Contents. Acknowledgments. xix

Basic Concepts in Number Theory and Finite Fields

Outline Topic 21 - Two Factor ANOVA

Semester 2, 2015/2016

Remedial Measures, Brown-Forsythe test, F test

Summarizing Measured Data

Keller: Stats for Mgmt & Econ, 7th Ed July 17, 2006

Math 423/533: The Main Theoretical Topics

Regression: Main Ideas Setting: Quantitative outcome with a quantitative explanatory variable. Example, cont.

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

Correlation & Simple Regression

STAT22200 Spring 2014 Chapter 8A

Inference for the Regression Coefficient

ANOVA CIVL 7012/8012

Exam details. Final Review Session. Things to Review

Biostatistics 380 Multiple Regression 1. Multiple Regression

McGill University. Faculty of Science MATH 204 PRINCIPLES OF STATISTICS II. Final Examination

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

36-309/749 Experimental Design for Behavioral and Social Sciences. Sep. 22, 2015 Lecture 4: Linear Regression

ANOVA (Analysis of Variance) output RLS 11/20/2016

Matrices and vectors A matrix is a rectangular array of numbers. Here s an example: A =

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =

The Random Effects Model Introduction

Factorial and Unbalanced Analysis of Variance

What is a Hypothesis?

Ch 2: Simple Linear Regression

Regression and Models with Multiple Factors. Ch. 17, 18

Multiple Comparisons. The Interaction Effects of more than two factors in an analysis of variance experiment. Submitted by: Anna Pashley

Confidence Intervals, Testing and ANOVA Summary

Econ 3790: Business and Economics Statistics. Instructor: Yogesh Uppal

Simple Linear Regression. Material from Devore s book (Ed 8), and Cengagebrain.com

Confidence Interval for the mean response

Introduction to Queueing Theory

16.3 One-Way ANOVA: The Procedure

Regression used to predict or estimate the value of one variable corresponding to a given value of another variable.

Topic 22 Analysis of Variance

Transcription:

One Factor Experiments Raj Jain Washington University in Saint Louis Saint Louis, MO 63130 Jain@cse.wustl.edu These slides are available on-line at: http://www.cse.wustl.edu/~jain/cse567-06/ 20-1

Overview! Computation of Effects! Estimating Experimental Errors! Allocation of Variation! ANOVA Table and F-Test! Visual Diagnostic Tests! Confidence Intervals For Effects! Unequal Sample Sizes 20-2

One Factor Experiments! Used to compare alternatives of a single categorical variable. For example, several processors, several caching schemes 20-3

Computation of Effects 20-4

Computation of Effects (Cont) 20-5

Example 20.1: Code Size Comparison! Entries in a row are unrelated. (Otherwise, need a two factor analysis.) 20-6

Example 20.1 Code Size (Cont) 20-7

Example 20.1: Interpretation! Average processor requires 187.7 bytes of storage.! The effects of the processors R, V, and Z are -13.3, -24.5, and 37.7, respectively. That is, " R requires 13.3 bytes less than an average processor " V requires 24.5 bytes less than an average processor, and " Z requires 37.7 bytes more than an average processor. 20-8

Estimating Experimental Errors! Estimated response for jth alternative:! Error:! Sum of squared errors (SSE): 20-9

Example 20.2 20-10

Allocation of Variation 20-11

Allocation of Variation (Cont)! Total variation of y (SST): 20-12

Example 20.3 20-13

Example 20.3 (Cont)! 89.6% of variation in code size is due to experimental errors (programmer differences). Is 10.4% statistically significant? 20-14

Analysis of Variance (ANOVA)! Importance Significance! Important Explains a high percent of variation! Significance High contribution to the variation compared to that by errors.! Degree of freedom = Number of independent values required to compute Note that the degrees of freedom also add up. 20-15

F-Test! Purpose: To check if SSA is significantly greater than SSE. Errors are normally distributed SSE and SSA have chisquare distributions. The ratio (SSA/ν Α )/(SSE/ν e ) has an F distribution. where ν A =a-1 = degrees of freedom for SSA ν e =a(r-1) = degrees of freedom for SSE Computed ratio > F [1- α; νa, ν e ] SSA is significantly higher than SSE. SSA/ν A is called mean square of A or (MSA). Similary, MSE=SSE/ν e 20-16

ANOVA Table for One Factor Experiments 20-17

Example 20.4: Code Size Comparison! Computed F-value < F from Table The variation in the code sizes is mostly due to experimental errors and not because of any significant difference among the processors. 20-18

Visual Diagnostic Tests Assumptions: 1. Factors effects are additive. 2. Errors are additive. 3. Errors are independent of factor levels. 4. Errors are normally distributed. 5. Errors have the same variance for all factor levels. Tests:! Residuals versus predicted response: No trend Independence Scale of errors << Scale of response Ignore visible trends.! Normal quantilte-quantile plot linear Normality 20-19

Example 20.5! Horizontal and vertical scales similar Residuals are not small Variation due to factors is small compared to the unexplained variation! No visible trend in the spread! Q-Q plot is S-shaped shorter tails than normal. 20-20

Confidence Intervals For Effects! Estimates are random variables! For the confidence intervals, use t values at a(r-1) degrees of freedom.! Mean responses:! Contrasts h j α j : Use for α 1 -α 2 20-21

Example 20.6: Code Size Comparison 20-22

Example 20.6 (Cont)! For 90% confidence, t [0.95; 12] = 1.782.! 90% confidence intervals:! The code size on an average processor is significantly different from zero.! Processor effects are not significant. 20-23

Example 20.6 (Cont)! Using h 1 =1, h 2 =-1, h 3 =0, ( h j =0):! CI includes zero one isn't superior to other. 20-24

Example 20.6 (Cont)! Similarly,! Any one processor is not superior to another. 20-25

Unequal Sample Sizes! By definition:! Here, r j is the number of observations at jth level. N =total number of observations: 20-26

Parameter Estimation 20-27

Analysis of Variance 20-28

Example 20.7: Code Size Comparison! All means are obtained by dividing by the number of observations added.! The column effects are 2.15, 13.75, and -21.92. 20-29

Example 20.6: Analysis of Variance 20-30

Example 20.6 ANOVA (Cont)! Sums of Squares:! Degrees of Freedom: 20-31

Example 20.6 ANOVA Table! Conclusion: Variation due processors is insignificant as compared to that due to modeling errors. 20-32

Example 20.6 Standard Dev. of Effects! Consider the effect of processor Z: Since,! Error in α 3 = Errors in terms on the right hand side:! e ij 's are normally distributed α 3 is normal with 20-33

Summary! Model for One factor experiments:! Computation of effects! Allocation of variation, degrees of freedom! ANOVA table! Standard deviation of errors! Confidence intervals for effects and contracts! Model assumptions and visual tests 20-34

Exercise 20.1 For a single factor design, suppose we want to write an expression for α j in terms of y ij 's: What are the values of a..j 's? From the above expression, the error in α j is seen to be: Assuming errors e ij are normally distributed with zero mean and variance σ e2, write an expression for variance of e αj. Verify that your answer matches that in Table 20.5. 20-35

Homework 20 Analyze the following one factor experiment: 1. Compute the effects 2. Prepare ANOVA table 3. Compute confidence intervals for effects and interpret 4. Compute Confidence interval for α 1 -α 3 5. Show graphs for visual tests and interpret 20-36