We need to define some concepts that are used in experiments.

Similar documents
ANOVA: Comparing More Than Two Means

STAT Chapter 10: Analysis of Variance

1 The Randomized Block Design

Chapter 15: Analysis of Variance

Inference for Regression Simple Linear Regression

Section 9.4. Notation. Requirements. Definition. Inferences About Two Means (Matched Pairs) Examples

Analysis of Variance

STAT 115:Experimental Designs

In ANOVA the response variable is numerical and the explanatory variables are categorical.

One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

with the usual assumptions about the error term. The two values of X 1 X 2 0 1

Unit 27 One-Way Analysis of Variance

Inferences for Regression

Chapter 10. Design of Experiments and Analysis of Variance

16.3 One-Way ANOVA: The Procedure

LAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2

Statistics For Economics & Business

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company

1 DV is normally distributed in the population for each level of the within-subjects factor 2 The population variances of the difference scores

SIMPLE REGRESSION ANALYSIS. Business Statistics

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

Introduction to Business Statistics QM 220 Chapter 12

Notes for Week 13 Analysis of Variance (ANOVA) continued WEEK 13 page 1

Multiple Regression Analysis

One-factor analysis of variance (ANOVA)

Econ 3790: Business and Economic Statistics. Instructor: Yogesh Uppal

Comparing the means of more than two groups

Sociology 6Z03 Review II

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018

Ch. 1: Data and Distributions

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College

Chapter 9 Inferences from Two Samples

Sampling Distributions: Central Limit Theorem

Hypothesis testing: Steps

Hypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -

While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1

Statistical methods for comparing multiple groups. Lecture 7: ANOVA. ANOVA: Definition. ANOVA: Concepts

ANOVA: Analysis of Variation

Chapter 12 - Lecture 2 Inferences about regression coefficient

CHAPTER 9: HYPOTHESIS TESTING

Analysis of Variance (ANOVA)

Hypothesis testing: Steps

Analysis of variance (ANOVA) ANOVA. Null hypothesis for simple ANOVA. H 0 : Variance among groups = 0

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI

Analysis of Variance

Formal Statement of Simple Linear Regression Model

HYPOTHESIS TESTING. Hypothesis Testing

The t-statistic. Student s t Test

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

This document contains 3 sets of practice problems.

1. The (dependent variable) is the variable of interest to be measured in the experiment.

Inference for Regression Inference about the Regression Model and Using the Regression Line

10/31/2012. One-Way ANOVA F-test

Assignment #7. Chapter 12: 18, 24 Chapter 13: 28. Due next Friday Nov. 20 th by 2pm in your TA s homework box

One-way Analysis of Variance. Major Points. T-test. Ψ320 Ainsworth

Week 14 Comparing k(> 2) Populations

OHSU OGI Class ECE-580-DOE :Design of Experiments Steve Brainerd

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

Data Analysis and Statistical Methods Statistics 651

Chapter 10: Analysis of variance (ANOVA)

The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization.

Two-Sample Inferential Statistics

Correlation Analysis

Multiple Linear Regression

Question. Hypothesis testing. Example. Answer: hypothesis. Test: true or not? Question. Average is not the mean! μ average. Random deviation or not?

CHAPTER 9, 10. Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities:

Analysis Of Variance Compiled by T.O. Antwi-Asare, U.G

STAT Chapter 8: Hypothesis Tests

ECO220Y Simple Regression: Testing the Slope

Chapter 8 Student Lecture Notes 8-1. Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance

Statistical Inference. Hypothesis Testing

What If There Are More Than. Two Factor Levels?

Lecture notes 13: ANOVA (a.k.a. Analysis of Variance)

INTRODUCTION TO ANALYSIS OF VARIANCE

a Sample By:Dr.Hoseyn Falahzadeh 1

Econ 3790: Business and Economics Statistics. Instructor: Yogesh Uppal

Chapter 23: Inferences About Means

Introduction to Statistical Inference Lecture 10: ANOVA, Kruskal-Wallis Test

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

Chapter 3 Multiple Regression Complete Example

PLSC PRACTICE TEST ONE

Lecture 9 Two-Sample Test. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech

Topic 6. Two-way designs: Randomized Complete Block Design [ST&D Chapter 9 sections 9.1 to 9.7 (except 9.6) and section 15.8]

Analysis of Variance. ภาว น ศ ร ประภาน ก ล คณะเศรษฐศาสตร มหาว ทยาล ยธรรมศาสตร

Sleep data, two drugs Ch13.xls

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

An Analysis of College Algebra Exam Scores December 14, James D Jones Math Section 01

BALANCED INCOMPLETE BLOCK DESIGNS

STAT Chapter 11: Regression

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:

Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance ECON 509. Dr.

CS 5014: Research Methods in Computer Science

Chapter 14 Student Lecture Notes Department of Quantitative Methods & Information Systems. Business Statistics. Chapter 14 Multiple Regression

Two sided, two sample t-tests. a) IQ = 100 b) Average height for men = c) Average number of white blood cells per cubic millimeter is 7,000.

This gives us an upper and lower bound that capture our population mean.

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants.

One-Way Analysis of Variance. With regression, we related two quantitative, typically continuous variables.

Sign test. Josemari Sarasola - Gizapedia. Statistics for Business. Josemari Sarasola - Gizapedia Sign test 1 / 13

Transcription:

Chapter 0 Analysis of Variance (a.k.a. Designing and Analysing Experiments) Section 0. Introduction In Chapter we mentioned some different ways in which we could get data: Surveys, Observational Studies, Experiments and Published Data. We have up to now analysed datasets in many ways, but we haven't studied how to collect the data. In this chapter we will look at the design of experiments and then we will see some new techniques for examining data collected in experiments. In an experiment the experimenter usually is interested in measuring the response of an individual or object to something applied to that individual or object. For example a pharmaceutical company may want to see the effect a drug has on patients with a particular disease. We need to define some concepts that are used in experiments.

Section 0. Definitions The Response Variable is the variable being measured in the experiment. Factors are variables applied to the object or individual in the experiment. We are interested in m asuring the effect of the Factors on the Response Variable. Quantitative Factors are measured on a numerical scale whereas Qualitative Factors are not measured on a numerical scale. Factor-Levels are the values of the factor used in the experiment. Treatments are the Factor-Level combinations used. An Experimental Unit is the object on which the response and factors are measured. A Designed Experiment is one for which the researcher controls the specifications of the treatments and the method of assigning the experimental units to each treatment. An Observational Experiment (Study) is one in which the researcher simply observes the treatments and the response on a sample of experimental units.

Section 0. Single Factor Anovas 0.. Introduction Suppose we have designed and conducted our experiment and it involved applying p different treatments to p different samples chosen in some manner. We are now interested in determining whether the responses of the experimental units differs according to treatment they received. As is usual in statistics we are interested in making an inference about a population using sample data. Here we have p populations each one represents the entire population who have in the past or will in the future receive each treatment. So here we have the following notation: P = Number of treatments being compared Population or Treatment 3 p Population or Treatment Mean μ μ μ 3 μ p Population or Treatment Variance σ σ σ 3 σ p Sample Size n n n 3 n p Sample Mean x x x 3 x p Sample Variance s s s 3 s p Also x is the mean of all the measurements

0.. So what are we at here? What are we trying to test here anyway? Well we're interested in seeing if there is any difference in the effect of each treatment on the experimental units. That means we are interested in testing if there is a difference between μ, μ, μ 3 etc. So in the single factor ANOVA we test the following hypotheses: H 0 : μ = μ = μ 3 =. = μ p H A : At least two of the μ's are different. Before we can do anything we must make an assumption to simplify life. 0..3 ASSUMPTION: We assume that σ = σ = σ 3 =.. = σ p and since they're all the same we just use the symbol σ. We also assume that we are dealing with a Completely Randomised Design. 0..4 Definition The Completely Randomised Design is an experimental design in which independent random samples of experimental units are selected for each treatment.

0..5 Now how will this test work? We will calculate sample means corresponding to each of the population means. We will check whether the difference between these sample means is small enough so that it would be explained by natural variability in the data. If the difference is instead larger than could be explained by natural variability then we will conclude that the population means actually are different. We could measure how far apart the individual sample means are by calculating SST the Sum of Squares for Treatments Definition: SST = [ n ( x - x ) + n ( x - x ) +. + n p ( x p - x ) ] A better measure however is MST - The Mean Square for Treatments. Definition: MST = SST/(p-) If all the individual x 's were the same and so also equal to the total x, then MST would be zero. The further the individual x 's are from each other then the bigger MST will be. If it is really big we will reject the null hypothesis and so conclude that at least two of the population means are different.

But we are left with the usual question, how big is big enough to reject? To begin to answer that question we first must notice that when H 0 is true the mean value of MST is σ. But when H 0 is false the mean of MST is not σ but instead is larger than σ. There is another statistic which estimates σ accurately and unbiased whether or not H 0 is true or not, it is called the Mean Square for Error, MSE. Definition: MSE = SSE/(N-p) where SSE = (n -) s + (n -) s +(n 3 -) s 3 + + (n p -) s p and N is the total number of observations in the dataset.

0..6 The Test So MST should in general be bigger than MSE if H 0 is false. The technique we will use therefore is to compute the ratio MST/MSE and if it is big enough we will reject H 0. This ratio will be our test statistic. So far when we have encountered test statistics they have been Z's or T's, this ratio is neither, it is from a new distribution the F. Definition: The test statistic for a single-factor ANOVA is MST F = MSE with p- numerator degrees of freedom and N-p denominator degrees of freedom. These tests are all upper tailed, we want to reject when the ratio is really big. So we look up F tables and reject H 0 if F calc is bigger than some critical value of F. All of what we have seen in this chapter so far can be summarised in the following table:

Single Factor Anova Table - Completely Randomised Design Source df SS MS F Treatments p- SST MST= SST/(p-) MST/MSE Error N-p SSE MSE= SSE/(N-p) Total N- SStotal 0..7 Anova Computations We have seen formulae for SSE and SST however in practice these formulae are very cumbersome for performing calculations. The following formulae are easier to use: SST = [n x + n x +. + n p x p ] - N x SSTotal = N x Nx SSE = SSTotal - SST

0..8 Example Detergent makers always claim that their product washes whiter. We are going to test whether there is any difference between three detergents, BOLD, DAZ and PERSIL. 9 students are asked to wear white shirts and to go out for a night's drinking and whatever. The next morning we randomly assign 3 of the 9 white shirts to be washed using BOLD, 3 using DAZ and 3 using PERSIL. After the wash is over we bring the 9 shirts over to a Biology lab and examine them under a microscope for stains. The surface area that remains dirty is determined and the following results are obtained: BOLD:,, 3 DAZ: 3, 4, 4 PERSIL:, 3, Test whether there is a difference between the population means use significance level 0.05. ANSWER: Analysis of Variance Source DF SS MS F P Treatment 4.667.333 4.0 0.07 Error 6 3.333 0.556 Total 8 8.000

Section 0.3 Randomised Block Design Sometimes randomisation can be improved upon. In the other half of this course you have seen paired T-tests. This idea of pairing can be extended to ANOVAs. Suppose we are conducting an experiment to measure the performance of a certain drug. The effect of this drug may be different on people depending on their ages, sex, blood pressure, weight etc. It is possible that the randomisation procedure does not evenly spread these characteristics among the different treatments. Sometimes it is better to force the people receiving each treatment to be the same. For example in this drug trial, we can pick groups of people with similar characteristics to receive the different treatments that way any difference we observe will be because of the treatments not because of different characteristics of the people. 0.3. Definition: A Randomised Block Design is an experimental procedure consisting of two steps:. The experimental units are divided into b blocks, the units chosen for a Block will be as similar as possible. There are p units in each Block where p is the number of Treatments being compared.. One experimental unit from each Block is randomly assigned to each treatment so in total there will be n = bp responses.

0.3. Assumptions. The observations corresponding to all blocktreatment combinations are Normally distributed.. The variances of all the Normal distributions are the same, σ. But the mean of all the Normal distributions may depend on the treatment applied and also on the block. 0.3.3 Notation and Formulae p = Number of Treatments b = Number of Blocks The average of all observations on the i th treatment is x Ti The average of all observations on the i th block is x Bi SST = b( x x) p = b x Nx p T i T i

b SSB = p( x x) b = p x Nx B i SSTotal = ( x x) N = x N Nx And SSE = SSTotal -SST - SSB 0.3.4 The Test B i The purpose of the Randomised Block Design is the same as the Completely Randomised Design ie to test: H 0 : μ = μ = μ 3 =. = μ p H A : At least two of the μ's are different.

The procedure for the test is to complete an ANOVA table in the following format: Source df SS MS F Treatments p- SST MST= SST/(p-) MST/MSE Blocks b- SSB MSB= MSB/MSE SSB/(b-) Error N-p-b+ SSE MSE= SSE/(N-p) Total N- SStotal The last column in this table includes two F statistics, the first one F=MST/MSE is the statistic used to test H 0 : The treatment means are all the same vs H A : At least two of the treatment means differ. The second test statistic F=MSB/MSE tests whether there is a difference in the Block means. Rejection of this test indicates that the Block means differ and that the approach of using a Randomised Block Design instead of a Completely Randomised Design was a good choice.

0.3.5 Example A single factor ANOVA was conducted using a Randomised Block Design. It yielded the following results: Block 3 4 3 6 8 4 6 0 3 3 3 7 Treatments Do the data provide sufficient evidence to suggest that the treatment means differ? Do the data provide sufficient evidence top indicate that blocking was effective in reducing the experimental error? Source df SS MS F Treatments Blocks Error Total

Section 0.4 Multiple Comparisons of Means 0.4. Introduction In the previous section we have seen how to determine if a set of n means are equal or if there is a difference between at least two of the means. In rejecting the F test of the single factor Anova all that we determined was that there was a difference between at least two of the means. We did not identify which of the means were different and which the same. In an experiment with 3 treatments there are three pairs of means which may be different (μ - μ ), (μ - μ 3 ) & (μ - μ 3 ). Having established that one of these pairs differs we now must test each pair seperately. In an experiment with P treatments there will be p ie: "p choose " possible combinations of pairs of means to be tested. Each pair of means will be tested essentially using a T-test. However there are two minor differences to the way we will perform these tests. Firstly we will actually use compute Confidence Intervals instead of performing Hypothesis Tests. Secondly we must use a different set of T-tables called Bonferroni -T tables because we are carrying out multiple tests.

0.4. Why Confidence Intervals are just like Hypothesis Tests Suppose we compute a 90% Confidence Interval the probability that the interval contains the actual value of the Population Characteristic is 0.9. That means the probability that the interval does not contain the actual value of the test statistic is 0.. The region outside the 90% Confidence Interval is therefore just like the Rejection Region of a Two Tailed Hypothesis Test with α = 0.. This is evident from the picture. So if we find that a 90% Confidence Interval for μ - μ does not contain the value 0 that is equivalent to rejecting the Null Hypothesis of the following Two Tailed Test with significance level α = 0.: H 0 : μ - μ = 0 Vs H A : μ - μ 0

0.4.3 Multiple Tests and Type errors When a Hypothesis test is performed at significance level α = 0., that means that there is a 0. probability of rejecting the Null Hypothesis when in fact it is true. When we perform two tests each with α = 0., the probability of making a Type error on the first test or on the second test is now larger than 0.. Remember P(A or B) = P(A) + P(B) - P(A and B) In an experiment with p treatments we will be performing p(p-)/ tests and the combined probability of making a Type error will be large. For this reason certain new versions of the T-Tables were invented called Bonferroni T-Tables. These tables take into account that you may want to perform several linked tests and they give critical values to be used in each test so that the OVERALL probability of a Type error for the Experiment as a whole is below a certain threshold.

0.4.4 The Bonferroni Multiple Comparisons Procedure When comparing p treatments or populations, we decide whether two population means differ by. Computing the p(p-)/ confidence intervals below. Checking if any intervals do not contain 0. 3. Intervals which do not contain 0 indicate a significant difference in the population means (μ's). μ μ :. x x ± ( Bonferroni Tcritical) MSE n + MSE n.. μ p μ p : x p x p ± ( Bonferroni Tcritical) MSE n p + MSE n p The Bonferroni T critical values are computed using the Error degrees of freedom.