Variance Estimates and the F Ratio. ERSH 8310 Lecture 3 September 2, 2009

Similar documents
ANALYTICAL COMPARISONS AMONG TREATMENT MEANS (CHAPTER 4)

Analytical Comparisons Among Treatment Means (Chapter 4) Analysis of Trend (Chapter 5) ERSH 8310 Fall 2009

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares

10/31/2012. One-Way ANOVA F-test

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

The One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)

Regression With a Categorical Independent Variable

HYPOTHESIS TESTING. Hypothesis Testing

Sampling Distributions: Central Limit Theorem

Introduction to the Analysis of Variance (ANOVA)

The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization.

One-way ANOVA. Experimental Design. One-way ANOVA

Lecture 7: Hypothesis Testing and ANOVA

Analysis of Variance: Repeated measures

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals

8/23/2018. One-Way ANOVA F-test. 1. Situation/hypotheses. 2. Test statistic. 3.Distribution. 4. Assumptions

1. What does the alternate hypothesis ask for a one-way between-subjects analysis of variance?

Lab #12: Exam 3 Review Key

Two-Sample Inferential Statistics

Lecture 11: Two Way Analysis of Variance

psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

LECTURE 5. Introduction to Econometrics. Hypothesis testing

Chap The McGraw-Hill Companies, Inc. All rights reserved.

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career.

Chapter 8 Student Lecture Notes 8-1. Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance

Hypothesis testing: Steps

Sleep data, two drugs Ch13.xls

INTRODUCTION TO ANALYSIS OF VARIANCE

Regression With a Categorical Independent Variable: Mean Comparisons

Department of Economics. Business Statistics. Chapter 12 Chi-square test of independence & Analysis of Variance ECON 509. Dr.

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8

One-factor analysis of variance (ANOVA)

Introduction to Analysis of Variance. Chapter 11

Rama Nada. -Ensherah Mokheemer. 1 P a g e

DISTRIBUTIONS USED IN STATISTICAL WORK

While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1

1 Descriptive statistics. 2 Scores and probability distributions. 3 Hypothesis testing and one-sample t-test. 4 More on t-tests

Regression With a Categorical Independent Variable

2 and F Distributions. Barrow, Statistics for Economics, Accounting and Business Studies, 4 th edition Pearson Education Limited 2006

Analysis of Variance: Part 1

Chapter 24. Comparing Means

CHAPTER 10 ONE-WAY ANALYSIS OF VARIANCE. It would be very unusual for all the research one might conduct to be restricted to

Unit 27 One-Way Analysis of Variance

ONE FACTOR COMPLETELY RANDOMIZED ANOVA

Chapter 7. Inference for Distributions. Introduction to the Practice of STATISTICS SEVENTH. Moore / McCabe / Craig. Lecture Presentation Slides

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)

COMPARING SEVERAL MEANS: ANOVA

OHSU OGI Class ECE-580-DOE :Design of Experiments Steve Brainerd

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

COGS 14B: INTRODUCTION TO STATISTICAL ANALYSIS

Battery Life. Factory

An inferential procedure to use sample data to understand a population Procedures

Assignment #7. Chapter 12: 18, 24 Chapter 13: 28. Due next Friday Nov. 20 th by 2pm in your TA s homework box

Comparing Several Means: ANOVA

Introduction to Business Statistics QM 220 Chapter 12

Factorial Independent Samples ANOVA

Analysis of Variance (ANOVA)

Hypothesis testing: Steps

Analysis Of Variance Compiled by T.O. Antwi-Asare, U.G

Independent Samples ANOVA

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

Inferences for Regression

Chapter 9 Inferences from Two Samples

Notes for Week 13 Analysis of Variance (ANOVA) continued WEEK 13 page 1

Chapter Seven: Multi-Sample Methods 1/52

Hypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -

Statistical Foundations:

Using SPSS for One Way Analysis of Variance

The t-statistic. Student s t Test

Statistics For Economics & Business

One-Way Analysis of Variance (ANOVA) Paul K. Strode, Ph.D.

What If There Are More Than. Two Factor Levels?

t-test for b Copyright 2000 Tom Malloy. All rights reserved. Regression

INTERVAL ESTIMATION AND HYPOTHESES TESTING

Hypothesis Tests and Estimation for Population Variances. Copyright 2014 Pearson Education, Inc.

DESAIN EKSPERIMEN Analysis of Variances (ANOVA) Semester Genap 2017/2018 Jurusan Teknik Industri Universitas Brawijaya

STAT 135 Lab 9 Multiple Testing, One-Way ANOVA and Kruskal-Wallis

CHAPTER 4 Analysis of Variance. One-way ANOVA Two-way ANOVA i) Two way ANOVA without replication ii) Two way ANOVA with replication

One-Way Analysis of Variance. With regression, we related two quantitative, typically continuous variables.

Applied Statistics for the Behavioral Sciences

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table

Econ 3790: Business and Economic Statistics. Instructor: Yogesh Uppal

Statistical Inference. Why Use Statistical Inference. Point Estimates. Point Estimates. Greg C Elvers

Statistics: revision

Calculating Fobt for all possible combinations of variances for each sample Calculating the probability of (F) for each different value of Fobt

Samples and Populations Confidence Intervals Hypotheses One-sided vs. two-sided Statistical Significance Error Types. Statistiek I.

Design of Experiments. Factorial experiments require a lot of resources

Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means

One-way between-subjects ANOVA. Comparing three or more independent means

44.2. Two-Way Analysis of Variance. Introduction. Prerequisites. Learning Outcomes

Lecture notes 13: ANOVA (a.k.a. Analysis of Variance)

Section 9.5. Testing the Difference Between Two Variances. Bluman, Chapter 9 1

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

PSYC 331 STATISTICS FOR PSYCHOLOGISTS

An Old Research Question

Transcription:

Variance Estimates and the F Ratio ERSH 8310 Lecture 3 September 2, 2009

Today s Class Completing the analysis (the ANOVA table) Evaluating the F ratio Errors in hypothesis testing A complete numerical example

The ANOVA Table COMPLETING THE ANALYSIS

Completing the Analysis Table 3 1 presents an ANOVA table look at the entries Note that variance is SS/df

Running Example To provide context, let s take the example shown in Chapter 2, Table 2.2 (p. 22) Control Drug A Drug B 16 4 2 18 7 10 10 8 9 12 10 13 19 1 11

Do You Remember? Now, let s compute a few things from this table: What are the treatment sums A j for each group j? n = j A j Y ij Using the treatment sums, what is the mean for each group? Y What is the grand sum? j = A n j j i= 1 = n j i= 1 Y n ij j T = Y = ij Aj

Building Bracket Terms Recall how to build the bracket terms: 1. Square all quantities in a given set (Y ij, A j, and T). 2. Sum these squared quantities (if more than one are present) 3. Divide this sum by the number of scores that went into each component So: [ Y ] Y [ A] 2 ij 2 j = A = n T 2 T = an [ ]

Sums of Squares Recall also that we used the bracket terms to construct the Sums of Squares: SS T = [Y] [T] T SS A = [A] [T] SS S/A = [Y] [A] What are these for Table 2.2?

Degrees of Freedom Degree of freedom (df): The number of independent scores in a Sum of Squares The general rule for computing the df of any sum of squares is: df = (number of independent observations) (number of restraints) or df = (number of independent observation) (number of population estimates)

Degrees of Freedom The total df comes from the sum of the within and between df df T = df A + df S/A = N 1 T A S/A or (a)(n) 1 = (a 1)+a(n 1).

DF Example Table 2.2 (p. 22) is replicated to the right 16 4 2 What are the total df for this table? Control Drug A Drug B 18 7 10 10 8 9 12 10 13 19 1 11 Bonus: What is df A? A What is df S/A?

Mean Squares The mean squares are given by: MS = SS/df. So far, we have two mean squares: SS df SS A S / A MS A = MS S / A = df A dfs / A Compute the Mean Squares for Tbl Table 22 2.2

The F Ratio The formula of the F ratio is F = MS A / MS S/A The degrees of freedom are: df A = df 1 = a 1 : (numerator df) df S/A = df 2 = a(n 1) : (denominator df) The F ratio is approximately 10 1.0 when the null hypothesis is true and is greater than 1.0 when the null hypothesis is false What is the F ratio for Table 2.2?

EVALUATING THE F RATIO

Sampling Distributions Recall from your previous statistics course the idea of a sampling distribution A frequency distribution of a statistic is called a sampling distribution of the statistic It comes from replications of the same experiment over and over, with replacement for the sample Each time, you record the mean.

Sampling Distribution Example To demonstrate sampling distributions, I would like your help and the help of our TAs 1. Please write down your height in inches on a small square of paper 2. Ask 5 different (and random) people p for their heights 3. Compute the mean and median of your sample 4. Write it next to your height on your sheet 5. Hand that to the back where our TAs are 6. Takea a 10 minute break

Survey Says Using your data, we computed the following: The true mean height (in inches) for our overall class The distribution of each of your sample means

Sampling Distributions The distribution of our sample means is called a sampling distribution Its shape follows well known statistical properties Here, we have the central limit theorem in play Any statistic can havea a sampling distribution Here is the sampling distribution for the median What we will construct is a sampling distribution for the F ratio when the null hypothesis is true

The Sampling Distribution of F Suppose that for an experiment a = 3, n = 5 (or N = 15), and the null hypothesis is true so that μ 1 = μ 2 = μ 3 Assume that we draw a very large number of such experiments each consisting of three groups of 5 scores, and that we compute the value of F for each case We can construct a graph relating F and frequency of g p g q y occurrence [see Figure 3 1 for F(2,12)]

Theoretical Distribution

The F Table The shape of the F distribution is determined by the degrees of freedom for the ratio We will use F(df num., df denom. ) or F df1,df2df2 An F table is found in Appendix A1(see A.1 pp. 571 575) Also you can compute exact p values in Excel Also, you can compute exact p values in Excel Type =fdist(f,df1, df2) and you will get a p value

Using the F Table A particular value of F in this table is specified by three factors: (1) the numerator df (2) the denominator df (3) the value of α Where α refers to the proportion of area to the right of an ordinate drawn at F a The α levels, α =.10,.05,.025,.01, and.001, are ones most commonly encountered, and these are listed in the table

The Distribution of F When the Null Hypothesis Is False The theoretical distribution when H 0 is false is called the noncentral F (i.e., F ) distribution This is used frequently to calculate power Used to figure out how big of a sample you need to detect a difference for a given effect size Very useful in planning research Always consult a statistician first!

Testing the Null Hypothesis The two hypotheses are: H 0 : μ 1 = μ 2 = μ 3 H 1 : not all μ s are equal. Assume that we: Conducted an experiment Computed the value of F Suppose we could agree on a dividing line for any F distribution, wherevalues of F falling above theline are considered to be unlikely (i.e., incompatible with H 0 ) and values of F falling below the line as considered to be likely (i.e., compatible with ihh 0 )

Decision Rules One decision rule is to reject the null hypothesis when the observed F falls within the region of incompatibility In practice, there is fairly common agreement on a In practice, there is fairly common agreement on a probability of α =.05 to define the region of incompatibility for the F distribution Doesn t always have to be this level, though

Decision Rules This probability may be called the significance level. The rejection rule is: Reject H 0 when F observed F (α) (df num.,df denom. ); retain H 0 otherwise It used to be that any reports would say that an F is significant at a given level of significance (e.g., α =.01) In modern applications a researcher reports the exact probability of F observed, the p value

Decision Rules, Continued The p value refers to the proportion of the sampling distribution of the F statistic fall at or above the F found in an experiment We do not need to consult an F table to determine significance, Simply compare the exact probability with the chosen significance ifi level l Reject H 0 if the exact probability is smaller than the chosen significance level For example, provided that α=.05, reject H 0 if p.05 and retain H 0 otherwise

Avoiding Common Misuses of Hypothesis Testing (p. 45 46) Never lose site of your data. Remember that not all null hypotheses represent interesting or plausible outcomes of the experiment and that their rejection is not inevitably informative Remember that the null hypothesis test only gives information about how likely the sampling operations are to have produced your effect When you find a significant effect, ask what it means Always interpret your results in the context of other studies in the field

ERRORS IN HYPOTHESIS TESTING

Errors in Hypothesis Testing If we reject H 0 when it is true, then we make a type I error If we retain H 0 when it is false, then we make a type II error Note that α is the probability of type I error, and that β is the probability of type II error (see Table 3 3 on page 47) Power refers to the probability of rejecting the null hypothesis when an alternative hypothesis is true (i.e., power = 1 β)

A COMPLETE NUMERICAL EXAMPLE

Vigilance Task While Sleep Deprived There are a = 4 conditions, namely, 4, 12, 20, and 28 hours without sleep There are n = 4 subjects randomly assigned to each of the different treatment conditions The vigilance task score represents the number of failures to spot objects on a radar screen during a 30 minutetest test period

Data (computation p. 51) Hours without sleep 4 hr 12 hr 20 hr 28 hr a 1 a 2 a 3 A 4 37 36 43 76 22 45 75 66 22 47 66 43 25 23 46 62

Construct the ANOVA Table Source SS Df MS F P-value A 3314.25 3 1104.75 7.34 0.005 S/A 1805.50 12 150.458 Ttl Total 5119.75 15

Do You Remember? Now, let s compute a few things from this table: What are the treatment sums A j for each group j? n = j A j Y ij Using the treatment sums, what is the mean for each group? Y What is the grand sum? j = A n j j i= 1 = n j i= 1 Y n ij j T = Y = ij Aj

Building Bracket Terms Recall how to build the bracket terms: 1. Square all quantities in a given set (Y ij, A j, and T). 2. Sum these squared quantities (if more than one are present) 3. Divide this sum by the number of scores that went into each component So: [ Y ] Y [ A] = A = n T 2 T = an [ ] ij 2 j

Sums of Squares Recall also that we used the bracket terms to construct the Sums of Squares: SS T = [Y] [T] T SS A = [A] [T] SS S/A = [Y] [A]

Degrees of Freedom The total df comes from the sum of the within and between df df T = df A + df S/A or (a)(n) 1 = (a 1)+a(n 1).

Mean Squares The mean squares are given by: MS = SS/df. So far, we have two mean squares: SS df SS A S / A MS A = MS S / A = df A dfs / A

The F Ratio The formula of the F ratio is F = MS A / MS S/A The degrees of freedom are: df A = df 1 = a 1 : (numerator df) df S/A = df 2 = a(n 1) : (denominator df)

But Now What You have computed the ANOVA And received a p value What were your hypotheses? Null: Alternative What is your decision? What does that mean?

Wrapping Up The F Ratio is the vehicle to test the null hypothesis (es) of the experiment The F Ratio is composed of two variances: between groups and within groups. The ratiofollows the idea of variance partitioning that willbe present throughout the class Don t get too wrapped up into the formulae presented today: most every calculation will be done using a computer

Lab: Up Next Sampling distributions of F How to calculate ANOVA models in SPSS Good clean fun Homework Assigned tonight due at the start of class on Wednesday Next week: Read Chapters 4 and 5