One-Sample and Two-Sample Means Tests

Similar documents
Chapter 9 Inferences from Two Samples

Hypothesis Testing hypothesis testing approach formulation of the test statistic

Stat 427/527: Advanced Data Analysis I

Introduction to hypothesis testing

Two-Sample Inferential Statistics

ANOVA: Analysis of Variation

Lecture 17. Ingo Ruczinski. October 26, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University

Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.

10/31/2012. One-Way ANOVA F-test

Didacticiel Études de cas. Parametric hypothesis testing for comparison of two or more populations. Independent and dependent samples.

Two-sample inference: Continuous data

Comparison of two samples

+ Specify 1 tail / 2 tail

Stats Review Chapter 14. Mary Stangler Center for Academic Success Revised 8/16

COMPLETELY RANDOM DESIGN (CRD) -Design can be used when experimental units are essentially homogeneous.

4.1 Hypothesis Testing

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

Statistics: revision

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

Advanced Experimental Design

Business Statistics. Lecture 10: Course Review

Chapter 24. Comparing Means. Copyright 2010 Pearson Education, Inc.

Introduction to Linear regression analysis. Part 2. Model comparisons

MBA 605, Business Analytics Donald D. Conant, Ph.D. Master of Business Administration

Inferences About the Difference Between Two Means

COMPARING SEVERAL MEANS: ANOVA

What Is ANOVA? Comparing Groups. One-way ANOVA. One way ANOVA (the F ratio test)

Psychology 282 Lecture #4 Outline Inferences in SLR

13: Additional ANOVA Topics

PLSC PRACTICE TEST ONE

Tentative solutions TMA4255 Applied Statistics 16 May, 2015

Analysis of variance (ANOVA) Comparing the means of more than two groups

Chapter 7. Inference for Distributions. Introduction to the Practice of STATISTICS SEVENTH. Moore / McCabe / Craig. Lecture Presentation Slides

Lecture 17: Small-Sample Inferences for Normal Populations. Confidence intervals for µ when σ is unknown

Chapter 23: Inferences About Means

T test for two Independent Samples. Raja, BSc.N, DCHN, RN Nursing Instructor Acknowledgement: Ms. Saima Hirani June 07, 2016

Note that we are looking at the true mean, μ, not y. The problem for us is that we need to find the endpoints of our interval (a, b).

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

Statistics for Managers Using Microsoft Excel Chapter 9 Two Sample Tests With Numerical Data

Introduction to the Analysis of Variance (ANOVA)

8/23/2018. One-Way ANOVA F-test. 1. Situation/hypotheses. 2. Test statistic. 3.Distribution. 4. Assumptions

Statistics Boot Camp. Dr. Stephanie Lane Institute for Defense Analyses DATAWorks 2018

Review 6. n 1 = 85 n 2 = 75 x 1 = x 2 = s 1 = 38.7 s 2 = 39.2

Introduction to Business Statistics QM 220 Chapter 12

2011 Pearson Education, Inc

Statistical Intervals (One sample) (Chs )

The Difference in Proportions Test

MS&E 226: Small Data

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

16.400/453J Human Factors Engineering. Design of Experiments II

Chapter 8 Heteroskedasticity

Chapter 12. ANalysis Of VAriance. Lecture 1 Sections:

CHAPTER 7. Hypothesis Testing

Hypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -

Inferences for Regression

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

I used college textbooks because they were the only resource available to evaluate measurement uncertainty calculations.

Statistics for IT Managers

Correlation Analysis

a Sample By:Dr.Hoseyn Falahzadeh 1

Chapter 12: Linear regression II

Testing for Normality

Difference between means - t-test /25

Statistics for Managers using Microsoft Excel 6 th Edition

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

An inferential procedure to use sample data to understand a population Procedures

Introduction to Statistical Data Analysis III

GROUPED DATA E.G. FOR SAMPLE OF RAW DATA (E.G. 4, 12, 7, 5, MEAN G x / n STANDARD DEVIATION MEDIAN AND QUARTILES STANDARD DEVIATION

MORE ON SIMPLE REGRESSION: OVERVIEW

Descriptive Statistics (And a little bit on rounding and significant digits)

Multiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600

MAT2377. Rafa l Kulik. Version 2015/November/23. Rafa l Kulik

Calculating Fobt for all possible combinations of variances for each sample Calculating the probability of (F) for each different value of Fobt

STA Module 10 Comparing Two Proportions

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =

Confidence Intervals, Testing and ANOVA Summary

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College

Visual interpretation with normal approximation

Density Temp vs Ratio. temp

Inference for the mean of a population. Testing hypotheses about a single mean (the one sample t-test). The sign test for matched pairs

Chap The McGraw-Hill Companies, Inc. All rights reserved.

Chapter 24. Comparing Means

Inference with Simple Regression

Sociology 6Z03 Review II

Ordinary Least Squares Regression Explained: Vartanian

In a one-way ANOVA, the total sums of squares among observations is partitioned into two components: Sums of squares represent:

Chapter 23. Inferences About Means. Monday, May 6, 13. Copyright 2009 Pearson Education, Inc.

Nature vs. nurture? Lecture 18 - Regression: Inference, Outliers, and Intervals. Regression Output. Conditions for inference.

Confidence Intervals. - simply, an interval for which we have a certain confidence.

Non-Parametric Two-Sample Analysis: The Mann-Whitney U Test

13: Additional ANOVA Topics. Post hoc Comparisons

Contrasts and Multiple Comparisons Supplement for Pages

Statistics Handbook. All statistical tables were computed by the author.

Sampling Distributions: Central Limit Theorem

Note that we are looking at the true mean, μ, not y. The problem for us is that we need to find the endpoints of our interval (a, b).

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

HYPOTHESIS TESTING II TESTS ON MEANS. Sorana D. Bolboacă

Soil Phosphorus Discussion

AP Statistics Cumulative AP Exam Study Guide

Transcription:

One-Sample and Two-Sample Means Tests

1 Sample t Test The 1 sample t test allows us to determine whether the mean of a sample data set is different than a known value. Used when the population variance is not known. Can be used when the sample size is small. Use n-1 degrees of freedom.

For example, we are interested in determining if the mean per capita income of West Virginia counties is different than the national average, and we suspect based on a priori (before hand) knowledge that it is lower. Our hypotheses are: H o : The per capita income of West Virginia counties is not significantly less than the national average. H a : The per capita income of West Virginia counties is significantly less than the national average.

Procedure Steps 1. Determine if West Virginia county per capita income is normally distributed.. Proceed with the 1 sample t test. We do not need to know the normality of US per capita income. All we need to know is the mean.

Per Capita Income Median Household Income Median Family Income Population Households United States $7,334 $51,914 $6,98 308,745,538 116,716,9 West Virginia $19,443 $38,380 $48896 1,85,994 763,831

W/S Test for Normality sd $3,075.8 range $16,778 q q q q critical w s $16,778 $3,075 5.46 3.90, 5.46 Since 3.90 < q5.46 < 5.46, accept Ho, the data are not significantly different than normal (w/s 5.46, p > 0.05). Just barely normal.

The barely normal result of the w/s test is due to its testing of k 3 and k 4, so it is more sensitive to leptokurtic data.

The t statistic is calculated as: t x µ s / n Where x is the sample mean, µ is the true mean, s is the sample standard deviation, and n is the sample size.

The effect of sample size on t test results: n 7 n 14 t (15 13).63 t / 7 /.65 0.76 (15 13) / 14 / 3.74 0.53 3.77 Small sample sizes make tests MORE conservative (e.g. harder to gain significance). Small but important differences may not be detected. Increasing the sample size allows us to detect smaller, but statistically significant, differences. Larger sample sizes make statistical test more powerful... but remember there is a point of diminishing returns.

West Virginia Mean: 1953. S: 3075.8 n: 54 United States Mean: 7334 t 1953. 7334.0 3075.8/ 54 8080.8 813.6 9.93 df 54 1 53 We take the calculated t value and the df to the t table to determine the probability. Although we ignore the sign on the t table, we can tell from the sign that WV per capita income is less than the national average.

t -9.93 df 53 Since our df is not listed, use the next lower df. Where does the calculated t (-9.93) fall?

Since -9.93 < 1.684, reject H 0. The per capita income of West Virginia is significantly less than the national average (t -9.93, p<0.005). West Virginia s per capita income is very low.

Note that the confidence intervals for the mean are calculated using the critical t value from the table. pr x s s tα, df µ x + tα, df 1 α n n If the df you need is not in the table, use the next LOWER value. 1953. 1.684 757.04 54 µ 1953. + 1.684 757.04 54 1 α 1861.4 µ 19885.0

Two-Sample T-Test for Means Used to compare one sample mean to another. Two different test: Equal variances Unequal variances Homoscedasticity the assumption of equal variances.

When data meet the assumption of normality we can use confidence intervals to quantify the area of uncertainty (in this case 5%). Here we are 95% confident that the means do not overlap and that these two groups are significantly different.

In this example the means could occur in an overlapping region and we are less than 95% certain that they are significantly different.

When the variances are not equal, the region of overlap is asymmetrical and the probabilities associated with the location of the mean are not the same. To avoid a Type 1 error we use a more conservative approach.

Test 1: Equal Variances The test statistic is: t X s 1 n 1 1 X + s n df n 1 +n - s 1 and s are the variances (or squared standard deviations) for each sample, n 1 and n are the group samples sizes. Therefore t is the distance between the two means, in standard deviation units, taking into consideration sample size. For the -sample t test we know means, therefore the degrees of freedom would be: df n 1 +n -.

When comparing two or more groups, each group MUST tested for normality individually. If we pool the data and test for normality, then we are assuming that the data are from the same population which is what we are trying to determine with the t test.

To determine if the variances are equal, use the equation: Numerator df F s1 df n1 1, n s ( 1) Denominator df Which is just the ratio of one variance to the other. The F results are then compared to the F table.

Note that there are several additional pages not shown

Example: -Sample T Test with Equal Variances Research question: Is there a difference in the sitting height between the Arctic and Great Plains native Americans? H 0 : There is no significant difference in the sitting heights of of Arctic and Great Plains native Americans. H a : There is a significant difference in the sitting heights of of Arctic and Great Plains native Americans. Arctic Plains N 8 16 Mean 864.9 88.7 cm s 488. 43.6 Range 71. 84.

The order of operations for conducting the test is: 1. Test each group for normality. Proceed with t test if both groups are normal.. Conduct an F test. A. If H o is accepted then proceed with equal variances t test. B. If Ho is rejected then proceed with Welch s approximate t test. 3. Conduct appropriate t test.

Test each group for normality: Arctic n 8 Range Arctic Variance 71. Arctic 488. Plains n 16 Range Plains Variance 84. Plains 43.6 q Arctic 71. 488. 3. q Critical.50,3.399 q Arctic 84. 43.6 4.049 q Critical 3.01,4.4 Since.50 < q 3. < 3.399, accept H o. Arctic sitting height is not different than normal (q 3., p>0.05). Since 3.01 < q 4.049 < 3.4.4, accept H o. Plains sitting height is not different than normal (q 4.049, p>0.05). Since both groups are normal, proceed to the variance (F) test.

Variance or F Test: H 0 : The variances are not significantly different. H a : The variances are significantly different. α 0.05 488. F 1.13 df (16 1, 8 1) 43.6 (15, 7) From the F table the critical value for 15,7 df is.71.

Note that on the F table the probabilities are read in the columns rather than the rows. The calculated F statistic falls here between 0.5 and 0.50.

Since our calculated F statistic is less than the critical value from the table (1.13 <.71) we assume that the variances are equal. The variances for two sitting heights are not significantly different (F 1.13,.5 < p < 0.50). Now we can proceed with the two-sample t test

Arctic Plains N 8 16 Mean 864.9 88.7 cm S 488. 43.6 Range 71. 84. t 864.9 88.7 488. 8 + 43.6 16 17.8 9.38 1.9 Since we do not care about direction (larger or smaller) the sign does not matter. If we do care about direction then the sign is very important.

For example: No direction implied (doesn t matter): t Arctic Plains 864.9 88.7 488. 43.6 + 8 16 17.8 9.38 1.9 Sign has no meaning. Arctic tribe shorter than plains tribe: t Arctic 864.9 488. 8 Plains 88.7 43.6 + 16 17.8 9.38 1.9 Sign means the arctic tribe s mean is less than the plains tribe s mean. Plains tribe taller than arctic tribe: t Plains 88.7 43.6 16 Arctic 864.9 488. + 8 17.8 9.38 1.9 Sign means plains the tribe s mean is greater than the arctic tribe s mean.

Remember that the positioning of each group in the numerator of the t equation is related to the sign of the results. Be very careful how you construct the equation s numerator, AND the way you interpret the results.

Note: use a -tailed test since we are just interested in whether there is a difference, not if one is greater or less than the other.

In Table A.3 the critical value for df is.074. Since 1.95 <.074 we therefore accept H o : There is no significant difference in the sitting heights of Arctic and Great Plains native Americans (t 1.95, 0.10 > p > 0.05). The probability range (e.g. 0.10 > p > 0.05) are from the t-table.

Note: -tailed t1.95 for a df(v) of falls between 0.10 and 0.05.

Test : Unequal Variances (Welch s approximate t) The test statistic is: s 1 and s are the variances (or squared standard deviations) for each sample, n 1 and n are the group samples sizes. The df may be non-integer, in which case the next smaller integer should be used. Take this number to the t table. 1 1 1 1 1 1 1 1 1 1 + + + n n s n n s n s n s df and n s n s X X t

Is the vegetation index for wetland 35 less than the index for 3? Ho: The variances are not significantly different. ID 3: Mean 544 Variance 34055 n 144 ID 35: Mean 1887 Variance 884 n 144 F 34055 11.8 FCritical 884 ( 144,144) 1.35 From F table for df 10,10 (next lower on the table) The wetland variances are significantly different (F 11.8, p < 0.001).

t t t ID35 1887 544 884 144 657 0 + 103.5 + 36.5 ID3 34055 144 t Critical and df 1.654 df 34055 144 34055 144 144 1 6579.5 391.1+.8 884 + 144 884 144 + 144 1 167 The vegetation index for wetland 35 is significantly less than the index for wetland 3 (t -103.5, p < 0.001).

Variable Distributions and Test Statistic Distributions The variable may have a negative exponential distribution (e.g. frequency of a Ebola cases per outbreak). Most locations have very few people with Ebola.

Take 100 samples from the previous distribution. Calculate a mean for each sample. The central limit theorem states that the sample means will be normally distributed. 4 sample means 48 sample means 7 sample means

Some sample means will be greater than the true mean, some will be less. However, if the number of sample means is large they will take on a the properties of a normal distribution. This is true even if the underlying population has a different distribution.

Robustness in Statistics Robust when a statistical technique remains useful even when one or more of its assumptions are violated. In general, if the sample size is large moderate deviations from the statistical technique s assumptions will not invalidate the results. The F and t tests are considered to be fairly robust other tests are not, so simply increasing the sample size may not be the best approach if the data violate assumptions.