Project proposal feedback. Unit 4: Inference for numerical variables Lecture 3: t-distribution. Statistics 104

Similar documents
Hypotheses. Poll. (a) H 0 : µ 6th = µ 13th H A : µ 6th µ 13th (b) H 0 : p 6th = p 13th H A : p 6th p 13th (c) H 0 : µ diff = 0

Announcements. Statistics 104 (Mine C etinkaya-rundel) Diamonds. From last time.. = 22

Unit4: Inferencefornumericaldata. 1. Inference using the t-distribution. Sta Fall Duke University, Department of Statistical Science

Announcements. Unit 3: Foundations for inference Lecture 3: Decision errors, significance levels, sample size, and power.

Lecture 19: Inference for SLR & Transformations

Nature vs. nurture? Lecture 18 - Regression: Inference, Outliers, and Intervals. Regression Output. Conditions for inference.

Review of Statistics 101

Announcements. Unit 7: Multiple linear regression Lecture 3: Confidence and prediction intervals + Transformations. Uncertainty of predictions

STA 101 Final Review

Weldon s dice. Lecture 15 - χ 2 Tests. Labby s dice. Labby s dice (cont.)

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Chapter 23. Inferences About Means. Monday, May 6, 13. Copyright 2009 Pearson Education, Inc.

Do students sleep the recommended 8 hours a night on average?

appstats27.notebook April 06, 2017

Chapter 27 Summary Inferences for Regression

Warm-up Using the given data Create a scatterplot Find the regression line

Lecture 11 - Tests of Proportions

Contingency Tables. Safety equipment in use Fatal Non-fatal Total. None 1, , ,128 Seat belt , ,878

Announcements. Unit 4: Inference for numerical variables Lecture 4: ANOVA. Data. Statistics 104

Announcements. Unit 6: Simple Linear Regression Lecture : Introduction to SLR. Poverty vs. HS graduate rate. Modeling numerical variables

Occupy movement - Duke edition. Lecture 14: Large sample inference for proportions. Exploratory analysis. Another poll on the movement

Quantitative Analysis and Empirical Methods

Business Analytics and Data Mining Modeling Using R Prof. Gaurav Dixit Department of Management Studies Indian Institute of Technology, Roorkee

Comparing Means from Two-Sample

INFERENCE FOR REGRESSION

Announcements. Lecture 18: Simple Linear Regression. Poverty vs. HS graduate rate

Example - Alfalfa (11.6.1) Lecture 14 - ANOVA cont. Alfalfa Hypotheses. Treatment Effect

EC2001 Econometrics 1 Dr. Jose Olmo Room D309

Unit5: Inferenceforcategoricaldata. 4. MT2 Review. Sta Fall Duke University, Department of Statistical Science

Unit 6 - Simple linear regression

Business Statistics. Lecture 10: Course Review

ACMS Statistics for Life Sciences. Chapter 13: Sampling Distributions

Contingency Tables. Contingency tables are used when we want to looking at two (or more) factors. Each factor might have two more or levels.

Confidence Intervals. Confidence interval for sample mean. Confidence interval for sample mean. Confidence interval for sample mean

Two-sample inference: Continuous data

Annoucements. MT2 - Review. one variable. two variables

Announcements. Lecture 1 - Data and Data Summaries. Data. Numerical Data. all variables. continuous discrete. Homework 1 - Out 1/15, due 1/22

Discrete distribution. Fitting probability models to frequency data. Hypotheses for! 2 test. ! 2 Goodness-of-fit test

Hypothesis Testing hypothesis testing approach formulation of the test statistic

Chapter 8 Handout: Interval Estimates and Hypothesis Testing

1 Descriptive statistics. 2 Scores and probability distributions. 3 Hypothesis testing and one-sample t-test. 4 More on t-tests

Lecture 41 Sections Mon, Apr 7, 2008

Harvard University. Rigorous Research in Engineering Education

Two-sample inference: Continuous data

Announcements. Final Review: Units 1-7

An Introduction to Mplus and Path Analysis

Regression Analysis II

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =

Ch18 links / ch18 pdf links Ch18 image t-dist table

Chapter 7. Inference for Distributions. Introduction to the Practice of STATISTICS SEVENTH. Moore / McCabe / Craig. Lecture Presentation Slides

Content by Week Week of October 14 27

Hypothesis testing: Steps

Truck prices - linear model? Truck prices - log transform of the response variable. Interpreting models with log transformation

Tables Table A Table B Table C Table D Table E 675

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing

One-sample categorical data: approximate inference

Chapter 8: Estimating with Confidence

Chapter 9 Inferences from Two Samples

Regression so far... Lecture 21 - Logistic Regression. Odds. Recap of what you should know how to do... At this point we have covered: Sta102 / BME102

Statistics for IT Managers

Chapter 24. Comparing Means. Copyright 2010 Pearson Education, Inc.

Announcements. Lecture 5: Probability. Dangling threads from last week: Mean vs. median. Dangling threads from last week: Sampling bias

Example. Multiple Regression. Review of ANOVA & Simple Regression /749 Experimental Design for Behavioral and Social Sciences

Advanced Experimental Design

Lecture 15 - ANOVA cont.

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

Introduction to Survey Analysis!

Hypothesis testing: Steps

Class 24. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

The t-statistic. Student s t Test

Lecture 9. Selected material from: Ch. 12 The analysis of categorical data and goodness of fit tests

Chapter 5: HYPOTHESIS TESTING

Review for Final. Chapter 1 Type of studies: anecdotal, observational, experimental Random sampling

The Purpose of Hypothesis Testing

20 Hypothesis Testing, Part I

PHP2510: Principles of Biostatistics & Data Analysis. Lecture X: Hypothesis testing. PHP 2510 Lec 10: Hypothesis testing 1

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 9.1-1

7.2 One-Sample Correlation ( = a) Introduction. Correlation analysis measures the strength and direction of association between

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

First we look at some terms to be used in this section.

Lecture 26: Chapter 10, Section 2 Inference for Quantitative Variable Confidence Interval with t

Inference for Regression Inference about the Regression Model and Using the Regression Line, with Details. Section 10.1, 2, 3

Confidence Intervals, Testing and ANOVA Summary

The goodness-of-fit test Having discussed how to make comparisons between two proportions, we now consider comparisons of multiple proportions.

Multiple linear regression

1 Correlation and Inference from Regression

AP Statistics Cumulative AP Exam Study Guide

Example - Alfalfa (11.6.1) Lecture 16 - ANOVA cont. Alfalfa Hypotheses. Treatment Effect

Midterm 2 - Solutions

STAT 4385 Topic 01: Introduction & Review

Hotelling s One- Sample T2

Mathematical Notation Math Introduction to Applied Statistics

Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.

An Introduction to Path Analysis

Sampling Distributions: Central Limit Theorem

Week Topics of study Home/Independent Learning Assessment (If in addition to homework) 7 th September 2015

Multiple Regression Analysis

Wolf River. Lecture 19 - ANOVA. Exploratory analysis. Wolf River - Data. Sta 111. June 11, 2014

Chapter 20 Comparing Groups

HYPOTHESIS TESTING: THE CHI-SQUARE STATISTIC

Transcription:

Announcements Project proposal feedback Unit 4: Inference for numerical variables Lecture 3: t-distribution Statistics 104 Mine Çetinkaya-Rundel October 22, 2013 Cases: units, not values Data prep: If categorical, use meaningful level labels that are not just numbers Use short names for dataset and variable names to streamline analysis Data collection: do not just copy paste from data source, avoid terminology you might not be familiar with Type of study: Avoid generic language in reasoning why the study is observational/experiment, phrase in context of your data Variables manipulated? Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 2 / 39 Announcements Project proposal feedback Scope of inference: generalizability + causation Large sample size does not mean generalizability Population at large : not informative, define target population representative sample generalizable; random assignment causal EDA if using two variables, also include analysis on a single variable (univariate analysis) first if there are outliers and it makes sense to mention what they are (highest grossing movie, etc.), make sure to mention them in your discussion to provide context describing distributions/comparing groups: also discuss shape and spread, not just center avoid repetitive graphs/tabulations summary([dataset]) is rarely useful Misc correlation: association between two numerical variables data plural, dataset singular Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 3 / 39 Review: what purpose does a large sample serve? As long as observations are independent, and the population distribution is not extremely skewed, a large sample would ensure that... the sampling distribution of the mean is nearly normal the estimate of the standard error, as s n, is reliable Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 4 / 39

The normality condition Introducing the t distribution The normality condition The t distribution The CLT, which states that sampling distributions will be nearly normal, holds true for any sample size as long as the population distribution is nearly normal. While this is a helpful special case, it s inherently difficult to verify normality in small data sets. We should exercise caution when verifying the normality condition for small samples. It is important to not only examine the data but also think about where the data come from. For example, ask: would I expect this distribution to be symmetric, and am I confident that outliers are rare? When n is small, and the population standard deviation(σ) is unknown (almost always), the uncertainty of the standard error estimate is addressed by using the t distribution. This distribution also has a bell shape, but its tails are thicker than the normal model s. Therefore observations are more likely to fall beyond two SDs from the mean than under the normal distribution. These extra thick tails are helpful for mitigating the effect of a less reliable estimate for the standard error of the sampling distribution (since n is small) normal t 4 2 0 2 4 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 5 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 6 / 39 Introducing the t distribution Introducing the t distribution The t distribution (cont.) t statistic Always centered at zero, like the standard normal (z) distribution. Has a single parameter: degrees of freedom (df). normal t, df=10 t, df=5 t, df=2 t, df=1 Test statistic for inference on a small sample mean The test statistic for inference on a small sample (n < 30) mean is the T statistic with df = n 1. T df = point estimate null value SE Why is df = n 1, i.e. what does it mean to lose a degree of freedom? 2 0 2 4 6 What happens to shape of the t distribution as df increases? Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 7 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 8 / 39

Example: Friday the 13 th Example: Friday the 13 th Friday the 13 th Friday the 13 th Between 1990-1992 researchers in the UK collected data on traffic flow, accidents, and hospital admissions on Friday 13 th and the previous Friday, Friday 6 th. Below is an excerpt from this data set on traffic flow. We can assume that traffic flow on given day at locations 1 and 2 are independent. type date 6 th 13 th diff location 1 traffic 1990, July 139246 138548 698 loc 1 2 traffic 1990, July 134012 132908 1104 loc 2 3 traffic 1991, September 137055 136018 1037 loc 1 4 traffic 1991, September 133732 131843 1889 loc 2 5 traffic 1991, December 123552 121641 1911 loc 1 6 traffic 1991, December 121139 118723 2416 loc 2 7 traffic 1992, March 128293 125532 2761 loc 1 8 traffic 1992, March 124631 120249 4382 loc 2 9 traffic 1992, November 124609 122770 1839 loc 1 10 traffic 1992, November 117584 117263 321 loc 2 We want to investigate if people s behavior is different on Friday 13 th compared to Friday 6 th. One approach is to compare the traffic flow on these two days. H 0 : Average traffic flow on Friday 6 th and 13 th are equal. H A : Average traffic flow on Friday 6 th and 13 th are different. Each case in the data set represents traffic flow recorded at the same location in the same month of the same year: one count from Friday 6 th and the other Friday 13 th. Are these two counts independent? Scanlon, T.J., Luben, R.N., Scanlon, F.L., Singleton, N. (1993), Is Friday the 13th Bad For Your Health?, BMJ, 307, 1584-1586. Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 9 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 10 / 39 Example: Friday the 13 th Example: Friday the 13 th Hypotheses Conditions What are the hypotheses for testing for a difference between the average traffic flow between Friday 6 th and 13 th? (a) H 0 : µ 6th = µ 13th H A : µ 6th µ 13th (b) H 0 : p 6th = p 13th H A : p 6th p 13th (c) H 0 : µ diff = 0 H A : µ diff 0 (d) H 0 : x diff = 0 H A : x diff = 0 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 11 / 39 Independence: We are told to assume that cases (rows) are independent. Sample size / skew: The sample distribution does not appear to be extremely skewed, but it s very difficult to assess with such a small sample size. We might want to think about whether we would expect the population distribution to be skewed or not probably not, it should be equally likely to have days with lower than average traffic and higher than average traffic. n < 30! So what do we do when the sample size is small? frequency 5 4 3 2 1 0 0 1000 2000 3000 4000 5000 Difference in traffic flow We can use simulation, but when working with small sample means we can also use the t distribution. Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 12 / 39

Application exercise: t test Evaluating hypotheses using the t distribution Use these data to evaluate whether there is a difference in the traffic flow between Friday 6 th and 13 th. Evaluating hypotheses using the t distribution type date 6 th 13 th diff location 1 traffic 1990, July 139246 138548 698 loc 1 2 traffic 1990, July 134012 132908 1104 loc 2 3 traffic 1991, September 137055 136018 1037 loc 1 4 traffic 1991, September 133732 131843 1889 loc 2 5 traffic 1991, December 123552 121641 1911 loc 1 6 traffic 1991, December 121139 118723 2416 loc 2 7 traffic 1992, March 128293 125532 2761 loc 1 8 traffic 1992, March 124631 120249 4382 loc 2 9 traffic 1992, November 124609 122770 1839 loc 1 10 traffic 1992, November 117584 117263 321 loc 2 x diff = 1836 s diff = 1176 n = 10 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 13 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 14 / 39 Evaluating hypotheses using the t distribution Evaluating hypotheses using the t distribution Using technology to find the p-value Using R: t-test for a mean # load data friday = read.csv("http://stat.duke.edu/ mc301/data/friday.csv") # load inference function source("http://stat.duke.edu/ mc301/r/inference.r") The p-value is, once again, calculated as the area tail area under the t distribution. Using R: > 2 * pt(4.94, df = 9, lower.tail = FALSE) [1] 0.0008022394 Using a web applet: http://www.socr.ucla.edu/htmls/socr Distributions.html # test for a mean inference(friday$diff, est = "mean", type = "ht", method = "theoretical", null = 0, alternative = "twosided") Single mean Summary statistics: mean = 1835.8 ; sd = 1176.0139 ; n = 10 H0: mu = 0 HA: mu!= 0 Standard error = 371.8882 Test statistic: T = 4.936 Degrees of freedom: 9 p-value = 8e-04 0 1 2 3 4 5 0 1000 2000 3000 4000 5000 0 friday$diff Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 15 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 16 / 39

Constructing confidence intervals using the t distribution Constructing confidence intervals using the t distribution What is the difference? Confidence interval for a small sample mean Confidence intervals are always of the form We concluded that there is a difference in the traffic flow between Friday 6 th and 13 th. But it would be more interesting to find out what exactly this difference is. We can use a confidence interval to estimate this difference. point estimate ± ME ME is always calculated as the product of a critical value and SE. Since small sample means follow a t distribution (and not a z distribution), the critical value is a t (as opposed to a z ). point estimate ± t SE Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 17 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 18 / 39 Constructing confidence intervals using the t distribution Constructing confidence intervals using the t distribution Finding the critical t (t ) Constructing a CI for a small sample mean 95% 0 t* =? df = 9 n = 10, df = 10 1 = 9, t is at the intersection of row df = 9 and two tail probability 0.05. Which of the following is the correct calculation of a 95% confidence interval for the difference between the traffic flow between Friday 6 th and 13 th? one tail 0.100 0.050 0.025 0.010 0.005 two tails 0.200 0.100 0.050 0.020 0.010 df 6 1.44 1.94 2.45 3.14 3.71 7 1.41 1.89 2.36 3.00 3.50 8 1.40 1.86 2.31 2.90 3.36 9 1.38 1.83 2.26 2.82 3.25 10 1.37 1.81 2.23 2.76 3.17 x diff = 1836 s diff = 1176 n = 10 SE = 372 (a) 1836 ± 1.96 372 (b) 1836 ± 2.26 372 (c) 1836 ± 2.26 372 (d) 1836 ± 2.26 1176 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 19 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 20 / 39

Constructing confidence intervals using the t distribution Interpreting the CI Synthesis Synthesis Which of the following is the best interpretation for the confidence interval we just calculated? Does the conclusion from the hypothesis test agree with the findings of the confidence interval? µdiff :6th 13th = (995, 2677) We are 95% confident that... (a) the difference between the average number of cars on the road on Friday 6th and 13th is between 995 and 2,677. Do the findings of the study suggest that people believe Friday 13th is a day of bad luck? (b) on Friday 6th there are 995 to 2,677 fewer cars on the road than on the Friday 13th, on average. (c) on Friday 6th there are 995 fewer to 2,677 more cars on the road than on the Friday 13th, on average. (d) on Friday 13th there are 995 to 2,677 fewer cars on the road than on the Friday 6th, on average. Statistics 104 (Mine C etinkaya-rundel) U4 - L3: t-distribution October 22, 2013 21 / 39 Synthesis October 22, 2013 22 / 39 Diamonds If n < 30, and σ is unknown, use the t distribution for inference on means. Weights of diamonds are measured in carats. 1 carat = 100 points, 0.99 carats = 99 points, etc. The difference between the size of a 0.99 carat diamond and a 1 carat diamond is undetectable to the naked human eye, but the price of a 1 carat diamond tends to be much higher than the price of a 0.99 diamond. We are going to test to see if there is a difference between the average prices of 0.99 and 1 carat diamonds. In order to be able to compare equivalent units, we divide the prices of 0.99 carat diamonds by 99 and 1 carat diamonds by 100, and compare the average point prices. Hypothesis testing: point estimate null value SE where df = n 1 and SE = U4 - L3: t-distribution Recap: Inference using a small sample mean Tdf = Statistics 104 (Mine C etinkaya-rundel) s n Confidence interval:? point estimate ± tdf SE Note: The example we used was for paired means (difference between dependent groups). We took the difference between the observations and used only these differences (one sample) in our analysis, therefore the mechanics are the same as when we are working with just one sample. Statistics 104 (Mine C etinkaya-rundel) U4 - L3: t-distribution October 22, 2013 23 / 39 Statistics 104 (Mine C etinkaya-rundel) U4 - L3: t-distribution October 22, 2013 24 / 39

Data Parameter and point estimate 80 70 60 50 40 30 Parameter of interest: Average difference between the point prices of all 0.99 carat and 1 carat diamonds. µ pt99 µ pt100 20 carat = 0.99 carat = 1 0.99 carat 1 carat pt99 pt100 x 44.50 53.43 s 13.32 12.22 n 23 30 Point estimate: Average difference between the point prices of sampled 0.99 carat and 1 carat diamonds. x pt99 x pt100 These data are a random sample from the diamonds data set in ggplot2 R package. Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 25 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 26 / 39 Hypotheses Conditions Which of the following is the correct set of hypotheses for testing if the average point price of 1 carat diamonds ( pt100 ) is higher than the average point price of 0.99 carat diamonds ( pt99 )? (a) H 0 : µ pt99 = µ pt100 H A : µ pt99 µ pt100 (b) H 0 : µ pt99 = µ pt100 H A : µ pt99 > µ pt100 (c) H 0 : µ pt99 = µ pt100 H A : µ pt99 < µ pt100 (d) H 0 : x pt99 = x pt100 H A : x pt99 < x pt100 Which of the following does not need to be satisfied in order to conduct this hypothesis test using theoretical methods? (a) Point price of one 0.99 carat diamond in the sample should be independent of another, and the point price of one 1 cars diamond should independent of another as well. (b) Point prices of 0.99 carat and 1 carat diamonds in the sample should be independent. (c) Distributions of point prices of 0.99 and 1 carat diamonds should not be extremely skewed. (d) Both sample sizes should be at least 30. Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 27 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 28 / 39

Sampling distribution for the difference of two means Test statistic Test statistic (cont.) Test statistic for inference on the difference of two small sample means The test statistic for inference on the difference of two small sample means (n 1 < 30 and/or n 2 < 30) mean is the T statistic. where SE = T df = point estimate null value SE s 2 1 n 1 + s2 2 n 2 and df = min(n 1 1, n 2 1) Note: The calculation of the df is actually much more complicated. For simplicity we ll use the above formula to estimate the true df when conducting the analysis by hand. Calculate the test statistic. 0.99 carat 1 carat pt99 pt100 x 44.50 53.43 s 13.32 12.22 n 23 30 T = point estimate null value SE = (44.50 53.43) 0 13.32 2 + 12.222 23 30 = 8.93 3.56 = 2.508 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 29 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 30 / 39 Test statistic (cont.) p-value Which of the following is the correct df for this hypothesis test? (a) 22 (b) 23 (c) 30 (d) 29 (e) 52 Which of the following is the correct p-value for this hypothesis test? (a) between 0.005 and 0.01 (b) between 0.01 and 0.025 (c) between 0.02 and 0.05 (d) between 0.01 and 0.02 T = 2.508 one tail 0.100 0.050 0.025 0.010 two tails 0.200 0.100 0.050 0.020 df 21 1.32 1.72 2.08 2.52 22 1.32 1.72 2.07 2.51 23 1.32 1.71 2.07 2.50 24 1.32 1.71 2.06 2.49 25 1.32 1.71 2.06 2.49 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 31 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 32 / 39

Synthesis Diamond prices What is the conclusion of the hypothesis test? How (if at all) would this conclusion change your behavior if you went diamond shopping? http:// rstudio-pubs-static.s3.amazonaws.com/ 2176 75884214fc524dc0bc2a140573da38bb.html Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 33 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 34 / 39 Using R: t-test for a difference of two means # load data diamond = read.csv("http://stat.duke.edu/ mc301/data/diamond.csv") Equivalent confidence level Confidence intervals for the difference of two means # test for a difference of means inference(diamond$ptprice, diamond$carat, est = "mean", type = "ht", method = "theoretical", null = 0, alternative = "less", order = c("pt99", "pt100")) Response variable: numerical, Explanatory variable: categorical Difference between two means Summary statistics: n_pt99 = 23, mean_pt99 = 44.5061, sd_pt99 = 13.3239 n_pt100 = 30, mean_pt100 = 53.4337, sd_pt100 = 12.2157 Observed difference between means (pt99-pt100) = -8.9276 H0: mu_pt99 - mu_pt100 = 0 HA: mu_pt99 - mu_pt100 < 0 Standard error = 3.563 Test statistic: T = -2.506 Degrees of freedom: 22 p-value = 0.0101 20 40 60 80 What is the equivalent confidence level for a one sided hypothesis test at α = 0.05? (a) 90% (b) 92.5% (c) 95% (d) 97.5% pt99 pt100-8.93 0 diamond$ptprice Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 35 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 36 / 39

Critical value Confidence intervals for the difference of two means What is the appropriate t for a confidence interval for the average difference between the point prices of 0.99 and 1 carat diamonds? Confidence intervals for the difference of two means Application exercise: t interval for comparing means Calculate a 90% confidence interval for the average difference between the point prices of 0.99 and 1 carat diamonds?, and interpret it in context. (a) 1.32 (b) 1.72 (c) 2.07 (d) 2.82 one tail 0.100 0.050 0.025 0.010 0.005 two tails 0.200 0.100 0.050 0.020 0.010 df 21 1.32 1.72 2.08 2.52 2.83 22 1.32 1.72 2.07 2.51 2.82 23 1.32 1.71 2.07 2.50 2.81 24 1.32 1.71 2.06 2.49 2.80 25 1.32 1.71 2.06 2.49 2.79 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 37 / 39 Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 38 / 39 Recap Recap: Inference for difference of two small sample means If n 1 < 30 and/or n 2 < 30 use the t distribution for inference on difference of means. Hypothesis testing: T df = point estimate null value SE where df = min(n 1 1, n 2 1) and SE = Confidence interval: s 2 1 n 1 + s2 2 n 2. point estimate ± t df SE Statistics 104 (Mine Çetinkaya-Rundel) U4 - L3: t-distribution October 22, 2013 39 / 39