Outline. PubH 5450 Biostatistics I Prof. Carlin. Confidence Interval for the Mean. Part I. Reviews

Similar documents
T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

PubH 5450 Biostatistics I Prof. Carlin. Lecture 13

T-test: means of Spock's judge versus all other judges 1 12:10 Wednesday, January 5, judge1 N Mean Std Dev Std Err Minimum Maximum

Introduction to Crossover Trials

Business Statistics. Lecture 10: Course Review

Chapter 7 Comparison of two independent samples

Exam 2 (KEY) July 20, 2009

Lecture 3: Inference in SLR

MBA 605, Business Analytics Donald D. Conant, Ph.D. Master of Business Administration

ECON Introductory Econometrics. Lecture 2: Review of Statistics

Section 9.4. Notation. Requirements. Definition. Inferences About Two Means (Matched Pairs) Examples

df=degrees of freedom = n - 1

Two Sample Problems. Two sample problems

E509A: Principle of Biostatistics. GY Zou

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests

General Linear Model (Chapter 4)

Psychology 282 Lecture #4 Outline Inferences in SLR

Single Sample Means. SOCY601 Alan Neustadtl

An inferential procedure to use sample data to understand a population Procedures

Statistics: CI, Tolerance Intervals, Exceedance, and Hypothesis Testing. Confidence intervals on mean. CL = x ± t * CL1- = exp

Epidemiology Principles of Biostatistics Chapter 10 - Inferences about two populations. John Koval

Nonparametric tests. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 704: Data Analysis I

Lectures 5 & 6: Hypothesis Testing

Chapter 9. Inferences from Two Samples. Objective. Notation. Section 9.2. Definition. Notation. q = 1 p. Inferences About Two Proportions

Basic Statistics. 1. Gross error analyst makes a gross mistake (misread balance or entered wrong value into calculation).

Outline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013

Repeated Measures Part 2: Cartoon data

A discussion on multiple regression models

Lecture 15: Inference Based on Two Samples

STA 101 Final Review

Statistics for Managers Using Microsoft Excel Chapter 9 Two Sample Tests With Numerical Data

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

Business Statistics. Lecture 5: Confidence Intervals

COMPLETELY RANDOM DESIGN (CRD) -Design can be used when experimental units are essentially homogeneous.

Two-Sample Inferential Statistics

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

Medical statistics part I, autumn 2010: One sample test of hypothesis

Chapter 24. Comparing Means

Hypothesis Testing hypothesis testing approach formulation of the test statistic

Stat 529 (Winter 2011) Experimental Design for the Two-Sample Problem. Motivation: Designing a new silver coins experiment

Inference for Binomial Parameters

PLS205 Lab 2 January 15, Laboratory Topic 3

Notes for Week 13 Analysis of Variance (ANOVA) continued WEEK 13 page 1

Chapter 7. Inference for Distributions. Introduction to the Practice of STATISTICS SEVENTH. Moore / McCabe / Craig. Lecture Presentation Slides

Review of Statistics 101

Chapter 9 Inferences from Two Samples

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares

4.1 Hypothesis Testing

7.2 One-Sample Correlation ( = a) Introduction. Correlation analysis measures the strength and direction of association between

Frequency table: Var2 (Spreadsheet1) Count Cumulative Percent Cumulative From To. Percent <x<=

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Statistics for IT Managers

SAS Commands. General Plan. Output. Construct scatterplot / interaction plot. Run full model

Statistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017

Two sample Test. Paired Data : Δ = 0. Lecture 3: Comparison of Means. d s d where is the sample average of the differences and is the

Statistical Methods I

Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model

Chapter 11. Analysis of Variance (One-Way)

ST505/S697R: Fall Homework 2 Solution.

ANOVA Analysis of Variance

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

CBA4 is live in practice mode this week exam mode from Saturday!

Many natural processes can be fit to a Poisson distribution

Chapter 5 Confidence Intervals

HYPOTHESIS TESTING. Hypothesis Testing

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8

Visual interpretation with normal approximation

Hypothesis Testing. Hypothesis: conjecture, proposition or statement based on published literature, data, or a theory that may or may not be true

Chapter 23. Inference About Means

UCLA STAT 251. Statistical Methods for the Life and Health Sciences. Hypothesis Testing. Instructor: Ivo Dinov,

Stat 427/527: Advanced Data Analysis I

The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization.

Ch 2: Simple Linear Regression

Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.

Lecture 18: Simple Linear Regression

Statistics for exp. medical researchers Comparison of groups, T-tests and ANOVA

Comparing two independent samples

LAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2

HYPOTHESIS TESTING II TESTS ON MEANS. Sorana D. Bolboacă

Inference for Distributions Inference for the Mean of a Population

Hypothesis Testing Problem. TMS-062: Lecture 5 Hypotheses Testing. Alternative Hypotheses. Test Statistic

Lecture 17: Small-Sample Inferences for Normal Populations. Confidence intervals for µ when σ is unknown

Nature vs. nurture? Lecture 18 - Regression: Inference, Outliers, and Intervals. Regression Output. Conditions for inference.

Statistics Primer. ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong

Outline for Today. Review of In-class Exercise Bivariate Hypothesis Test 2: Difference of Means Bivariate Hypothesis Testing 3: Correla

Lecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series

SAS Procedures Inference about the Line ffl model statement in proc reg has many options ffl To construct confidence intervals use alpha=, clm, cli, c

Chapter 24. Comparing Means. Copyright 2010 Pearson Education, Inc.

Design of Engineering Experiments Part 2 Basic Statistical Concepts Simple comparative experiments

Lecture 9 Two-Sample Test. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech

Lecture 17. Ingo Ruczinski. October 26, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University

Chapter 7: Statistical Inference (Two Samples)

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI

6 Single Sample Methods for a Location Parameter

Paper Equivalence Tests. Fei Wang and John Amrhein, McDougall Scientific Ltd.

Business Analytics and Data Mining Modeling Using R Prof. Gaurav Dixit Department of Management Studies Indian Institute of Technology, Roorkee

In many situations, there is a non-parametric test that corresponds to the standard test, as described below:

1 Statistical inference for a population mean

Introduction to Econometrics. Review of Probability & Statistics

Transcription:

Outline Outline PubH 5450 Biostatistics I Prof. Carlin Lecture 11 Confidence Interval for the Mean Known σ (population standard deviation): Part I Reviews σ x ± z 1 α/2 n Small n, normal population. Large n, any population. Unknown σ: s x ± t n 1,1 α/2 n Small n, normal population. Large n, near-normal population.

OneSampleTest:H 0 : µ = µ 0 p-values Known σ (population standard deviation), the z-statistic z x µ 0 σ/ n, has a N (0, 1) distribution if the population is normal, or if n is large. Unknown σ, thet-statistic H 1 : µ µ 0 : p = P( Z > z ) orp( T > t ). H 1 : µ>µ 0 : p = P(Z > z) orp(t > t). H 1 : µ<µ 0 : p = P(Z < z) orp(t < t). t x µ 0 s/ n, has a t n 1 distribution if the population is normal, or if n is large and the population is nearly normal. Interpretations Hypothesis First The long-run probability that the 95% CI for the mean covers the true population mean is 95%. If the population mean is truly µ 0, the probability of observing a z or t statistic as extreme or more extreme than yours is the p-value. In both cases, the probabilities refer to relative frequencies in an infinite number of repetitions of the experiment. If the hypothesis depends on the data ( data snooping ), then the requirement for identical experiments is violated, and the interpretation of the p-value becomes difficult/impossible. That is, the p-value will tend to overstate the evidence against a null that was generated by the data itself.

Misinterpretation of the CI Misinterpretations of the p-value For a 95% CI (a, b): The probability that the true population parameter is within (a, b) is 0.95. NO NO NO NO NO NO NO NO NO! If t =2.7 with d.f. = 18, and p =0.01, The probability that the null hypothesis is true is 0.01. The probability of observing z when null hypothesis is true is 0.01. The probability of observing such a difference due to chance is 0.01. The probability of finding a significant result in a replicate experiment is 0.99. p-values and Critical values Part II For H 0 : µ = µ 0,andatwo-sidedH 1, we reject the null when p =Pr ( T > t ) < 0.05, where t is the observed t-statistic. The bigger t is, the smaller the p-value. Alternatively, we can find t 0 > 0suchthat P( T > t 0 )=0.05 andthenrejectthenullif t > t 0. t 0 = t n 1,1 α/2 (what we called t in the CI context) Here, t 0 is called the critical value of the test.

Rejection Region Rejection Region We reject H 0 at significance level α if and only if t = x µ 0 s/ n > t n 1,1 α/2 x µ 0 > t n 1,1 α/2 x >µ 0 + t n 1,1 α/2 s n s n OR x <µ 0 t n 1,1 α/2 s n Definition The rejection region is the range of values of x (or whatever sample statistic we are using) for which H 0 is rejected. CI and Rejection Region CIs and Testing For H 0 : µ = µ 0, and significance level α, We do not reject H 0 if ( x µ 0 t n 1,1 α/2 s, µ 0 + t n n 1,1 α/2 s ). n Alternatively, we do not reject H 0 if ( µ 0 x t n 1,1 α/2 s n, x + t n 1,1 α/2 s ), n i.e., if the null mean value lies inside the 100(1 α)% CI for µ. In many situations, confidence intervals and hypothesis tests are equivalent. The confidence interval is often the non-rejection region of the corresponding hypothesis test. Always report the sample size, relevant sample statistic, and its CI. Report the p-value (and the sampling distribution for the test statistic) when appropriate, rather than just reject or fail to reject at some level α.

Paired Samples Summary Sometimes we want to compare two groups, but only within matched pairs, e.g., measurements of the same subjects before and after they receive some treatment While this initially appears to be a two-sample problem, the pairing and the lack of independence between the samples means that the best approach here is to first convert the problem into a one-sample problem by taking the differences of the pairs. SeeM&MExample7.7! If (X i, Y i ) are the paired samples, define D i = X i Y i,and construct a CI for the true mean difference µ D = µ X µ Y using our usual one-sample CI techniques. Or, test the hypothesis H 0 : µ x = µ Y µ D =0 using the usual one-sample test techniques! For paired samples, the sample size is the number of pairs, not the total number of data points A common mistake is treating paired two-sample problems as independent two-sample problems a big difference! In SAS: PROC TTEST using PAIRED statement PROC TTEST (V8.0+) data radon2 ; input before after @@; cards ; 91.9 97.8 111.4 122.3 105.4 95.0 103.8 99.6 96.6 119.3 104.8 101.7 ; proc ttest data = radon2 alpha = 0.05; paired before after ; run ; Statistics Lower CL Upper CL Lower CL Difference N Mean Mean Mean Std Dev before - after 6-16.27-3.633 9.0045 7.517 Statistics Upper CL Difference Std Dev Std Dev Std Err Minimum Maximum before - after 12.043 29.536 4.9163-22.7 10.4

PROC TTEST (V8.0+) One-Sample Test T-Tests Difference DF t Value Pr > t before - after 5-0.74 0.4931 data radondiff ; set radon2; diff = after before ; proc ttest data = radondiff h0 = 0.0 alpha = 0.05; var diff ; run ; Two-sample problems Part III The goal of inference is to compare the responses in two groups. Each group is considered to be a sample from a distinct distribution. The responses in each group are independent of those in the other group.

Two-sample z-statistic Unknown but Definition Ifasampleofsizen 1 is drawn from N (µ 1,σ1 2 ) and a sample of size n 2 is drawn from N (µ 2,σ2 2), suppose x 1 and x 2 are the sample means for each sample. Then the two-sample z statistic ) ( ) ( x1 x 2 µ1 µ 2 z = σ 2 1 n 1 + σ2 2 n 2 has the standard normal N (0, 1) distribution. When the population variances σ1 2 and σ2 2 are not known, the situation is a bit more complex. If the two populations can be assumed to have the same variance (i.e., σ1 2 = σ2 2 = σ2 ), then we can use the pooled estimator of σ 2, sp 2 = (n 1 1)s1 2 +(n 2 1)s2 2 n 1 + n 2 2 n1 i=1 = (x 1i x 1 ) 2 + n 2 i=1 (x 2i x 2 ) 2. n 1 + n 2 2 [Crummy] F -test for equal variances Pooled-sample t-test We can use the F -statistic, F = s2 1 s 2 2 to test for the equality of the two population variances, i.e., H 0 : σ2 1 σ 2 2 = 1. The alternative is H 1 : σ2 1 σ 2 2 1. F has an F -distribution with (n 1 1, n 2 1) degrees of freedom. Stat software is available to give p-values for this F -test... BUT the F is so nonrobust to (easily influenced by) departures from normality, it is rarely used in practice! (see M&M p.556) Definition When the two populations have the same variance, we can test H 0 : µ 1 = µ 2 using the pooled-sample t-statistic where s p = t = ( x 1 x 2 ) s p 1 n 1 + 1 n 2 s 2 p, the pooled standard error estimate defined above.

Pooled-sample t-test Confidence Interval Under H 0 : µ 1 = µ 2, the pooled sample t statistic has a t distribution with n 1 + n 2 2 degrees of freedom. For a two-sided alternative H 1 : µ 1 µ 2,thep-value is once again SeeM&MExample7.20 p = P( T > t ). Basedonsimilarargumentasinone-samplecase,wecanalso define a 100(1 α)% CI for the difference of the two population means: ( x 1 x 2 ) ± t s p 1 n 1 + 1 n 2 where t = t n1 +n 2 2,1 α/2 and s p is again the pooled standard error estimate. SeeM&MExample7.21. Unknown and For the more general case when the variances are not equal, the two-sample t-statistic is defined as: t = ( x 1 x 2 ) (1) s 2 1 n 1 + s2 2 n 2 which does not quite have a t-distribution under H 0 : µ 1 = µ 2... Do you still want to test for µ 1 = µ 2 when the variances are unequal? If you still do, we can use a t distribution to approximate the actual sampling distribution of the two-sample t-statistic. The degrees of freedom is not n 1 + n 2 2. It is smaller, and there are several possible formulae for calculating it. Typically we use either the smaller of n1 1andn 2 2, or the ludicrous formula on M&M p.536 Note that smaller df generally corresponds to wider confidence interval or bigger critical value (i.e., harder to reject H 0 ).

M&M Examples 7.14, 7.18!