Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Similar documents
Lecture 3. Experiments with a Single Factor: ANOVA Montgomery 3.1 through 3.3

Lecture 3. Experiments with a Single Factor: ANOVA Montgomery 3-1 through 3-3

Lecture 4. Checking Model Adequacy

Single Factor Experiments

Comparison of a Population Means

Chapter 11. Analysis of Variance (One-Way)

Central Limit Theorem ( 5.3)

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

SAS Procedures Inference about the Line ffl model statement in proc reg has many options ffl To construct confidence intervals use alpha=, clm, cli, c

SAS Commands. General Plan. Output. Construct scatterplot / interaction plot. Run full model

INTERVAL ESTIMATION AND HYPOTHESES TESTING

Sociology 6Z03 Review II

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

Lecture 11: Simple Linear Regression

Normal (Gaussian) distribution The normal distribution is often relevant because of the Central Limit Theorem (CLT):

Ch 2: Simple Linear Regression

EXAMINERS REPORT & SOLUTIONS STATISTICS 1 (MATH 11400) May-June 2009

Statistical inference (estimation, hypothesis tests, confidence intervals) Oct 2018

Confidence Intervals, Testing and ANOVA Summary

+ Specify 1 tail / 2 tail

Simple linear regression

Topic 17 - Single Factor Analysis of Variance. Outline. One-way ANOVA. The Data / Notation. One way ANOVA Cell means model Factor effects model

Lecture 10: 2 k Factorial Design Montgomery: Chapter 6

Lecture 3: Inference in SLR

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants.

Table 1: Fish Biomass data set on 26 streams

Linear models and their mathematical foundations: Simple linear regression

Space Telescope Science Institute statistics mini-course. October Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses

Statistical Inference

Math 475. Jimin Ding. August 29, Department of Mathematics Washington University in St. Louis jmding/math475/index.

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

CH.9 Tests of Hypotheses for a Single Sample

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Review of Statistics 101

Course information: Instructor: Tim Hanson, Leconte 219C, phone Office hours: Tuesday/Thursday 11-12, Wednesday 10-12, and by appointment.

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

I i=1 1 I(J 1) j=1 (Y ij Ȳi ) 2. j=1 (Y j Ȳ )2 ] = 2n( is the two-sample t-test statistic.

IENG581 Design and Analysis of Experiments INTRODUCTION

Bias Variance Trade-off

Lecture 12: 2 k Factorial Design Montgomery: Chapter 6

Lecture 15. Hypothesis testing in the linear model

Institute of Actuaries of India

Chapter 1 Linear Regression with One Predictor

STAT 263/363: Experimental Design Winter 2016/17. Lecture 1 January 9. Why perform Design of Experiments (DOE)? There are at least two reasons:

Chapter 8 - Statistical intervals for a single sample

EXST Regression Techniques Page 1. We can also test the hypothesis H :" œ 0 versus H :"

Stat 427/527: Advanced Data Analysis I

Performance Evaluation and Comparison

Math 494: Mathematical Statistics

STT 843 Key to Homework 1 Spring 2018

STAT 525 Fall Final exam. Tuesday December 14, 2010

Lecture 9. ANOVA: Random-effects model, sample size

Stat 704 Data Analysis I Probability Review

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Problem 1 (20) Log-normal. f(x) Cauchy

Multiple comparisons - subsequent inferences for two-way ANOVA

Statistics for Engineers Lecture 9 Linear Regression

Analysis of Variance Bios 662

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

Chapter 7 Comparison of two independent samples

STAT 4385 Topic 01: Introduction & Review

Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests

Lecture 10: Generalized likelihood ratio test

1 Statistical inference for a population mean

On Assumptions. On Assumptions

6 Single Sample Methods for a Location Parameter

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

1 Hypothesis testing for a single mean

ME3620. Theory of Engineering Experimentation. Spring Chapter IV. Decision Making for a Single Sample. Chapter IV

Topic 20: Single Factor Analysis of Variance

1 One-way Analysis of Variance

Outline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013

Lecture 5: Comparing Treatment Means Montgomery: Section 3-5

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS

Chapter Seven: Multi-Sample Methods 1/52

Nonparametric tests. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 704: Data Analysis I

Testing Independence

STAT420 Midterm Exam. University of Illinois Urbana-Champaign October 19 (Friday), :00 4:15p. SOLUTIONS (Yellow)

Exercises and Answers to Chapter 1

Introductory Econometrics

Statistics. Statistics

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Inference for Binomial Parameters

Lectures on Simple Linear Regression Stat 431, Summer 2012

STATISTICS 141 Final Review

Distribution-Free Procedures (Devore Chapter Fifteen)

Preliminary Statistics Lecture 5: Hypothesis Testing (Outline)

STAT 401A - Statistical Methods for Research Workers

Introduction to Business Statistics QM 220 Chapter 12

STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5)

Inferences for Regression

Lecture 7: Latin Square and Related Design

Two-factor studies. STAT 525 Chapter 19 and 20. Professor Olga Vitek

Multivariate Statistical Analysis

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Statistics 512: Applied Linear Models. Topic 9

Introduction to Statistical Inference

Preliminary Statistics. Lecture 3: Probability Models and Distributions

Lec 1: An Introduction to ANOVA

Transcription:

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Fall, 2013 Page 1

Random Variable and Probability Distribution Discrete random variable Y : Finite possible values {y 1, y 2, y 3,..., y k } Probability mass function {p(y 1 ), p(y 2 ),... p(y k )} satisfying p(y i ) 0 and k p(y i ) = 1. i=1 Continuous random variable Y : Possible values form an interval Probability density function f(y) satisfying f(y) 0 and f(y)dy = 1. Fall, 2013 Page 2

Mean and Variance Mean µ = E(Y ): center, location, etc. Variance σ 2 = Var(Y ): spread, dispersion, etc. Discrete Y : µ = k i=1 y ip(y i ); σ 2 = k i=1 (y i µ) 2 p(y i ) Continuous Y : µ = yf(y)dy; σ 2 = (y µ) 2 f(y)dy Formulas for calculating mean and variance If Y 1 and Y 2 are independent, then E(Y 1 Y 2 ) = E(Y 1 )E(Y 2 ) Var(aY 1 ± by 2 ) = a 2 Var(Y 1 ) + b 2 Var(Y 2 ) Other formulas refer to Page 30 (Montgomery, 8th Edition) Fall, 2013 Page 3

Statistical Analysis and Inference: Learn about population from (randomly) drawn data/sample Model and parameter: Assume population (Y ) follows a certain model (distribution) that depends on a set of unknown constants (parameters) denoted by θ: Y f(y, θ). Example 1: Y N(µ, σ 2 ) Y 1 2πσ 2 exp{ (y µ)2 2σ 2 }; where θ = (µ, σ 2 ) Example 2: Y 1 and Y 2 are mean yields of tomato plants fed with fertilizer mixtures A and B, respectively: Y 1 = µ 1 + ϵ 1 ; ϵ 1 N(0, σ1) 2 Y 2 = µ 2 + ϵ 2 ; ϵ 2 N(0, σ2) 2 θ = (µ 1, σ1, 2 µ 2, σ2) 2 Fall, 2013 Page 4

Random sample or observations Random Sample (conceptual) Random Sample (realized) X 1, X 2,..., X n f(x, θ) x 1, x 2,..., x n f(x, θ) Example 1: 0.0 4.9-0.5-1.2 2.1 2.8 1.2 0.8 0.9-0.9 Example 2: A: 19.6 17.9 18.0 20.3 19.3 17.1 16.7 19.2 19.9 19.3 B: 19.6 19.9 21.8 18.4 19.4 21.4 20.5 20.0 18.2 19.9 Fall, 2013 Page 5

Statistical Inference: Estimating Parameter θ Statistics: a statistic is a function of the sample. Y 1,..., Y n : ˆθ = g(y 1, Y 2,..., Y n ) called estimator y 1,..., y n : ˆθ = g(y 1, y 2,..., y n ) called estimate Example 1: Estimators for µ and σ 2 ˆµ = Ȳ = n i=1 Y i n ; ˆσ 2 = S 2 = n i=1 (Y i Ȳ )2 n 1 Estimates ˆµ = ȳ = 1.01; ˆσ 2 = s 2 = 3.49 Fall, 2013 Page 6

Example 2: Estimators: for i = 1, 2. Estimates: ˆµ i = Ȳi = ni j=1 Y ij n i ; ˆσ 2 i = S 2 i = ni j=1 (Y ij Ȳi) 2 n i 1 Assume σ 2 1 = σ 2 2 : ȳ 1 = 18.73; s 2 1 = 1.50; ȳ 2 = 19.91; s 2 2 = 1.30; S 2 pool = (n 1 1)S 2 1 + (n 2 1)S 2 2 n 1 + n 2 2 ; s 2 pool = 1.40 Fall, 2013 Page 7

Statistical Inference: Testing Hypotheses Use test statistics and their distributions to judge hypotheses regarding parameters. H 0 : null hypothesis vs H 1 : alternative hypothesis Example 1: H 0 : µ = 0 vs H 1 : µ 0 Example 2.1: H 0 : µ 2 = µ 1 vs H 1 : µ 2 > µ 1 Example 2.2: H 0 : σ 2 1 = σ 2 2 vs H 1 : σ 2 1 σ 2 2 Details refer to Table 2-3 on Page 47 and Table 2-7 on Page 53 Test statistics: Measures the amount of deviation of estimates from H 0 Example 1: T = Ȳ 0 S/ n H 0 t(n 1); T obs = 1.71 Fall, 2013 Page 8

Example 2: T = (Ȳ2 Ȳ1) 0 H0 t(n 1 + n 2 2); T obs = 2.22 1 S pool n 1 + 1 n 2 Decision Rules Given significance level α, there are two approaches: Compare observed test statistic with critical value Compute the P -value of observed test statistic Reject H 0, if the P -value α. Fall, 2013 Page 9

Statistical Inference: Testing Hypotheses P -value is the probability that test statistic takes on a value that is at least as extreme as the observed value of the statistic when H 0 is true. Extreme in the sense of the alternative hypothesis H 1. Example 1: P value = P(T 1.71 or T 1.71 t(9)) =.12 Conclusion: fail to reject H 0 because 12% 5%. Example 2: P value = P(T 2.22 t(18)) = 0.02 Conclusion: reject H 0 because 2% 5%. Fall, 2013 Page 10

Type I Error, Type II Error and Power of a Decision Rule Type I error: when H 0 is true, reject H 0. α = P (type I error) = P (reject H 0 H 0 is true) Type II error: when H 0 is false, not reject H 0. β = P (type II error) = P (not reject H 0 H 0 is false) Power Power = 1 β = P (reject H 0 H 0 is false) Details refer to Chapter 2, Stat511, etc. In testing hypotheses, we usually control α (the significance level) and prefer decision rules with small β (or high power). Requirements on β (or power) are usually used to calculate necessary sample size. Fall, 2013 Page 11

Statistical Inference: Confidence Intervals: Interval statements regarding parameter θ 100(1-α) percent confidence interval for θ: (L, U ) Both L and U are statistics (calculated from a sample), such that P (L < θ < U) = 1 α Given a real sample x 1, x 2,..., x n, l = L(x 1,..., x n ) and u = U(x 1,..., x n ) lead to a confidence interval (l, u). Question: P (l < θ < u) =? Fall, 2013 Page 12

Example 1. A 95% Confidence Interval for µ: (L, U) = (Y t 0.025 (9) S n, Y + t 0.025 (9) S n ) For the given sample; (l, u) = (1.01 2.26 1.87 10, 1.01 + 2.26 1.87 10 ) = (.33, 2.35) Example 2. A 95% Confidence interval for µ 2 µ 1 : (L, U) = Y 2 Y 1 ± t 0.025 (18)S pool 1/n1 + 1/n 2 (l, u) = (19.91 18.83) ± 2.10 1.18 1/10 + 1/10 = (.07, 2.29) Connection between two-sided hypothesis testing and C.I. If the C.I. contains zero, fail to reject H 0 ; otherwise, reject H 0. Fall, 2013 Page 13

Sampling Distributions Distributions of statistics used in estimation, testing and C.I. construction Random sample: Y 1, Y 2,..., Y n N(µ, σ 2 ) Sample mean Y = (Y 1 + Y 2 + + Y n )/n E(Y ) = E( 1 n Yi ) = 1 n E(Yi ) = 1 n nµ = µ Var(Y ) = Var( 1 n Yi ) = 1 n 2 Var(Y i ) = 1 n 2 nσ2 = σ 2 /n Y follows N(µ, σ2 n ) Fall, 2013 Page 14

The Central Limit Theorem Y 1, Y 2,..., Y n are n independent and identically distributed random variables with E(Y i ) = µ and Var(Y i ) = σ 2. Then Z n = Y µ σ/ n approximately follows the standard normal distribution N(0, 1). Remark 1. Do not need to assume the original population distribution is normal 2. When the population distribution is normal, then Z n exactly follows N(0, 1). Fall, 2013 Page 15

Sampling Distributions: Sample Variance S 2 = (Y 1 Y ) 2 + (Y 2 Y ) 2 + + (Y n Y ) 2 n 1 E(S 2 ) = σ 2 (n 1)S 2 σ 2 = Chi-squared distribution n i=1 (Y i Y ) 2 σ 2 If Z 1, Z 2,..., Z k are i.i.d as N(0, 1), then χ 2 n 1 W = Z 2 1 + Z 2 2 + + Z 2 k follows a Chi-squared distribution with degree of freedom k, denoted by χ 2 k Fall, 2013 Page 16

Density functions of χ 2 k Fall, 2013 Page 17

Sampling Distributions t-distribution: t(k) If Z N(0, 1), W χ 2 k and Z and W independent, then T k = Z W/k follows a t-distribution with d.f. k, i.e., t(k). For example, in t-test: T = Y µ 0 S/ n = n(y µ0 )/σ S2 /σ 2 = Z W/(n 1) t(n 1) Remark: As n goes to infinity, t(n 1) converges to N(0, 1). Fall, 2013 Page 18

Density functions of t(k) distributions Fall, 2013 Page 19

Sampling Distribution F -distributions: F k1,k 2 Suppose random variables W 1 χ 2 k 1, W 2 χ 2 k 2, and W 1 and W 2 are independent, then F = W 1/k 1 W 2 /k 2 follows F k1,k 2 with numerator d.f. k 1 and denominator d.f. k 2. Example: H 0 : σ1 2 = σ2 2, the test statistic is F = S2 1 S 2 2 = S2 1/σ 2 S 2 2 /σ2 = W 1/(n 1 1) W 2 /(n 2 1) F n 1 1,n 2 1 Refer to Section 2.6 for details. Fall, 2013 Page 20

Density functions of F -distributions Fall, 2013 Page 21

Normal Probability Plot used to check if a sample is from a normal distribution Y 1, Y 2,..., Y n is a random sample from a population with mean µ and variance σ 2. Order Statistics: Y (1), Y (2),..., Y (n) where Y (i) is the ith smallest value. if the population is normal, i.e., N(µ, σ 2 ), then E(Y (i) ) µ + σr αi with α i = i 3/8 n+1/4 where r αi is the 100α i th percentile of N(0, 1) for 1 i n. Given a sample y 1, y 2,..., y n, the plot of (r αi, y (i) ) is called the normal probability plot or QQ plot. the points falling around a straight line indicate normality of the population; Deviation from a straight line pattern indicates non-normality (the pen rule) Fall, 2013 Page 22

Example 1 y (i) -1.2-0.9-0.5 0.0 0.8 0.9 1.2 2.1 2.8 4.9 α i.06.16.26.35.45.55.65.74.84.94 r αi -1.6-1.0 -.7 -.4 -.1.1.4.7 1.0 1.6 Note: r αi were obtained from the Z-chart (table) x1-1 0 1 2 3 4 5-1.5-1.0-0.5 0.0 0.5 1.0 1.5 z1 Fall, 2013 Page 23

QQ Plot 1 Quantiles of Standard Normal xf -2-1 0 1 2 0 1 2 3 4 5 Fall, 2013 Page 24

QQ Plot 1 (continued): True Population Distribution 0 2 4 6 8 10 0 1 2 3 4 5 Concave-upward shape indicates right-skewed distn xf Fall, 2013 Page 25

QQ plot 2. Quantiles of Standard Normal xxf -2-1 0 1 2 4 6 8 10 Fall, 2013 Page 26

QQ plot 2 (continued) 0 5 10 15 2 4 6 8 10 Concave-downward shape indicates left-skewed distn xxf Fall, 2013 Page 27

QQ plot 3. Quantiles of Standard Normal xt -2-1 0 1 2-5 0 5 Fall, 2013 Page 28

QQ plot 3 (continued) 0 5 10 15 20 25-5 0 5 flipped S shape indicates a distribution with two heavier tails xt Fall, 2013 Page 29

SAS Code for QQ plot data one; input observation @@; datalines; 0.89 2.79 2.27 2.58 1.72 2.93-0.82-1.40 0.08 1.97 0.84-2.74 2.62 3.48 1.95 2.23 1.02-0.76 0.20-1.69-1.69 0.89 1.98 1.61 0.22 2.60-0.52 0.40 2.71 2.19 ; proc univariate data=one; var observation; histogram observation / normal; qqplot observation /normal (L=1 mu=est sigma=est); run; quit; Fall, 2013 Page 30

Output Fall, 2013 Page 31

Fall, 2013 Page 32

Determine Sample Size Type II error: β=p( fail to reject H 0 H 1 is correct) In testing hypotheses, one first wants to control type I error. If type II error is too large, the conclusion would be too conservative. Example 2 H 0 : µ 2 µ 1 = 0 vs H 1 : µ 2 µ 1 0 Significance level: α = 5% For convenience, we assume two samples have the same size n Decision Rule based on two-sample t-test: reject H 0, if Equivalently Y 2 Y 1 S pool 1/n+1/n > t 0.025 (2n 2) or < t 0.025 (2n 2) fail to reject H 0 if t 0.025 (2n 2) Y 2 Y 1 S pool 1/n+1/n t 0.025 (2n 2) The type I error of the decision rule is 5%, we want to know how large n should be so that the decision rule has type II error less than a threshold, say, 5%. Fall, 2013 Page 33

Recall β = P (type II) = P ( accept H 0 H 1 holds ) Hence β = P ( t 0.025 (2n 2) Y 2 Y 1 S pool 1/n + 1/n t 0.025 (2n 2) H 1 ) Under H 1, the test statistic does not follow t(2n 2), in fact, it follows a noncentral t-distribution with df 2n 2 and noncentral parameter δ = µ 2 µ 1 σ 2/n. Hence β is a function of µ 2 µ 1 /2σ, and n, β = β( µ 2 µ 1 /2σ, n) Fall, 2013 Page 34

Determine Sample Size (continued) Let d = µ 2 µ 1 2σ. So β = β(d, n), which is the probability of type II error when µ 1 and µ 2 are apart by d. Intuitively, the smaller d is, the larger n needs to be such that β 5%. In terms of power (1-β(d, n)). The smaller d is, the larger n needs to be in order to detect µ 1 and µ 2 are different from each other. Suppose we are interested in making the correct decision when µ 1 and µ 2 are apart by at least d = 1 with high probability (power), that is, we want to guarantee the type II error at d = 1, β(1, n) to be small enough, say < 5%. How many data points we need to collect?: Find the smallest n such that β(1, n) < 5% Calculate β(d, n) for d > 0 and fixed n and plot β(d, n) against d, until the smallest n is found. Fall, 2013 Page 35

Case 1: n=4 Fall, 2013 Page 36

Case 2: n=7 Fall, 2013 Page 37

Case 3: n=9 Fall, 2013 Page 38

Operating characteristic Curves Curves of β(d, n) versus d for various given n are called operating characteristic curves, O.C. Curves, which can be used to determine sample size O.C. Curves for two-sided t test (next slide) n = n 1 + n 2 1. From the curves, n 1 + n 2 1 16 If equal sample size is required, then n 1 = n 2 9. O.C. Curves for ANOVA involving fixed effects and random effects are given in Tables V-VI in the Appendix (not required). Fall, 2013 Page 39

O.C. Curves for two-sided t test Fall, 2013 Page 40

SAS code for plotting O.C. Curves data one; n=9;df=2*(n-1);alpha=0.05; do d=0 to 1 by 0.10; nc=d*sqrt(2*n); rlow=tinv(alpha/2,df); rhigh=tinv(1-alpha/2,df); p=probt(rhigh,df,nc)-probt(rlow,df,nc); output; end; proc print data=one; symbol1 v=circle i=sm5; title1 operating characteristic curve ; axis1 label=( prob of accepting H_0 ); axis2 label=( d ); proc gplot; plot p*d/haxis=axis2 vaxis=axis1; run; quit; Fall, 2013 Page 41