Topic 9: Sampling Distributions of Estimators

Similar documents
Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

Stat 421-SP2012 Interval Estimation Section

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

Expectation and Variance of a random variable

Lecture 7: Properties of Random Samples

Stat 200 -Testing Summary Page 1

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes.

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

Statistics 511 Additional Materials

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

MATH/STAT 352: Lecture 15

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

Module 1 Fundamentals in statistics

Stat 319 Theory of Statistics (2) Exercises

Interval Estimation (Confidence Interval = C.I.): An interval estimate of some population parameter is an interval of the form (, ),

Chapter 6 Sampling Distributions

7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals

Random Variables, Sampling and Estimation

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Probability and statistics: basic terms

Common Large/Small Sample Tests 1/55

Properties and Hypothesis Testing

Parameter, Statistic and Random Samples

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

Lecture 5. Materials Covered: Chapter 6 Suggested Exercises: 6.7, 6.9, 6.17, 6.20, 6.21, 6.41, 6.49, 6.52, 6.53, 6.62, 6.63.

Computing Confidence Intervals for Sample Data

Lecture 18: Sampling distributions

Lecture 2: Monte Carlo Simulation

[ ] ( ) ( ) [ ] ( ) 1 [ ] [ ] Sums of Random Variables Y = a 1 X 1 + a 2 X 2 + +a n X n The expected value of Y is:

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

MBACATÓLICA. Quantitative Methods. Faculdade de Ciências Económicas e Empresariais UNIVERSIDADE CATÓLICA PORTUGUESA 9. SAMPLING DISTRIBUTIONS

Tests of Hypotheses Based on a Single Sample (Devore Chapter Eight)

MOST PEOPLE WOULD RATHER LIVE WITH A PROBLEM THEY CAN'T SOLVE, THAN ACCEPT A SOLUTION THEY CAN'T UNDERSTAND.

Simulation. Two Rule For Inverting A Distribution Function

Frequentist Inference

Last Lecture. Wald Test

(7 One- and Two-Sample Estimation Problem )

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

April 18, 2017 CONFIDENCE INTERVALS AND HYPOTHESIS TESTING, UNDERGRADUATE MATH 526 STYLE

Statistical inference: example 1. Inferential Statistics

Chapter 6 Principles of Data Reduction

Basis for simulation techniques

A statistical method to determine sample size to estimate characteristic value of soil parameters

4. Partial Sums and the Central Limit Theorem

t distribution [34] : used to test a mean against an hypothesized value (H 0 : µ = µ 0 ) or the difference

Lecture 3. Properties of Summary Statistics: Sampling Distribution

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Sample Size Determination (Two or More Samples)

Kurskod: TAMS11 Provkod: TENB 21 March 2015, 14:00-18:00. English Version (no Swedish Version)

Overview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

Chapter 13: Tests of Hypothesis Section 13.1 Introduction

7.1 Convergence of sequences of random variables

Chapter 13, Part A Analysis of Variance and Experimental Design

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

Class 23. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Math 152. Rumbos Fall Solutions to Review Problems for Exam #2. Number of Heads Frequency

This is an introductory course in Analysis of Variance and Design of Experiments.

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

Distributions of Functions of. Normal Random Variables Version 27 Jan 2004

Hypothesis Testing. Evaluation of Performance of Learned h. Issues. Trade-off Between Bias and Variance

1.010 Uncertainty in Engineering Fall 2008

Final Examination Solutions 17/6/2010

Statisticians use the word population to refer the total number of (potential) observations under consideration

of the matrix is =-85, so it is not positive definite. Thus, the first

Agreement of CI and HT. Lecture 13 - Tests of Proportions. Example - Waiting Times

1 Inferential Methods for Correlation and Regression Analysis

Distribution of Random Samples & Limit theorems

Asymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values

Because it tests for differences between multiple pairs of means in one test, it is called an omnibus test.

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)

Summary. Recap ... Last Lecture. Summary. Theorem

Exam II Review. CEE 3710 November 15, /16/2017. EXAM II Friday, November 17, in class. Open book and open notes.

Mathacle. PSet Stats, Concepts In Statistics Level Number Name: Date: Confidence Interval Guesswork with Confidence

Bayesian Methods: Introduction to Multi-parameter Models

Chapter 6 Part 5. Confidence Intervals t distribution chi square distribution. October 23, 2008

STATISTICAL INFERENCE

Confidence Level We want to estimate the true mean of a random variable X economically and with confidence.

Comparing Two Populations. Topic 15 - Two Sample Inference I. Comparing Two Means. Comparing Two Pop Means. Background Reading

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn

Binomial Distribution

BIOS 4110: Introduction to Biostatistics. Breheny. Lab #9

7.1 Convergence of sequences of random variables

Statistical Intervals for a Single Sample

STAT431 Review. X = n. n )

Parameter, Statistic and Random Samples

Chapter 8: Estimating with Confidence

UNIT 8: INTRODUCTION TO INTERVAL ESTIMATION

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett

STA Learning Objectives. Population Proportions. Module 10 Comparing Two Proportions. Upon completing this module, you should be able to:

A quick activity - Central Limit Theorem and Proportions. Lecture 21: Testing Proportions. Results from the GSS. Statistics and the General Population

A Confidence Interval for μ

x = Pr ( X (n) βx ) =

TAMS24: Notations and Formulas

1 Review of Probability & Statistics

1 Introduction to reducing variance in Monte Carlo simulations

Big Picture. 5. Data, Estimates, and Models: quantifying the accuracy of estimates.

Transcription:

Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0

Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be derived from the joit distributio of X 1... X. It is called the samplig distributio because it is based o the joit distributio of the radom sample. Give a samplig distributio, we ca calculate the probability that a estimator will ot differ from the parameter θ by more tha a specified amout obtai iterval estimates rather tha poit estimates after we have a sample- a iterval estimate is a radom iterval such that the true parameter lies withi this iterval with a give probability (say 95%). choose betwee to estimators- we ca, for istace, calculate the mea-squared error of the estimator, E θ [( ˆθ θ) 2 ] usig the distributio of ˆθ. Samplig distributios o sample size- bigger samples are more expesive, but estimators based o them have lower variace, so we strike the appropriate trade-off betwee cost ad accuracy. Page 1

Examples: Applicatio: sample size ad precisio 1. What if X i N(θ, 4), ad we wat E( X θ) 2.1? This is simply the variace of X, ad we kow X N(θ, 4/). 4.1 if 40 2. Cosider a radom sample of size from a Uiform distributio o [0, θ], ad the statistic U = max{x 1,..., X }. The CDF of U is give by: 0 if u 0 ( ) F(X) = uθ if 0 < u < θ 1 if u θ We ca ow use this to see how large our sample must be if we wat a certai level of precisio i our estimate for θ. Suppose we wat the probability that our estimate lies withi.1θ for ay level of θ to be bigger tha 0.95: Pr( U θ.1θ) = Pr(θ U.1θ) = Pr(U.9θ) = 1 F(.9θ) = 1 0.9 We wat this to be bigger tha 0.95, or 0.9 0.05. With the LHS decreasig i, we choose log(.05) log(.9) = 28.43. Our miimum sample size is therefore 29. Page 2

Joit distributio of sample mea ad sample variace For a radom sample from a ormal distributio, we kow that the M.L.E.s are the sample mea ad the sample variace 1 (X i X ) 2. We kow that X N(µ, σ2 ) ad ( X i µ σ )2 χ 2 ( sum of squares of stadard ormals) If we replace the populatio mea µ with the sample mea X, the resultig sum of squares, has a χ 2 1 distributio. Theorem: If X 1,... X form a radom sample from a ormal distributio with mea µ ad variace σ 2, the the sample mea X ad the sample variace 1 (X i X ) 2 are idepedet radom variables ad X N(µ, σ2 ) (X i X ) 2 σ 2 χ 2 1 Note: This is oly for ormal samples. Page 3

Applicatio: mea ad variace estimates We have a ormal radom sample ad would like the M.L.E.s of the mea ad stadard deviatio to be withi oe-fifth of a stadard deviatio of the respective parameters, µ ad σ with some threshold probability. Suppose we wat to choose a sample size such that Pr( X µ 1 5 σ) 2 1 If we use Chebyshev s iequality, we get this probability is greater tha 25, so settig this equal to 1 2, we have = 50 Usig the exact distributio of X, Pr( X µ 1 5 σ) = Pr( X µ σ 1 5 ) Sice we ow have a stadard ormal r.v., we kow Pr(Z >.68) =.25, so we eed the smallest greater tha (.68 5) 2 = 11.6, so = 12 (Stata 14: ivormal(.75)=.6745) Now if we wat to determie so that Pr[( X µ 1 5 σ ad ( ˆσ σ 5 1 σ] 1 2 By the previous theorem, X ad ˆσ are idepedet, so the LHS is the product p 1 p 2 = Pr( X µ 1 5 σ)pr( ˆσ σ 1 5 σ) p 1 = Pr( Z 5 ) = 1 2 (1 Φ( 5 ). p 2 = Pr(.8σ < ˆσ < 1.2σ) = Pr(.64 < ˆσ2 σ 2 Sice V = ˆσ2 σ 2 < 1.44) χ 2 1, we ca search over values of to fid oe that gives us a product of probabilities equal to 1 2. For = 21, p 1 =.64 p 2 =.79 so p 1 p 2 =.5. display chi2(20, 30.24)-chi2(20, 13.44) (sice 21*.64=13.44 ad 21*1.44=30.24) Page 4

The t-distributio Let Z N(0, 1), let Y χ 2 v, ad let Z ad Y be idepedet radom variables. The X = Z Yv t v The p.d.f of the t-distributio is give by: f(x; v) = v+1 Γ( 2 ) (1 Γ( v 2 ) + x2 ) ( v+1 2 ) πv v Features of the t-distributio: Oe ca see from the above desity fuctio that the t-desity is symmetric with a maximum value at x = 0. The shape of the desity is similar to that of the stadard ormal (bell-shaped) but with fatter tails. Page 5

Relatio to radom ormal samples RESULT 1: Defie S 2 = (X i X ) 2 The radom variable U = (X µ) S 2 1 t 1 (X µ) σ Proof: We kow that N(0, 1) ad that S2 σ 2 χ 2 1. Dividig the first radom variable by the square root of the secod, divided by its degrees of freedom, the σ i the umerator ad deomiator cacels to obtai U. Implicatio: We caot make statemets about X µ usig the ormal distributio if σ 2 is ukow. This result allows us to use its estimate ˆσ 2 = (X i X ) 2 / sice (X µ) ˆσ/ 1 t 1 RESULT 2 As, U Z N(0, 1) To see why: ad 1 U ca be writte as is close to 1. 1 (X µ) ˆσ t 1. As gets large ˆσ gets very close to σ F 1 (.55) =.129 for t 10,.127 for t 20 ad.126 for the stadard ormal distributio. The differeces betwee these values icreases for higher values of their distributio fuctios (why?) Page 6

Cofidece itervals for the mea Give σ 2, let us see how we ca obtai a iterval estimate for µ, i.e. a iterval which is likely to cotai µ with a pre-specified probability. ( ) Sice (X µ) σ/ N(0, 1), Pr 2 < (X µ) σ/ < 2 =.955 But this evet is equivalet to the evets 2σ < X µ < 2σ ad X 2σ < µ < X + 2σ With kow σ, each of the radom variables X 2σ ad X + 2σ are statistics. Therefore, we have derived a radom iterval withi which the populatio parameter lies with probability.955, i.e. ( Pr X 2σ < µ < X + 2σ ) =.955 = γ Notice that there are may itervals for the same γ, this is the shortest oe. Now, give our sample, our statistics take particular values ad the resultig iterval either cotais or does ot cotai µ. We ca therefore o loger talk about the probability that it cotais µ because the experimet has already bee performed. We say that (x 2σ < µ < x + 2σ ) is a 95.5% cofidece iterval for µ. Alteratively, we may say that µ lies i the above iterval with cofidece γ or that the above iterval is a cofidece iterval for µ with cofidece coefficiet γ Page 7

Cofidece Itervals for meas..examples Example 1: X 1,..., X forms a radom sample from a ormal distributio with ukow µ ad σ 2 = 10. x is foud to be 7.164 with = 40. A 80% cofidece iterval for the mea µ 10 is give by (7.164 1.282 40 ), 7.164 + 1.282 10 40 ) or (6.523, 7.805). The cofidece coefficiet. is.8 Example 2: Let X deote the sample mea of a radom sample of size 25 from a σ distributio with variace 100 ad mea µ. I this case, = 2 ad, makig use of the cetral limit theorem the followig statemet is approximately true: ( Pr 1.96 < (X µ) 2 ) ( ) < 1.96 =.95 or Pr X 3.92 < µ < X + 3.92 =.95 If the sample mea is give by x = 67.53, a approximate 95% cofidece iterval for the sample mea is give by (63.61, 71.45). Example 3: Suppose we are iterested i a cofidece iterval for the mea of a ormal distributio but do ot kow σ 2. We kow that (X µ) ˆσ/ 1 t 1 ad ca use the t-distributio with ( 1) degrees of freedom to costruct our iterval estimate. With = 10, x = 3.22, ˆσ = 1.17, a 95% cofidece iterval is give by (3.22 (2.262)(1.17)/ 9, 3.22 + (2.262)(1.17)/ 9) = (2.34, 4.10) (display ivt(9,.975) gives you 2.262) Page 8

Cofidece Itervals for differeces i meas Let X 1,..., X ad Y 1,..., Y m deote idepedet ormal radom samples. X i N(µ 1, σ 2 ) ad Y i N(µ 2, σ 2 ) respectively. Sample meas ad variaces are X, Ȳ, ˆσ2 1, ˆσ2 2. We kow (usig previous results) that: X ad Ȳ are ormally ad idepedetly distributed with meas µ 1 ad µ 2 ad variaces σ2 σ2 ad m ( X Ȳm) N(µ 1 µ 2, σ2 ˆσ 2 1 σ 2 χ 2 1 ad m ˆσ2 2 σ 2 + σ2 m ) so ( X Ȳm) (µ 1 µ 2 ) N(0, 1) σ 2 + σ2 m χ 2 m 1, so their sum ( ˆσ2 1 + m ˆσ2 2 )/σ2 χ 2 +m 2. Therefore U = ( X Ȳm) (µ 1 µ 2 ) ( ) t +m 2 ˆσ 2 1 +m ˆσ2 2 1 (+m 2) + m 1 Deote the deomiator of U by R. Page 9 Suppose we wat a 95% cofidece iterval for the differece i the meas: ( ) Usig the above t-distributio, we fid a umber b for which Pr b < X < b =.95 The radom iterval ( X Ȳ) br, ( X Ȳ) + br will ow cotai the true differece i meas with 95% probability. A cofidece iterval is ow based o sample values, ( x ȳ m ) ad correspodig sample variaces. Based o the CLT, we ca use the same procedure eve whe our samples are ot ormal.

The F-distributio RESULT : Let Y χ 2 m, Z χ 2, ad let Y ad Z be idepedet radom variables. The F = Y/m Z/ = Y mz has a F-distributio with m ad degrees of freedom. The F-desity is give by: f(x) = Γ( m+ 2 )m m/2 /2 Γ( m 2 )Γ( 2 ) x m/2 1 (mx + ) (m+)/2 I (0, )(x) The parameters m ad are sometimes referred to as the umerator ad deomiator degrees of freedom respectively. While the t-desity is symmetric, the F desity is defied o the positive real umbers ad is skewed to the right. From the defiitio of the F distributio (the ratio of two χ 2 distributios ), it follows that 1 F v1,v 2 = F v2,v 1 So if X F(m, ), 1 X F(, m). The square of a radom variable with a t distributio with degrees of freedom has a F distributio with (1, ) degrees of freedom. Page 10 The F test turs out to be useful i may hypothesis testig problems