Topic 9: Sampling Distributions of Estimators

Similar documents
Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

Stat 421-SP2012 Interval Estimation Section

Expectation and Variance of a random variable

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

Lecture 3. Properties of Summary Statistics: Sampling Distribution

Statistics 511 Additional Materials

Lecture 7: Properties of Random Samples

Random Variables, Sampling and Estimation

Interval Estimation (Confidence Interval = C.I.): An interval estimate of some population parameter is an interval of the form (, ),

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

Properties and Hypothesis Testing

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

MATH/STAT 352: Lecture 15

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes.

Module 1 Fundamentals in statistics

Chapter 6 Sampling Distributions

[ ] ( ) ( ) [ ] ( ) 1 [ ] [ ] Sums of Random Variables Y = a 1 X 1 + a 2 X 2 + +a n X n The expected value of Y is:

(7 One- and Two-Sample Estimation Problem )

Bayesian Methods: Introduction to Multi-parameter Models

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

MBACATÓLICA. Quantitative Methods. Faculdade de Ciências Económicas e Empresariais UNIVERSIDADE CATÓLICA PORTUGUESA 9. SAMPLING DISTRIBUTIONS

7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals

Stat 200 -Testing Summary Page 1

Stat 319 Theory of Statistics (2) Exercises

Statistical inference: example 1. Inferential Statistics

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Lecture 2: Monte Carlo Simulation

Parameter, Statistic and Random Samples

Kurskod: TAMS11 Provkod: TENB 21 March 2015, 14:00-18:00. English Version (no Swedish Version)

Chapter 6 Principles of Data Reduction

1 Inferential Methods for Correlation and Regression Analysis

Computing Confidence Intervals for Sample Data

Simulation. Two Rule For Inverting A Distribution Function

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

Probability and statistics: basic terms

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

t distribution [34] : used to test a mean against an hypothesized value (H 0 : µ = µ 0 ) or the difference

Lecture 18: Sampling distributions

Tests of Hypotheses Based on a Single Sample (Devore Chapter Eight)

Comparing Two Populations. Topic 15 - Two Sample Inference I. Comparing Two Means. Comparing Two Pop Means. Background Reading

Common Large/Small Sample Tests 1/55

This is an introductory course in Analysis of Variance and Design of Experiments.

MOST PEOPLE WOULD RATHER LIVE WITH A PROBLEM THEY CAN'T SOLVE, THAN ACCEPT A SOLUTION THEY CAN'T UNDERSTAND.

Lecture 5. Materials Covered: Chapter 6 Suggested Exercises: 6.7, 6.9, 6.17, 6.20, 6.21, 6.41, 6.49, 6.52, 6.53, 6.62, 6.63.

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

Frequentist Inference

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

Department of Mathematics

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

4. Partial Sums and the Central Limit Theorem

Statistical Theory MT 2008 Problems 1: Solution sketches

A quick activity - Central Limit Theorem and Proportions. Lecture 21: Testing Proportions. Results from the GSS. Statistics and the General Population

Big Picture. 5. Data, Estimates, and Models: quantifying the accuracy of estimates.

Last Lecture. Wald Test

Chapter 6 Part 5. Confidence Intervals t distribution chi square distribution. October 23, 2008

TAMS24: Notations and Formulas

1 Review of Probability & Statistics

STA Learning Objectives. Population Proportions. Module 10 Comparing Two Proportions. Upon completing this module, you should be able to:

Because it tests for differences between multiple pairs of means in one test, it is called an omnibus test.

A statistical method to determine sample size to estimate characteristic value of soil parameters

1.010 Uncertainty in Engineering Fall 2008

Statistical Theory MT 2009 Problems 1: Solution sketches

KLMED8004 Medical statistics. Part I, autumn Estimation. We have previously learned: Population and sample. New questions

April 18, 2017 CONFIDENCE INTERVALS AND HYPOTHESIS TESTING, UNDERGRADUATE MATH 526 STYLE

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

Sampling Distributions, Z-Tests, Power

Hypothesis Testing. Evaluation of Performance of Learned h. Issues. Trade-off Between Bias and Variance

Economics 250 Assignment 1 Suggested Answers. 1. We have the following data set on the lengths (in minutes) of a sample of long-distance phone calls

7.1 Convergence of sequences of random variables

Mathacle. PSet Stats, Concepts In Statistics Level Number Name: Date: Confidence Interval Guesswork with Confidence

Data Analysis and Statistical Methods Statistics 651

Confidence intervals summary Conservative and approximate confidence intervals for a binomial p Examples. MATH1005 Statistics. Lecture 24. M.

Questions and Answers on Maximum Likelihood

Overview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions


Statisticians use the word population to refer the total number of (potential) observations under consideration

Mathematical Notation Math Introduction to Applied Statistics

Class 23. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Agreement of CI and HT. Lecture 13 - Tests of Proportions. Example - Waiting Times

Lecture 23: Minimal sufficiency

Joint Probability Distributions and Random Samples. Jointly Distributed Random Variables. Chapter { }

Chapter 13, Part A Analysis of Variance and Experimental Design

A Confidence Interval for μ

CHAPTER 8 FUNDAMENTAL SAMPLING DISTRIBUTIONS AND DATA DESCRIPTIONS. 8.1 Random Sampling. 8.2 Some Important Statistics

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn

Chapter 13: Tests of Hypothesis Section 13.1 Introduction

Exam II Review. CEE 3710 November 15, /16/2017. EXAM II Friday, November 17, in class. Open book and open notes.

AMS 216 Stochastic Differential Equations Lecture 02 Copyright by Hongyun Wang, UCSC ( ( )) 2 = E X 2 ( ( )) 2

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10

Basis for simulation techniques

Distribution of Random Samples & Limit theorems

Problem Set 4 Due Oct, 12

ENGI 4421 Confidence Intervals (Two Samples) Page 12-01

1 Introduction to reducing variance in Monte Carlo simulations

AP Statistics Review Ch. 8

Distributions of Functions of. Normal Random Variables Version 27 Jan 2004

Transcription:

Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0

Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be derived from the joit distributio of X 1... X. It is called the samplig distributio because it is based o the joit distributio of the radom sample. Give a samplig distributio, we ca calculate the probability that a estimator will ot differ from the parameter θ by more tha a specified amout obtai iterval estimates rather tha poit estimates after we have a sample- a iterval estimate is a radom iterval such that the true parameter lies withi this iterval with a give probability (say 95%). choose betwee to estimators- we ca, for istace, calculate the mea-squared error of the estimator, E θ [( ˆθ θ) 2 ] usig the distributio of ˆθ. Samplig distributios of estimators deped o sample size, ad we wat to kow exactly how the distributio chages as we chage this size so that we ca make the right trade-offs betwee cost ad accuracy. Page 1

Samplig distributios: sample size ad precisio Examples: 1. What if X i N(θ, 4), ad we wat E( X θ) 2.1? This is simply the variace of X, ad we kow X N(θ, 4/). 4.1 if 40 2. Cosider a radom sample of size from a Uiform distributio o [0, θ], ad the statistic U = max{x 1,..., X }. The CDF of U is give by: 0 if u 0 ( ) F(X) = uθ if 0 < u < θ 1 if u θ We ca ow use this to see how large our sample must be if we wat a certai level of precisio i our estimate for θ. Suppose we wat the probability that our estimate lies withi.1θ for ay level of θ to be bigger tha 0.95: Pr( U θ.1θ) = Pr(θ U.1θ) = Pr(U.9θ) = 1 F(.9θ) = 1 0.9 We wat this to be bigger tha 0.95, or 0.9 0.05. With the LHS decreasig i, we choose log(.05) log(.9) = 28.43. Our miimum sample size is therefore 29. Page 2

Joit distributio of sample mea ad sample variace For a radom sample from a ormal distributio, we kow that the M.L.E.s are the sample mea ad the sample variace 1 (X i X ) 2. Also, X N(µ, σ2 ) ( X i µ )2 χ 2 (sice it is the sum of squares of stadard ormal radom variables). σ If we replace the populatio mea µ with the sample mea X, the resultig sum of squares, has a χ 2 1 distributio. Theorem: If X 1,... X form a radom sample from a ormal distributio with mea µ ad variace σ 2, the the sample mea X ad the sample variace 1 (X i X ) 2 are idepedet radom variables ad X N(µ, σ2 ) (X i X ) 2 σ 2 χ 2 1 Note: This is oly for ormal samples. Work through the applicatio of this theorem o p. 475 of your textbook, where you are asked to compute the probability that the sample mea ad sample stadard deviatio of a sample draw from a N(µ, σ 2 ) are withi.2σ of their populatio values. Page 3

The t-distributio Let Z N(0, 1), let Y χ 2 v, ad let Z ad Y be idepedet radom variables. The X = Z Yv t v The p.d.f of the t-distributio is give by: f(x; v) = v+1 Γ( 2 ) (1 Γ( v 2 ) + x2 ) ( v+1 2 ) πv v Features of the t-distributio: Oe ca see from the above desity fuctio that the t-desity is symmetric with a maximum value at x = 0. The shape of the desity is similar to that of the stadard ormal (bell-shaped) but with fatter tails. Page 4

Relatio to radom ormal samples RESULT 1: Defie S 2 = (X i X ) 2 The radom variable U = (X µ) S 2 1 t 1 (X µ) σ Proof: We kow that N(0, 1) ad that S2 σ 2 χ 2 1. Dividig the first radom variable by the square root of the secod, divided by its degrees of freedom, the σ i the umerator ad deomiator cacels to obtai U. Implicatio: We caot make statemets about X µ usig the ormal distributio if σ 2 is ukow. This result allows us to use its estimate ˆσ 2 = (X i X ) 2 / sice (X µ) ˆσ/ 1 t 1 RESULT 2 Give X, Z, Y, as above. As X d Z N(0, 1) To see why: ad 1 U ca be writte as is close to 1. 1 (X µ) ˆσ t 1. As gets large ˆσ gets very close to σ F 1 (.55) =.129 for t 10,.127 for t 20 ad.126 for the stadard ormal distributio. The differeces betwee these values icreases for higher values of their distributio fuctios (why?) Page 5

Cofidece itervals for the mea Give σ 2, let us see how we ca obtai a iterval estimate for µ, i.e. a iterval which is likely to cotai µ with a pre-specified probability. ( ) Sice (X µ) σ/ N(0, 1), Pr 2 < (X µ) σ/ < 2 =.955 But this evet is equivalet to the evets 2σ < X µ < 2σ ad X 2σ < µ < X + 2σ With kow σ, each of the radom variables X 2σ ad X + 2σ are statistics. Therefore, we have derived a radom iterval withi which the populatio parameter lies with probability.955, i.e. ( Pr X 2σ < µ < X + 2σ ) =.955 = γ Notice that there are may itervals for the same γ, this is the shortest oe. Now, give our sample, our statistics take particular values ad the resultig iterval either cotais or does ot cotai µ. We ca therefore o loger talk about the probability that it cotais µ because the experimet has already bee performed. We say that (x 2σ < µ < x + 2σ ) is a 95.5% cofidece iterval for µ. Alteratively, we may say that µ lies i the above iterval with cofidece γ or that the above iterval is a cofidece iterval for µ with cofidece coefficiet γ Page 6

Cofidece Itervals for meas..examples Example 1: X 1,..., X forms a radom sample from a ormal distributio with ukow µ ad σ 2 = 10. x is foud to be 7.164 with = 40. A 80% cofidece iterval for the mea µ 10 is give by (7.164 1.282 40 ), 7.164 + 1.282 10 40 ) or (6.523, 7.805). The cofidece coefficiet. is.8 Example 2: Let X deote the sample mea of a radom sample of size 25 from a σ distributio with variace 100 ad mea µ. I this case, = 2 ad, makig use of the cetral limit theorem the followig statemet is approximately true: ( Pr 1.96 < (X µ) 2 ) ( ) < 1.96 =.95 or Pr X 3.92 < µ < X + 3.92 =.95 If the sample mea is give by x = 67.53, a approximate 95% cofidece iterval for the sample mea is give by (63.61, 71.45). Example 3: Suppose we are iterested i a cofidece iterval for the mea of a ormal distributio but do ot kow σ 2. We kow that (X µ) ˆσ/ 1 t 1 ad ca use the t-distributio with ( 1) degrees of freedom to costruct our iterval estimate. With = 10, x = 3.22, ˆσ = 1.17, a 95% cofidece iterval is give by (3.22 (2.262)(1.17)/ 9, 3.22 + (2.262)(1.17)/ 9) = (2.34, 4.10) (display ivt(9,.975) gives you 2.262) Page 7

Cofidece Itervals for differeces i meas Let X 1,..., X ad Y 1,..., Y m deote idepedet radom samples from two distributios, N(µ 1, σ 2 ) ad N(µ 2, σ 2 ), with sample meas deoted by X, Ȳ ad sample variaces by ˆσ 2 1 ad ˆσ2 2. We ve established that: X ad Ȳ are ormally ad idepedetly distributed with meas µ 1 ad µ 2 ad variaces σ2 ad σ2 m Usig our results o the distributio of liear combiatios of ormally distributed variables, we kow that X Ȳm is ormally distributed with mea µ 1 µ 2 ad variace σ 2 + σ2 m. The radom variable ( X Ȳm) (µ 1 µ 2 ) has a stadard ormal distributio ad will σ 2 + σ2 m form the umerator of the T radom variable that we are goig to use. We also kow that ˆσ2 1 σ 2 ad m ˆσ2 2 σ 2 have χ 2 distributios with ( 1) ad (m 1) degrees of freedom respectively, so their sum ( ˆσ 2 1 + m ˆσ2 2 )/σ2 has a χ 2 distributio with ( + m 2) degrees of freedom ad the radom variable ˆσ 2 1 +m ˆσ2 2 σ 2 (+m 2) ca appear as the deomiator of a radom variable which has a t distributio with ( + m 2) degrees of freedom. Page 8

Cofidece Itervals for differeces i meas..cotd We have therefore established that X = ( X Ȳm) (µ 1 µ 2 ( ) has a t-distributio with ˆσ 2 1 +m ˆσ2 2 (+m 2) 1 + 1 m ( + m 2) degrees of freedom. To simplify otatio, deote the deomiator of the above expressio by R. Give our samples, X 1,..., X ad Y 1,..., Y m, we ca ow costruct cofidece itervals for differeces i the meas of the correspodig populatios, µ 1 µ 2. We do this i the usual way: Suppose we wat a 95% cofidece iterval for the differece i the meas, we fid a umber b such that, usig the t-distributio with ( + m 2) degrees of freedom, ( ) Pr b < X < b =.95 The radom iterval ( X Ȳ) br, ( X Ȳ) + br will ow cotai the true differece i meas with 95% probability. A cofidece iterval is ow based o sample values, ( x ȳ m ) ad correspodig sample variaces. Based o the CLT, we ca use the same procedure eve whe our samples are ot ormal. Page 9