Lecture 8. October 22, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Similar documents
Lecture 4. August 24, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Lecture Notes 3 Convergence (Chapter 5)

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

X = X X n, + X 2

Mathematical statistics

One-sample categorical data: approximate inference

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20

Lecture 2. October 21, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

. Find E(V ) and var(v ).

MAS113 Introduction to Probability and Statistics

Discrete Probability Refresher

Review of Probability. CS1538: Introduction to Simulations

COMP2610/COMP Information Theory

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

3 Multiple Discrete Random Variables

Quick Tour of Basic Probability Theory and Linear Algebra

Chernoff Bounds. Theme: try to show that it is unlikely a random variable X is far away from its expectation.

6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips

MATH 151, FINAL EXAM Winter Quarter, 21 March, 2014

Maximum-Likelihood Estimation: Basic Ideas

Lecture 2 Sep 5, 2017

Basic Probability. Introduction

COMPSCI 240: Reasoning Under Uncertainty

CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions)

Petter Mostad Mathematical Statistics Chalmers and GU

Confidence Intervals Unknown σ

Kousha Etessami. U. of Edinburgh, UK. Kousha Etessami (U. of Edinburgh, UK) Discrete Mathematics (Chapter 7) 1 / 13

E X A M. Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours. Number of pages incl.

CSE 103 Homework 8: Solutions November 30, var(x) = np(1 p) = P r( X ) 0.95 P r( X ) 0.

This does not cover everything on the final. Look at the posted practice problems for other topics.

Lecture 4: September Reminder: convergence of sequences

1 Basic continuous random variable problems

Lecture 2: Repetition of probability theory and statistics

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf

Expectation is linear. So far we saw that E(X + Y ) = E(X) + E(Y ). Let α R. Then,

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

The Central Limit Theorem

Lecture 2: Review of Basic Probability Theory

MAS223 Statistical Inference and Modelling Exercises

Introduction to Probability

Interval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean

1 Basic continuous random variable problems

Evaluating Hypotheses

CSE 312 Final Review: Section AA

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

Mathematical statistics

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

Statistics and Econometrics I

Lecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality

MATH Solutions to Probability Exercises

Lecture 10. Variance and standard deviation

simple if it completely specifies the density of x

Algorithms for Uncertainty Quantification

EECS 70 Discrete Mathematics and Probability Theory Fall 2015 Walrand/Rao Final

Statistics: Learning models from data

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Fundamental Tools - Probability Theory IV

Lecture 1: Review on Probability and Statistics

University of Regina. Lecture Notes. Michael Kozdron

7.1 Basic Properties of Confidence Intervals

Topic 7: Convergence of Random Variables

IEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 2

Statistics 3858 : Maximum Likelihood Estimators

EE514A Information Theory I Fall 2013

CS 361: Probability & Statistics

1 Review of The Learning Setting

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants

Stochastic Processes and Monte-Carlo Methods. University of Massachusetts: Spring 2018 version. Luc Rey-Bellet

Business Statistics. Lecture 3: Random Variables and the Normal Distribution

Distributions of linear combinations

Joint Distribution of Two or More Random Variables

BIO5312 Biostatistics Lecture 6: Statistical hypothesis testings

8 Laws of large numbers

Lecture 5. October 21, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

IEOR E4703: Monte-Carlo Simulation

Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing

Lecture 4: Sampling, Tail Inequalities

STA 111: Probability & Statistical Inference

i=1 X i/n i=1 (X i X) 2 /(n 1). Find the constant c so that the statistic c(x X n+1 )/S has a t-distribution. If n = 8, determine k such that

Expectation of Random Variables

A Brief Review of Probability, Bayesian Statistics, and Information Theory

Review. December 4 th, Review

Lecture 6 - Random Variables and Parameterized Sample Spaces

Expectation, inequalities and laws of large numbers

1 Probability theory. 2 Random variables and probability theory.

Probability Density Functions and the Normal Distribution. Quantitative Understanding in Biology, 1.2

Continuous Distributions

ECE 450 Homework #3. 1. Given the joint density function f XY (x,y) = 0.5 1<x<2, 2<y< <x<4, 2<y<3 0 else

Discrete Random Variables

1 Hypothesis testing for a single mean

Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices.

Econ 113. Lecture Module 2

7 Random samples and sampling distributions

Class 26: review for final exam 18.05, Spring 2014

Bias Variance Trade-off

Chapter 7: Theoretical Probability Distributions Variable - Measured/Categorized characteristic

Some Basic Concepts of Probability and Information Theory: Pt. 2

Probability reminders

Transcription:

Lecture 8 Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University October 22, 2007

1 2 3 4 5 6

1 Define convergent series 2 Define the Law of Large Numbers 3 Define the Central Limit Theorem 4 Create Wald confidence using the

Numerical limits Imagine a sequence a 1 =.9, a 2 =.99, a 3 =.999,... Clearly this sequence converges to 1 Definition of a limit: For any fixed distance we can find a point in the sequence so that the sequence is closer to the limit than that distance from that point on a n 1 = 10 n

of random variables The problem is harder for random variables Consider X n the sample average of the first n of a collection of iid observations Example X n could be the average of the result of n coin flips (i.e. the sample proportion of heads) We say that X n converges in probability to a limit if for any fixed distance the probability of X n being closer (further away) than that distance from the limit converges to one (zero) P( X n limit < ɛ) 1

The Law of Large Numbers Establishing that a random sequence converges to a limit is hard Fortunately, we have a theorem that does all the work for us, called the Law of Large Numbers The law of large numbers states that if X 1,... X n are iid from a population with mean µ and variance σ 2 then X n converges in probability to µ (There are many variations on the ; we are using a particularly lazy one)

Proof using Chebyshev s inequality Recall Chebyshev s inequality states that the probability that a random variable variable is more than k standard deviations from its mean is less than 1/k 2 Therefore for the sample mean P { X n µ k sd( X n ) } 1/k 2 Pick a distance ɛ and let k = ɛ/sd( X n ) P( X n µ ɛ) sd( X n ) 2 ɛ 2 = σ2 nɛ 2

average 0.6 0.2 0.0 0.2 0 20 40 60 80 100 iteration

Useful facts Functions of convergent random sequences converge to to the function evaluated at the limit This includes sums, products, differences,... Example ( X n ) 2 converges to µ 2 Notice that this is different than ( Xi 2 )/n which converges to E[Xi 2 ] = σ 2 + µ 2 We can use this to prove that the sample variance converges to σ 2

Continued (Xi X n ) 2 /(n 1) = = X 2 i n 1 n( X n ) 2 n 1 n n 1 X 2 i n n n 1 ( X n ) 2 p 1 (σ 2 + µ 2 ) 1 µ 2 = σ 2 Hence we also know that the sample standard deviation converges to σ

Discussion An estimator is consistent if it converges to what you want to estimate The basically states that the sample mean is consistent We just showed that the sample variance and the sample standard deviation are consistent as well Recall also that the sample mean and the sample variance are unbiased as well (The sample standard deviation is not unbiased, by the way)

The Central Limit Theorem The Central Limit Theorem () is one of the most important theorems in statistics For our purposes, the states that the distribution of averages of iid variables, properly normalized, becomes that of a standard normal as the sample size increases The applies in an endless variety of settings

The Let X 1,..., X n be a collection of iid random variables with mean µ and variance σ 2 Let X n be their sample average Then P ( ) X n µ σ/ n z Φ(z) Notice the form of the normalized quantity X n µ σ/ n Estimate Mean of estimate =. Std. Err. of estimate

Example Simulate a standard normal random variable by rolling n (six sided) Let X i be the outcome for die i Then note that µ = E[X i ] = 3.5 Var(X i ) = 2.92 SE 2.92/n = 1.71/ n Standardized mean X n 3.5 1.71/ n

1 die rolls 2 die rolls 6 die rolls 0.0 0.2 0.4 0.0 0.2 0.4 0.0 0.2 0.4 3 1 1 2 3 3 1 1 2 3 3 1 1 2 3

Coin Let X i be the 0 or 1 result of the i th flip of a possibly unfair coin The sample proportion, say ˆp, is the average of the coin flips E[X i ] = p and Var(X i ) = p(1 p) Standard error of the mean is p(1 p)/n Then ˆp p p(1 p)/n will be approximately normally distributed

1 coin flips 10 coin flips 20 coin flips 3 1 1 2 3 0.0 0.2 0.4 0.0 0.2 0.4 3 1 1 2 3 0.0 0.2 0.4 3 1 1 2 3 1 coin flips 10 coin flips 20 coin flips 0.0 0.2 0.4 0.0 0.2 0.4 0.0 0.2 0.4 3 1 1 2 3 3 1 1 2 3 3 1 1 2 3

in practice In practice the is mostly useful as an approximation ( ) Xn µ P σ/ n z Φ(z). Recall 1.96 is a good approximation to the.975 th quantile of the standard normal Consider.95 P ( 1.96 X ) n µ σ/ n 1.96 = P ( X n + 1.96σ/ n µ X n 1.96σ/ n ),

Therefore, according to the, the probability that the random interval X n ± z 1 α/2 σ/ n contains µ is approximately 95%, where z 1 α/2 is the 1 α/2 quantile of the standard normal distribution This is called a 95% confidence interval for µ Slutsky s theorem, allows us to replace the unknown σ with s

Sample proportions In the event that each X i is 0 or 1 with common success probability p then σ 2 = p(1 p) The interval takes the form ˆp ± z 1 α/2 p(1 p) n Replacing p by ˆp in the standard error results in what is called a Wald confidence interval for p Also note that p(1 p) 1/4 for 0 p 1 Let α =.05 so that z 1 α/2 = 1.96 2 then 2 p(1 p) n 1 2 4n = 1 n Therefore ˆp ± 1 n is a quick CI estimate for p