The Central Limit Theorem

Similar documents
The Central Limit Theorem

MAT Mathematics in Today's World

The normal distribution

ACMS Statistics for Life Sciences. Chapter 13: Sampling Distributions

STA Why Sampling? Module 6 The Sampling Distributions. Module Objectives

Two-sample inference: Continuous data

Lecture 8 Sampling Theory

Chapter 18 Sampling Distribution Models

Chapter 18. Sampling Distribution Models. Copyright 2010, 2007, 2004 Pearson Education, Inc.

Descriptive statistics

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Example. If 4 tickets are drawn with replacement from ,

Sampling Distribution Models. Chapter 17

Lecture 7: Confidence interval and Normal approximation

STA Module 4 Probability Concepts. Rev.F08 1

Two-sample inference: Continuous data

The Central Limit Theorem

The normal distribution

Binomial and Poisson Probability Distributions

MA 1125 Lecture 33 - The Sign Test. Monday, December 4, Objectives: Introduce an example of a non-parametric test.

Chapter 18. Sampling Distribution Models /51

MA 1125 Lecture 15 - The Standard Normal Distribution. Friday, October 6, Objectives: Introduce the standard normal distribution and table.

Two-sample Categorical data: Testing

Chapter 15 Sampling Distribution Models

Sampling Distribution Models. Central Limit Theorem

CS 361: Probability & Statistics

Random processes. Lecture 17: Probability, Part 1. Probability. Law of large numbers

Lecture 4: Random Variables and Distributions

One-sample categorical data: approximate inference

Combinatorial Analysis

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014

Machine Learning for Signal Processing Expectation Maximization Mixture Models. Bhiksha Raj 27 Oct /

Announcements. Lecture 5: Probability. Dangling threads from last week: Mean vs. median. Dangling threads from last week: Sampling bias

Objective - To understand experimental probability

STA Module 8 The Sampling Distribution of the Sample Mean. Rev.F08 1

Binomial Distribution. Collin Phillips

Conditional Probability

13-5 Probabilities of Independent and Dependent Events

MATH 1150 Chapter 2 Notation and Terminology

PHP2510: Principles of Biostatistics & Data Analysis. Lecture X: Hypothesis testing. PHP 2510 Lec 10: Hypothesis testing 1

STAT/SOC/CSSS 221 Statistical Concepts and Methods for the Social Sciences. Random Variables

SIMPLE RANDOM WALKS: IMPROBABILITY OF PROFITABLE STOPPING

CSE 103 Homework 8: Solutions November 30, var(x) = np(1 p) = P r( X ) 0.95 P r( X ) 0.

Statistics Primer. A Brief Overview of Basic Statistical and Probability Principles. Essential Statistics for Data Analysts Using Excel

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

COS 424: Interacting with Data. Lecturer: Dave Blei Lecture #2 Scribe: Ellen Kim February 7, 2008

Parameter Learning With Binary Variables

Sections 5.1 and 5.2

P (A) = P (B) = P (C) = P (D) =

Study and research skills 2009 Duncan Golicher. and Adrian Newton. Last draft 11/24/2008

A primer on Bayesian statistics, with an application to mortality rate estimation

MATH 151, FINAL EXAM Winter Quarter, 21 March, 2014

RVs and their probability distributions

Math 243 Section 3.1 Introduction to Probability Lab

Last few slides from last time

Probability Experiments, Trials, Outcomes, Sample Spaces Example 1 Example 2

Math/Stat 352 Lecture 10. Section 4.11 The Central Limit Theorem

Ma/CS 6b Class 15: The Probabilistic Method 2

Probability Distributions

Statistics, continued

Statistics and Sampling distributions

CHAPTER 18 SAMPLING DISTRIBUTION MODELS STAT 203

Discrete Probability

1 Random variables and distributions

Probability Distributions

STAT 201 Chapter 5. Probability

Chapter 7: Hypothesis Testing

Chapter 2: Probability Part 1

CS 543 Page 1 John E. Boon, Jr.

Statistical Inference for Means

Chapter 8: An Introduction to Probability and Statistics

Two-sample Categorical data: Testing

Probabilities and distributions

*Karle Laska s Sections: There is no class tomorrow and Friday! Have a good weekend! Scores will be posted in Compass early Friday morning

Central Limit Theorem and the Law of Large Numbers Class 6, Jeremy Orloff and Jonathan Bloom

UNIT NUMBER PROBABILITY 6 (Statistics for the binomial distribution) A.J.Hobson

Probability, For the Enthusiastic Beginner (Exercises, Version 1, September 2016) David Morin,

Permutation tests. Patrick Breheny. September 25. Conditioning Nonparametric null hypotheses Permutation testing

MAT2377. Ali Karimnezhad. Version September 9, Ali Karimnezhad

Stat 20 Midterm 1 Review

Scientific Measurement

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

Conditional Probability, Independence and Bayes Theorem Class 3, Jeremy Orloff and Jonathan Bloom

Markov Chains and Pandemics

Statistics, Probability Distributions & Error Propagation. James R. Graham

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary

Discussion 01. b) What is the probability that the letter selected is a vowel?

Lecture 1: Description of Data. Readings: Sections 1.2,

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

University of Jordan Fall 2009/2010 Department of Mathematics

Mathematical Foundations of Computer Science Lecture Outline October 18, 2018

Essentials of Statistics and Probability

Please do NOT write in this box. Multiple Choice Total

Name: Firas Rassoul-Agha

Contents. Decision Making under Uncertainty 1. Meanings of uncertainty. Classical interpretation

Bayesian Updating with Continuous Priors Class 13, Jeremy Orloff and Jonathan Bloom

The t-test Pivots Summary. Pivots and t-tests. Patrick Breheny. October 15. Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/18

Probability and Discrete Distributions

Supporting Australian Mathematics Project. A guide for teachers Years 11 and 12. Probability and statistics: Module 25. Inference for means

Mean/Average Median Mode Range

Transcription:

The Central Limit Theorem Patrick Breheny March 1 Patrick Breheny STA 580: Biostatistics I 1/23

Kerrich s experiment A South African mathematician named John Kerrich was visiting Copenhagen in 1940 when Germany invaded Denmark Kerrich spent the next five years in an interment camp To pass the time, he carried out a series of experiments in probability theory One of them involved flipping a coin 10,000 times Patrick Breheny STA 580: Biostatistics I 2/23

The law of averages We know that a coin lands heads with probability 50% Thus, after many tosses, the law of averages says that the number of heads should be about the same as the number of tails...... or does it? Patrick Breheny STA 580: Biostatistics I 3/23

Kerrich s results Number of Number of Heads - tosses (n) heads 0.5 Tosses 10 4-1 100 44-6 500 255 5 1,000 502 2 2,000 1,013 13 3,000 1,510 10 4,000 2,029 29 5,000 2,533 33 6,000 3,009 9 7,000 3,516 16 8,000 4,034 34 9,000 4,538 38 10,000 5,067 67 Patrick Breheny STA 580: Biostatistics I 4/23

Kerrich s results plotted Number of heads minus half the number of tosses 50 0 50 10 50 100 500 1000 5000 10000 Number of tosses Instead of getting closer, the numbers of heads and tails are getting farther apart Patrick Breheny STA 580: Biostatistics I 5/23

Repeating the experiment 50 times This is no fluke: Number of heads minus half the number of tosses 100 50 0 50 100 10 50 100 500 1000 5000 10000 Number of tosses Patrick Breheny STA 580: Biostatistics I 6/23

Where s the law of averages? So where s the law of averages? Well, the law of averages does not say that as n increases the number of heads will be close to the number of tails What it says instead is that, as n increases, the average number of heads will get closer and closer to the long-run average (in this case, 0.5) The technical term for this is that the sample average, which is an estimate, converges to the expected value, which is a parameter Patrick Breheny STA 580: Biostatistics I 7/23

Repeating the experiment 50 times, Part II Percentage of heads 0 20 40 60 80 100 10 50 100 500 1000 5000 10000 Number of tosses Patrick Breheny STA 580: Biostatistics I 8/23

Trends in Kerrich s experiment Trend #1: The expected value of the average Trend #2: The standard error Trend #3: The distribution of the average There are three very important trends going on in this experiment These trends can be observed visually from the computer simulations or proven via the binomial distribution We ll work with both approaches so that you can get a sense of they both work and reinforce each other Before we do so, I ll introduce two additional, important facts about the binomial distribution: its mean (expected value) and standard deviation Patrick Breheny STA 580: Biostatistics I 9/23

Trend #1: The expected value of the average Trend #2: The standard error Trend #3: The distribution of the average The expected value of the binomial distribution If we flip a coin once, how many heads do we expect? How about twice? Three times? n times? In each of those situations, what do we expect the average to be? Tosses 1 2 3 n Expected # of heads 0.5 1 1.5 np Expected average 0.5 0.5 0.5 p The expected value of the binomial distribution is np The expected value of a percent (the average of a binomial distribution) is p Patrick Breheny STA 580: Biostatistics I 10/23

The standard error of the mean Trend #1: The expected value of the average Trend #2: The standard error Trend #3: The distribution of the average It can be shown that the standard deviation of the binomial distribution is np(1 p) What will the variability (standard deviation) be for one flip? Two? n? What about the variability of the average? Tosses 1 2 3 n SD (# of heads) 0.5 0.71 0.87 np(1 p) SD (average) 0.5 0.35 0.29 p(1 p)/n Patrick Breheny STA 580: Biostatistics I 11/23

Standard errors The law of averages Trend #1: The expected value of the average Trend #2: The standard error Trend #3: The distribution of the average Note that, as n goes up, the variability of the # of heads goes up, but the variability of the average goes down just as we saw in our simulation Indeed, the variability goes to 0 as n gets larger and larger this is the law of averages The standard deviation of the average is given a special name in statistics to distinguish it from the sample standard deviation of the data The standard deviation of the average is called the standard error The term standard error can also be applied to mean the variability of any estimate, to distinguish it from the variability of individual tosses or people Patrick Breheny STA 580: Biostatistics I 12/23

The square root law The law of averages Trend #1: The expected value of the average Trend #2: The standard error Trend #3: The distribution of the average The relationship between the variability of a individual (toss) and the variability of the average (of a large number of tosses) is a very important relationship, sometimes called the square root law: SE = SD n, where SE is the standard error of the mean and SD is the standard deviation of an individual (toss) We showed that this is true for tosses of a coin, but it is in fact true for all averages Once again, we see this phenomenon visually in our simulation results Patrick Breheny STA 580: Biostatistics I 13/23

The distribution of the mean Trend #1: The expected value of the average Trend #2: The standard error Trend #3: The distribution of the average Finally, let s look at the distribution of the mean by creating histograms of the mean in our simulation 2 flips 7 flips 25 flips Frequency 0 10000 20000 30000 40000 50000 Frequency 0 5000 10000 15000 20000 25000 Frequency 0 5000 10000 15000 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.2 0.4 0.6 0.8 Mean Mean Mean Patrick Breheny STA 580: Biostatistics I 14/23

The theorem How large does n have to be? In summary, there are three very important phenomena going on here concerning the sampling distribution of the sample average: #1 The expected value is always equal to the population average #2 The standard error is always equal to the population standard deviation divided by the square root of n #3 As n gets larger, the sampling distribution looks more and more like the normal distribution Furthermore, these three properties of the sampling distribution of the sample average hold for any distribution not just the binomial Patrick Breheny STA 580: Biostatistics I 15/23

(cont d) The theorem How large does n have to be? This result is called the central limit theorem, and it is one of the most important, remarkable, and powerful results in all of statistics In the real world, we rarely know the distribution of our data But the central limit theorem says: we don t have to Patrick Breheny STA 580: Biostatistics I 16/23

(cont d) The theorem How large does n have to be? Furthermore, as we have seen, knowing the mean and standard deviation of a distribution that is approximately normal allows us to calculate anything we wish to know with tremendous accuracy and the sampling distribution of the mean is always approximately normal The only caveats: Observations must be independently drawn from and representative of the population applies to the sampling distribution of the mean not necessarily to the sampling distribution of other statistics How large does n have to be before the distribution becomes close enough in shape to the normal distribution? Patrick Breheny STA 580: Biostatistics I 17/23

How large does n have to be? The theorem How large does n have to be? Rules of thumb are frequently recommended that n = 20 or n = 30 is large enough to be sure that the central limit theorem is working There is some truth to such rules, but in reality, whether n is large enough for the central limit theorem to provide an accurate approximation to the true sampling distribution depends on how close to normal the population distribution is If the original distribution is close to normal, n = 2 might be enough If the underlying distribution is highly skewed or strange in some other way, n = 50 might not be enough Patrick Breheny STA 580: Biostatistics I 18/23

Example #1 The law of averages The theorem How large does n have to be? Population n=10 Density 0.00 0.05 0.10 0.15 0.20 Density 0.0 0.1 0.2 0.3 0.4 0.5 6 4 2 0 2 4 6 x 3 2 1 0 1 2 3 Sample means Patrick Breheny STA 580: Biostatistics I 19/23

Example #2 The law of averages The theorem How large does n have to be? For example, imagine an urn containing the numbers 1, 2, and 9: n=20 Density 0.0 0.2 0.4 0.6 0.8 2 3 4 5 6 7 Sample mean n=50 Density 0.0 0.2 0.4 0.6 0.8 2 3 4 5 6 Sample mean Patrick Breheny STA 580: Biostatistics I 20/23

Example #2 (cont d) The law of averages The theorem How large does n have to be? n=100 Density 0.0 0.2 0.4 0.6 0.8 1.0 2.5 3.0 3.5 4.0 4.5 5.0 5.5 Sample mean Patrick Breheny STA 580: Biostatistics I 21/23

Example #3 The law of averages The theorem How large does n have to be? Weight tends to be skewed to the right (far more people are overweight than underweight) Let s perform an experiment in which the NHANES sample of adult men is the population I am going to randomly draw twenty-person samples from this population (i.e. I am re-sampling the original sample) Patrick Breheny STA 580: Biostatistics I 22/23

Example #3 (cont d) The law of averages The theorem How large does n have to be? n=20 Density 0.00 0.01 0.02 0.03 0.04 160 180 200 220 240 260 Sample mean Patrick Breheny STA 580: Biostatistics I 23/23