COMPSCI 240: Reasoning Under Uncertainty

Similar documents
Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages

Introduction to Probability

CS145: Probability & Computing

Practice Problem - Skewness of Bernoulli Random Variable. Lecture 7: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example

Fundamental Tools - Probability Theory IV

Chapter 6: Large Random Samples Sections

MAS113 Introduction to Probability and Statistics

Midterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example

Proving the central limit theorem

Limiting Distributions

Lecture Notes 3 Convergence (Chapter 5)

Twelfth Problem Assignment

CSE 312 Final Review: Section AA

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Limiting Distributions

6.1 Moment Generating and Characteristic Functions

Lecture 4: September Reminder: convergence of sequences

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

2 Random Variable Generation

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

STAT 430/510 Probability Lecture 12: Central Limit Theorem and Exponential Distribution

Concentration inequalities and tail bounds

COMP2610/COMP Information Theory

Math438 Actuarial Probability

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2017 Kannan Ramchandran March 21, 2017.

Stochastic Models (Lecture #4)

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability

Disjointness and Additivity

COMPSCI 240: Reasoning Under Uncertainty

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

1 Exercises for lecture 1

Expectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda

8 Laws of large numbers

Statistics and Sampling distributions

Lecture 8. October 22, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example

ECE302 Spring 2015 HW10 Solutions May 3,

Lecture 1: August 28

Midterm Exam 2 (Solutions)

Continuous Expectation and Variance, the Law of Large Numbers, and the Central Limit Theorem Spring 2014

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Notes 6 : First and second moment methods

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

Chapter 5. Chapter 5 sections

THEORY OF PROBABILITY VLADIMIR KOBZAR

Two hours. Statistical Tables to be provided THE UNIVERSITY OF MANCHESTER. 14 January :45 11:45

Stat410 Probability and Statistics II (F16)

Lecture 8 Sampling Theory

Applied Statistics I

Probabilistic Inversion of Laplace Transforms

Lecture 18: Central Limit Theorem. Lisa Yan August 6, 2018

{X i } realize. n i=1 X i. Note that again X is a random variable. If we are to

Chapter 5 continued. Chapter 5 sections

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

Notes on the second moment method, Erdős multiplication tables

Course information: Instructor: Tim Hanson, Leconte 219C, phone Office hours: Tuesday/Thursday 11-12, Wednesday 10-12, and by appointment.

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Expectation of Random Variables

18.175: Lecture 8 Weak laws and moment-generating/characteristic functions

Probability, Random Processes and Inference

Stat 704 Data Analysis I Probability Review

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Lecture 2: Review of Basic Probability Theory

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Lecture 11. Probability Theory: an Overveiw

STAT Chapter 5 Continuous Distributions

Continuous distributions

Limit Theorems. STATISTICS Lecture no Department of Econometrics FEM UO Brno office 69a, tel

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Probability inequalities 11

Random variables, Expectation, Mean and Variance. Slides are adapted from STAT414 course at PennState

Most data analysis starts with some data set; we will call this data set P. It will be composed of a set of n

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

Northwestern University Department of Electrical Engineering and Computer Science

Lecture 4: Sampling, Tail Inequalities

Discrete Probability Refresher

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1

Chapter 4. Chapter 4 sections

Lecture 4: Random Variables and Distributions

II. The Normal Distribution

Chapter 7: SAMPLING DISTRIBUTIONS & POINT ESTIMATION OF PARAMETERS. Part 2: Differences in sample means

Test Problems for Probability Theory ,

SDS 321: Introduction to Probability and Statistics

STA205 Probability: Week 8 R. Wolpert

Mathematical statistics

Lesson 100: The Normal Distribution. HL Math - Santowski

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.

Probability and Statistics Notes

Topic 7: Convergence of Random Variables

MATH Notebook 5 Fall 2018/2019

Lecture 3: Statistical sampling uncertainty

14.30 Introduction to Statistical Methods in Economics Spring 2009

Statistical Hypothesis Testing

CDA5530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables

1.1 Review of Probability Theory

University of Regina. Lecture Notes. Michael Kozdron

3. Review of Probability and Statistics

Transcription:

COMPSCI 240: Reasoning Under Uncertainty Andrew Lan and Nic Herndon University of Massachusetts at Amherst Spring 2019

Lecture 20: Central limit theorem & The strong law of large numbers

Markov and Chebyshev Bounds Markov Bound Informally: If a nonnegative RV has a small mean, then the probability that this RV takes a large value must also be small. Formally: For a non-negative random variable X, P(X a) E(X) a Chebyshev Bound Informally: If a RV has small variance, then the probability that it takes a value far from its mean is also small. Note that the Chebyshev inequality does not require the random variable to be nonnegative. Formally: For a random variable X, P( X E(X) c) Var(X) c 2 3 / 15

The Weak Law of Large Numbers Informally: If n is large, the bulk of the distribution of the sample mean (X n ) of a sequence of i.i.d. with mean µ and variance σ 2 will converge to (be concentrated around) µ. Formally: Let X 1, X 2, be a sequence of i.i.d. (either discrete or continuous) random variable with mean µ. For every ɛ > 0, we have P ( ) X n µ ɛ 0 as n. 4 / 15

Convergence in probability Let Y 1, Y 2,... be a sequence of random variables (not necessarily independent), and let a be a real number. We say that the sequence Y n converges to a in probability, if for every ɛ > 0, we have lim P( Y n a ɛ) = 0 n Put it another way: ɛ, δ > 0, n 0 such that n n 0 P( Y n a ɛ) δ Our measurement is accurate, with this much confidence. 5 / 15

The Strong Law of Large Numbers Let X 1, X 2, be a sequence of i.i.d. (either discrete or continuous) random variable with mean µ and variance σ 2. Then, the sequence of sample mean X n converges to µ as n, with probability 1: ( ) P lim X n = µ = 1. n Its sample mean X n, which is a RV, will converge to the true mean µ, which is a constant, with a probability 1 when we have an infinitely large sample size. More specifically, an event of X n = µ has a probability of 1. Example: Let X i Bern(p), then ( ) 1 n P lim X i = p = 1. n n i=1 6 / 15

The Central Limit Theorem The LLN states that X n converges to µ when n is large. The distribution of the sample mean X n is concentrated around µ. But what does the distribution of X n look like? The Central Limit Theorem can define this. 7 / 15

The Central Limit Theorem Let us define a variable by normalizing X n with its mean and standard deviation In the same manner as we normalized a Normal RV to derive the Standard Normal RV. Z n = X n µ σ n or equivalently Z n = X 1 + + X n nµ σ n Then, the PDF of Z n converges to the standard normal PDF as n Z n N(0, 1) as n 8 / 15

The Central Limit Theorem The CLT is surprisingly general and extremely powerful. It states that X i can have any forms of (discrete, continuous, or a mixture) probability distribution, but its sample mean converges to a Standard Normal distribution as n becomes large. Conceptually, this is important as it indicates that the sum of a large number of i.i.d RV is approximately normal. Practically, this is important as it eliminates the need for detailed probabilistic models as long as we have a large sample size. We can still approximate its sample mean using the Standard Normal distribution as long as we know µ and σ. 9 / 15

The Central Limit Theorem Let us run a simulation to see if this work! Consider a continuous exponential RV whose λ = 0.01 Sampling distribution of X n when n = 2 1200 50000 samples of X n, where n = 2 and X is a Exp RV 1000 800 Frequency 600 400 200 0-200 -100 0 100 200 300 400 500 600 700 Sample Mean of X n 10 / 15

The Central Limit Theorem Let us run a simulation to see if this work! Consider a continuous exponential RV whose λ = 0.01 Sampling distribution of X n when n = 4 1000 50000 samples of X n, where n = 4 and X is a Exp RV Frequency 900 800 700 600 500 400 300 200 100 0-100 0 100 200 300 400 500 Sample Mean of X n 11 / 15

The Central Limit Theorem Let us run a simulation to see if this work! Consider a continuous exponential RV whose λ = 0.01 Sampling distribution of X n when n = 20 800 50000 samples of X n, where n = 20 and X is a Exp RV 700 600 Frequency 500 400 300 200 100 0 0 50 100 150 200 250 X n 12 / 15

The Central Limit Theorem Let us run a simulation to see if this work! Consider a continuous exponential RV whose λ = 0.01 Sampling distribution of X n when n = 100 900 50000 samples of X n, where n = 100 and X is a Exp RV 800 700 600 Frequency 500 400 300 200 100 0 60 70 80 90 100 110 120 130 140 150 160 Sample Mean of X n 13 / 15

Example Question: Suppose salaries at a very large company have a mean of $62,000 and a standard deviation of $32,000. If a single employee is randomly selected, what is the probability that his/her salary exceeds $66,000? Solution: We cannot solve this problem since we do not have the true distribution function of the salaries. 14 / 15

Example Question: Suppose salaries at a very large company have a mean of $62,000 and a standard deviation of $32,000. If 100 employees are randomly selected, what is the probability that their average salary exceeds $66,000? Solution: We define a new random variable Z = X n µ σ n Then, P(X n > 66000) = P ( Z > ) 66000 62000 32000 100 = P(Z > 1.25) = 1 Φ(1.25) 15 / 15