Continuous Expectation and Variance, the Law of Large Numbers, and the Central Limit Theorem Spring 2014

Similar documents
Central Limit Theorem and the Law of Large Numbers Class 6, Jeremy Orloff and Jonathan Bloom

PROBABILITY: AN INTERACTIVE REVIEW, II CS1951A INTRO TO DATA SCIENCE DAN POTTER

Confidence Intervals for Normal Data Spring 2014

Expectation, Variance and Standard Deviation for Continuous Random Variables Class 6, Jeremy Orloff and Jonathan Bloom

Class 8 Review Problems 18.05, Spring 2014

Class 8 Review Problems solutions, 18.05, Spring 2014

Null Hypothesis Significance Testing p-values, significance level, power, t-tests

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

Class 26: review for final exam 18.05, Spring 2014

Exam 2 Practice Questions, 18.05, Spring 2014

Exam 1 Practice Questions I solutions, 18.05, Spring 2014

Post-exam 2 practice questions 18.05, Spring 2014

Recitation 2: Probability

This does not cover everything on the final. Look at the posted practice problems for other topics.

Joint Probability Distributions and Random Samples (Devore Chapter Five)

SDS 321: Introduction to Probability and Statistics

Lecture 10. Variance and standard deviation

Confidence Intervals for the Mean of Non-normal Data Class 23, Jeremy Orloff and Jonathan Bloom

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017

Homework for 1/13 Due 1/22

Probability Distributions

Lecture 9. Expectations of discrete random variables

Compute f(x θ)f(θ) dθ

6.041/6.431 Fall 2010 Quiz 2 Solutions

STAT Chapter 5 Continuous Distributions

Section 9.1. Expected Values of Sums

Chapter 6: Functions of Random Variables

CS145: Probability & Computing

Review for Exam Spring 2018

18.05 Practice Final Exam

Arkansas Tech University MATH 3513: Applied Statistics I Dr. Marcel B. Finan

Lecture 8 Sampling Theory

STATISTICS 1 REVISION NOTES

Lecture 16. Lectures 1-15 Review

Math 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) =

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

8 Laws of large numbers

Brief Review of Probability

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

MAS113 Introduction to Probability and Statistics

Math 416 Lecture 3. The average or mean or expected value of x 1, x 2, x 3,..., x n is

Practice Exam 1: Long List 18.05, Spring 2014

Bell-shaped curves, variance

Tom Salisbury

Special distributions

STA 111: Probability & Statistical Inference

CS 5014: Research Methods in Computer Science. Bernoulli Distribution. Binomial Distribution. Poisson Distribution. Clifford A. Shaffer.

STAT 430/510: Lecture 10

MATH Notebook 5 Fall 2018/2019

Continuous Data with Continuous Priors Class 14, Jeremy Orloff and Jonathan Bloom

Massachusetts Institute of Technology

Examples of random experiment (a) Random experiment BASIC EXPERIMENT

Choosing priors Class 15, Jeremy Orloff and Jonathan Bloom

MAS108 Probability I

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

The Central Limit Theorem

ECON Fundamentals of Probability

Continuous Random Variables

Analysis of Engineering and Scientific Data. Semester

Example A. Define X = number of heads in ten tosses of a coin. What are the values that X may assume?

1 Presessional Probability

Covariance and Correlation Class 7, Jeremy Orloff and Jonathan Bloom

A Brief Analysis of Central Limit Theorem. SIAM Chapter Florida State University

CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions)

Outline PMF, CDF and PDF Mean, Variance and Percentiles Some Common Distributions. Week 5 Random Variables and Their Distributions

Math 381 Discrete Mathematical Modeling

COMPSCI 240: Reasoning Under Uncertainty

*Karle Laska s Sections: There is no class tomorrow and Friday! Have a good weekend! Scores will be posted in Compass early Friday morning

F X (x) = P [X x] = x f X (t)dt. 42 Lebesgue-a.e, to be exact 43 More specifically, if g = f Lebesgue-a.e., then g is also a pdf for X.

Disjointness and Additivity

18.05 Exam 1. Table of normal probabilities: The last page of the exam contains a table of standard normal cdf values.

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Limiting Distributions

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

Continuous Distributions

(It's not always good, but we can always make it.) (4) Convert the normal distribution N to the standard normal distribution Z. Specically.

18.175: Lecture 8 Weak laws and moment-generating/characteristic functions

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Chapter 7: Theoretical Probability Distributions Variable - Measured/Categorized characteristic

18.440: Lecture 28 Lectures Review

Bayesian Updating: Discrete Priors: Spring 2014

Jointly Distributed Random Variables

Comparison of Bayesian and Frequentist Inference

Confidence Intervals for Normal Data Spring 2018

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Lecture 10: Normal RV. Lisa Yan July 18, 2018

Chapter 4. Chapter 4 sections

n(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3

Introduction to Probability

Confidence Intervals for the Sample Mean

Statistics for Applications. Chapter 1: Introduction 1/43

STAT 414: Introduction to Probability Theory

Massachusetts Institute of Technology

STAT 418: Probability and Stochastic Processes

More on Distribution Function

Functions of Several Random Variables (Ch. 5.5)

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

SDS 321: Practice questions

Chapter 5. Means and Variances

14.30 Introduction to Statistical Methods in Economics Spring 2009

Transcription:

Continuous Expectation and Variance, the Law of Large Numbers, and the Central Limit Theorem 18.5 Spring 214.5.4.3.2.1-4 -3-2 -1 1 2 3 4 January 1, 217 1 / 31

Expected value Expected value: measure of location, central tendency X continuous with range [a, b] and pdf f (x): E (X ) = b a xf (x) dx. X discrete with values x 1,..., x n and pmf p(x i ): n E (X ) = x i p(x i ). i=1 View these as essentially the same formulas. January 1, 217 2 / 31

Variance and standard deviation Standard deviation: measure of spread, scale For any random variable X with mean µ Var(X ) = E ((X µ) 2 ), σ = Var(X ) X continuous with range [a, b] and pdf f (x): b Var(X ) = (x µ) 2 f (x) dx. X discrete with values x 1,..., x n and pmf p(x i ): a n Var(X ) = (x i µ) 2 p(x i ). i=1 View these as essentially the same formulas. January 1, 217 3 / 31

Properties Properties: (the same for discrete and continuous) 1. E (X + Y ) = E (X ) + E (Y ). 2. E (ax + b) = ae (X ) + b. 3. If X and Y are independent then Var(X + Y ) = Var(X ) + Var(Y ). 4. Var(aX + b) = a 2 Var(X ). 5. Var(X ) = E (X 2 ) E (X ) 2. January 1, 217 4 / 31

Board question 2 The random variable X has range [,1] and pdf cx. (a) Find c. (b) Find the mean, variance and standard deviation of X. (c) Find the median value of X. (d) Suppose X 1,... X 16 are independent identically-distributed copies of X. Let X be their average. What is the standard deviation of X? (e) Suppose Y = X 4. Find the pdf of Y. answer: See next slides. January 1, 217 5 / 31

Solution 1 (a) Total probability is 1: cx 2 dx = 1 c = 3. 1 (b) µ = 3x 3 dx = 3/4. 1 9 3 σ 2 = ( (x 3/4) 2 3x 2 dx) = 3 9 + = σ = 3/8 = 1 4 3/5.194 (c) Set F (q.5 ) =.5, solve for q.5 : F (x) = F (q.5 ) = q.5 3 =.5. We get, q.5 = (.5) 1/3. 5 8 16 8. x 3 3u 2 du = x. Therefore, (d) Because they are independent Var(X 1 +... + X 16 ) = Var(X 1 ) + Var(X 2 ) +... + Var(X 16 ) = 16Var(X ). 16Var(X ) Var(X ) σ X Thus, Var(X ) = =. Finally, σ = =.194/4. 16 2 16 X 4 January 1, 217 6 / 31

Solution continued (e) Method 1 use the cdf: 1 1 3 F Y (y) = P(X 4 < y) = P(X < y 4 ) = F X (y 4 ) = y 4. 3 1 4 Y 4 Now differentiate. f Y (y) = F (y) = y. Method 2 use the pdf: We have 4 y = x dy = 4x 3 dx dy 4y 3/4 = dx dy 3y 2/4 dy 3 This implies f X (x) dx = f X (y 1/4 ) = = dy 4y 3/4 4y 3/4 4y 1/4 3 Therefore f Y (y) = 4y 1/4 January 1, 217 7 / 31

Quantiles Quantiles give a measure of location. left tail area = prob. =.6 φ(z) q.6 =.253 z F (q.6 ) =.6 1 Φ(z) q.6 =.253 z q.6 : left tail area =.6 F (q.6 ) =.6 January 1, 217 8 / 31

Concept question Each of the curves is the density for a given random variable. The median of the black plot is always at q. Which density has the greatest median? 1. Black 2. Red 3. Blue 4. All the same 5. Impossible to tell (A) Curves coincide to here. (B) q answer: See next frame. q January 1, 217 9 / 31

Solution (A) Curves coincide to here. Area to the left of the median =.5 q Plot A: 4. All three medians are the same. Remember that probability is computed as the area under the curve. By definition the median q is the point where the shaded area in Plot A.5. Since all three curves coincide up to q. That is, the shaded area in the figure is represents a probability of.5 for all three densities. Continued on next slide. January 1, 217 1 / 31

Solution continued (B) Plot B: 2. The red density has the greatest median. Since q is the median for the black density, the shaded area in Plot B is.5. Therefore the area under the blue curve (up to q) is greater than.5 and that under the red curve is less than.5. This means the median of the blue density is to the left of q (you need less area) and the median of the red density is to the right of q (you need more area). q January 1, 217 11 / 31

Law of Large Numbers (LoLN) Informally: An average of many measurements is more accurate than a single measurement. Formally: Let X 1, X 2,... be i.i.d. random variables all with mean µ and standard deviation σ. Let X 1 + X 2 +... + X n 1 n X n = = X i. n n i=1 Then for any (small number) a, we have lim P( X n µ < a) = 1. n No guarantees but: By choosing n large enough we can make X n as close as we want to µ with probability close to 1. January 1, 217 12 / 31

Concept Question: Desperation You have $1. You need $1 by tomorrow morning. Your only way to get it is to gamble. If you bet $k, you either win $k with probability p or lose $k with probability 1 p. Maximal strategy: Bet as much as you can, up to what you need, each time. Minimal strategy: Make a small bet, say $5, each time. 1. If p =.45, which is the better strategy? (a) Maximal (b) Minimal (c) They are the same 2. If p =.8, which is the better strategy? (a) Maximal (b) Minimal (c) They are the same answer: On next slide January 1, 217 13 / 31

Solution to previous two problems answer: If p =.45 use maximal strategy; If p =.8 use minimal strategy. If you use the minimal strategy the law of large numbers says your average winnings per bet will almost certainly be the expected winnings of one bet. The two tables represent p =.45 and p =.8 respectively. Win -1 1 Win -1 1 p.55.45 p.2.8 The expected value of a $5 bet when p =.45 is -$.5 Since on average you will lose $.5 per bet you want to avoid making a lot of bets. You go for broke and hope to win big a few times in a row. It s not very likely, but the maximal strategy is your best bet. The expected value when p =.8 is $3. Since this is positive you d like to make a lot of bets and let the law of large numbers (practically) guarantee you will win an average of $6 per bet. So you use the minimal strategy. January 1, 217 14 / 31

Histograms Made by binning data. Frequency: height of bar over bin = number of data points in bin. Density: area of bar is the fraction of all data points that lie in the bin. So, total area is 1. frequency density 4 3 2 1.8.6.4.2.25.75 1.25 1.75 2.25 x.25.75 1.25 1.75 2.25 x Check that the total area of the histogram on the right is 1. January 1, 217 15 / 31

Board question 1. Make both a frequency and density histogram from the data below. Use bins of width.5 starting at. The bins should be right closed. 1 1.2 1.3 1.6 1.6 2.1 2.2 2.6 2.7 3.1 3.2 3.4 3.8 3.9 3.9 2. Same question using unequal width bins with edges, 1, 3, 4. 3. For question 2, why does the density histogram give a more reasonable representation of the data. January 1, 217 16 / 31

Solution Frequency Frequency. 1. 2. 3. 2 4 6 8 1 2 3 4 Density..2.4 1 2 3 4 Histograms with equal width bins 1 2 3 4 Density..2.4 1 2 3 4 Histograms with unequal width bins January 1, 217 17 / 31

LoLN and histograms LoLN implies density histogram converges to pdf:.5.4.3.2.1-4 -3-2 -1 1 2 3 4 Histogram with bin width.1 showing 1 draws from a standard normal distribution. Standard normal pdf is overlaid in red. January 1, 217 18 / 31

Standardization Random variable X with mean µ and standard deviation σ. X µ Standardization: Y =. σ Y has mean and standard deviation 1. Standardizing any normal random variable produces the standard normal. If X normal then standardized X stand. normal. We use reserve Z to mean a standard normal random variable. January 1, 217 19 / 31

Concept Question: Standard Normal within 1 σ 68% Normal PDF within 2 σ 95% 68% within 3 σ 99% 95% 3σ 2σ σ 99% 1. P( 1 < Z < 1) is (a).25 (b).16 (c).68 (d).84 (e).95 σ 2σ 3σ z 2. P(Z > 2) (a).25 answer: 1c, 2a (b).16 (c).68 (d).84 (e).95 January 1, 217 2 / 31

Central Limit Theorem Setting: X 1, X 2,... i.i.d. with mean µ and standard dev. σ. For each n: 1 X n = (X 1 + X 2 +... + X n ) n average S n = X 1 + X 2 +... + X n sum. Conclusion: For large n: ( ) σ 2 X n N µ, n S n N nµ, nσ 2 Standardized S n or X n N(, 1) S n nµ X n µ That is, = N(, 1). nσ σ/ n January 1, 217 21 / 31

CLT: pictures Standardized average of n i.i.d. uniform random variables with n = 1, 2, 4, 12..4.35.3.25.2.15.1.5-3 -2-1 1 2 3.5.4.3.2.1-3 -2-1 1 2 3.4.35.3.25.2.15.1.5-3 -2-1 1 2 3.4.35.3.25.2.15.1.5-3 -2-1 1 2 3 January 1, 217 22 / 31

CLT: pictures 2 The standardized average of n i.i.d. exponential random variables with n = 1, 2, 8, 64. 1.8.6.4.2-3 -2-1 1 2 3.7.6.5.4.3.2.1-3 -2-1 1 2 3.5.4.3.2.1-3 -2-1 1 2 3.5.4.3.2.1-3 -2-1 1 2 3 January 1, 217 23 / 31

CLT: pictures 3 The standardized average of n i.i.d. Bernoulli(.5) random variables with n = 1, 2, 12, 64..4.35.3.25.2.15.1.5-3 -2-1 1 2 3.4.35.3.25.2.15.1.5-3 -2-1 1 2 3.4.35.3.25.2.15.1.5-3 -2-1 1 2 3.4.35.3.25.2.15.1.5-4 -3-2 -1 1 2 3 4 January 1, 217 24 / 31

CLT: pictures 4 The (non-standardized) average of n Bernoulli(.5) random variables, with n = 4, 12, 64. (Spikier.) 1.4 1.2 1.8.6.4.2-1 -.5.5 1 1.5 2 7 6 5 4 3 2 1 3 2.5 2 1.5 1.5 -.2.2.4.6.8 1 1.2 1.4 -.2.2.4.6.8 1 1.2 1.4 January 1, 217 25 / 31

Table Question: Sampling from the standard normal distribution As a table, produce a single random sample from (an approximate) standard normal distribution. The table is allowed nine rolls of the 1-sided die. Note: µ = 5.5 and σ 2 = 8.25 for a single 1-sided die. Hint: CLT is about averages. answer: The average of 9 rolls is a sample from the average of 9 independent random variables. The CLT says this average is approximately normal with µ = 5.5 and σ = 8.25/ 9 = 2.75 If x is the average of 9 rolls then standardizing we get x 5.5 z = 2.75 is (approximately) a sample from N(, 1). January 1, 217 26 / 31

Board Question: CLT 1. Carefully write the statement of the central limit theorem. 2. To head the newly formed US Dept. of Statistics, suppose that 5% of the population supports Ani, 25% supports Ruthi, and the remaining 25% is split evenly between Efrat, Elan, David and Jerry. A poll asks 4 random people who they support. What is the probability that at least 55% of those polled prefer Ani? 3. What is the probability that less than 2% of those polled prefer Ruthi? answer: On next slide. January 1, 217 27 / 31

Solution answer: 2. Let A be the fraction polled who support Ani. So A is the average of 4 Bernoulli(.5) random variables. That is, let X i = 1 if the ith person polled prefers Ani and if not, so A = average of the X i. The question asks for the probability A >.55. Each X i has µ =.5 and σ 2 =.25. So, E (A) =.5 and σ 2 =.25/4 or σ A = 1/4 =.25. A Because A is the average of 4 Bernoulli(.5) variables the CLT says it is approximately normal and standardizing gives So Continued on next slide A.5 Z.25 P(A >.55) P(Z > 2).25 January 1, 217 28 / 31

Solution continued 3. Let R be the fraction polled who support Ruthi. The question asks for the probability the R <.2. Similar to problem 2, R is the average of 4 Bernoulli(.25) random variables. So E (R) =.25 and σr 2 = (.25)(.75)/4 = σ R = 3/8. R.25 So Z. So, 3/8 P(R <.2) P(Z < 4/ 3).15 January 1, 217 29 / 31

Bonus problem Not for class. Solution will be posted with the slides. An accountant rounds to the nearest dollar. We ll assume the error in rounding is uniform on [-.5,.5]. Estimate the probability that the total error in 3 entries is more than $5. answer: Let X j be the error in the j th entry, so, X j U(.5,.5). We have E(X j ) = and Var(X j ) = 1/12. The total error S = X 1 +... + X 3 has E (S) =, Var(S) = 3/12 = 25, and σ S = 5. Standardizing we get, by the CLT, S/5 is approximately standard normal. That is, S/5 Z. So P(S < 5 or S > 5) P(Z < 1 or Z > 1).32. January 1, 217 3 / 31

MIT OpenCourseWare https://ocw.mit.edu 18.5 Introduction to Probability and Statistics Spring 214 For information about citing these materials or our Terms of Use, visit: https://ocw.mit.edu/terms.