SDS 321: Introduction to Probability and Statistics

Similar documents
Parameter, Statistic and Random Samples

This section is optional.

Stat 319 Theory of Statistics (2) Exercises

Math 152. Rumbos Fall Solutions to Review Problems for Exam #2. Number of Heads Frequency

Mathematical Statistics - MS

5. Likelihood Ratio Tests

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables

LECTURE 8: ASYMPTOTICS I

Chapter 13: Tests of Hypothesis Section 13.1 Introduction

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)].

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

5. Limit Theorems, Part II: Central Limit Theorem. ECE 302 Fall 2009 TR 3 4:15pm Purdue University, School of ECE Prof.

Lecture 19: Convergence

Frequentist Inference

ST5215: Advanced Statistical Theory

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

Sample Size Determination (Two or More Samples)

Estimation for Complete Data

Lecture 01: the Central Limit Theorem. 1 Central Limit Theorem for i.i.d. random variables

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Chapter 6 Sampling Distributions

Lecture 33: Bootstrap

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS

Lecture 2: Concentration Bounds

Lecture 20: Multivariate convergence and the Central Limit Theorem

Lecture 2: Poisson Sta*s*cs Probability Density Func*ons Expecta*on and Variance Es*mators

Topic 9: Sampling Distributions of Estimators

Elements of Statistical Methods Lots of Data or Large Samples (Ch 8)

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

TMA4245 Statistics. Corrected 30 May and 4 June Norwegian University of Science and Technology Department of Mathematical Sciences.

STATISTICAL INFERENCE

Expectation and Variance of a random variable

MATH/STAT 352: Lecture 15

ECE 330:541, Stochastic Signals and Systems Lecture Notes on Limit Theorems from Probability Fall 2002

Unbiased Estimation. February 7-12, 2008

f(x)dx = 1 and f(x) 0 for all x.

A statistical method to determine sample size to estimate characteristic value of soil parameters

Statisticians use the word population to refer the total number of (potential) observations under consideration

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes.

Summary. Recap ... Last Lecture. Summary. Theorem

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

Class 23. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Advanced Stochastic Processes.

Simulation. Two Rule For Inverting A Distribution Function

4. Partial Sums and the Central Limit Theorem

Lecture 2: Monte Carlo Simulation

1 = δ2 (0, ), Y Y n nδ. , T n = Y Y n n. ( U n,k + X ) ( f U n,k + Y ) n 2n f U n,k + θ Y ) 2 E X1 2 X1

The standard deviation of the mean

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

EE 4TM4: Digital Communications II Probability Theory

Last Lecture. Wald Test

Common Large/Small Sample Tests 1/55

Convergence of random variables. (telegram style notes) P.J.C. Spreij

n outcome is (+1,+1, 1,..., 1). Let the r.v. X denote our position (relative to our starting point 0) after n moves. Thus X = X 1 + X 2 + +X n,

Lecture 12: November 13, 2018

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Distribution of Random Samples & Limit theorems

Topic 9: Sampling Distributions of Estimators

1 Convergence in Probability and the Weak Law of Large Numbers

April 18, 2017 CONFIDENCE INTERVALS AND HYPOTHESIS TESTING, UNDERGRADUATE MATH 526 STYLE

Chapter 6 Principles of Data Reduction

Understanding Samples

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables

NANYANG TECHNOLOGICAL UNIVERSITY SYLLABUS FOR ENTRANCE EXAMINATION FOR INTERNATIONAL STUDENTS AO-LEVEL MATHEMATICS

Central limit theorem and almost sure central limit theorem for the product of some partial sums

1 Introduction to reducing variance in Monte Carlo simulations

Random Variables, Sampling and Estimation

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss

IIT JAM Mathematical Statistics (MS) 2006 SECTION A

Statistical inference: example 1. Inferential Statistics

f(x i ; ) L(x; p) = i=1 To estimate the value of that maximizes L or equivalently ln L we will set =0, for i =1, 2,...,m p x i (1 p) 1 x i i=1

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

Introductory statistics

TAMS24: Notations and Formulas

6.041/6.431 Spring 2009 Final Exam Thursday, May 21, 1:30-4:30 PM.

Notes 5 : More on the a.s. convergence of sums

7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals

ECE 6980 An Algorithmic and Information-Theoretic Toolbox for Massive Data

Lecture Notes 15 Hypothesis Testing (Chapter 10)

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

Properties and Hypothesis Testing

Tests of Hypotheses Based on a Single Sample (Devore Chapter Eight)

6.3 Testing Series With Positive Terms

Monte Carlo method and application to random processes

Basics of Probability Theory (for Theory of Computation courses)

Lecture 15: Density estimation

KLMED8004 Medical statistics. Part I, autumn Estimation. We have previously learned: Population and sample. New questions

Asymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

Agenda: Recap. Lecture. Chapter 12. Homework. Chapt 12 #1, 2, 3 SAS Problems 3 & 4 by hand. Marquette University MATH 4740/MSCS 5740

32 estimating the cumulative distribution function

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

5. INEQUALITIES, LIMIT THEOREMS AND GEOMETRIC PROBABILITY

STATISTICAL METHODS FOR BUSINESS

Learning Theory: Lecture Notes

Introduction to Econometrics (3 rd Updated Edition) Solutions to Odd- Numbered End- of- Chapter Exercises: Chapter 3

Transcription:

SDS 321: Itroductio to Probability ad Statistics Lecture 23: Cotiuous radom variables- Iequalities, CLT Puramrita Sarkar Departmet of Statistics ad Data Sciece The Uiversity of Texas at Austi www.cs.cmu.edu/ psarkar/teachig 1

Roadmap Examples of Markov ad Chebyshev Weak law of large umbers ad CLT Normal approximatio to Biomial 2

Markov s iequality Example You have 20 idepedet Poisso(1) radom variables X 1,..., X 20. Use 20 the Markov iequality to boud P( X i 15) i=1 3

Markov s iequality Example You have 20 idepedet Poisso(1) radom variables X 1,..., X 20. Use 20 the Markov iequality to boud P( X i 15) i=1 P( X i 15) E[ i X i ] = 20 15 15 = 4 3 i 3

Markov s iequality Example You have 20 idepedet Poisso(1) radom variables X 1,..., X 20. Use 20 the Markov iequality to boud P( X i 15) i=1 P( X i 15) E[ i X i ] = 20 15 15 = 4 3 i How useful is this? 3

Chebyshev s iequality Example You have idepedet Poisso(1) radom variables X 1,..., X. Use the Chebyshev iequality to boud P( X 1 1)? 4

Chebyshev s iequality Example You have idepedet Poisso(1) radom variables X 1,..., X. Use the Chebyshev iequality to boud P( X 1 1)? P( X 1 1) var(x 1) = 1 = 1 10 = 1 100... Whe = 10 Whe = 100 4

Weak law of large umbers The WLLN basically states that the sample mea of a large umber of radom variables is very close to the true mea with high probability. Cosider a sequece of i.i.d radom variables X 1,... X with mea µ ad variace σ 2. Let M = X 1 + + X. 5

Weak law of large umbers The WLLN basically states that the sample mea of a large umber of radom variables is very close to the true mea with high probability. Cosider a sequece of i.i.d radom variables X 1,... X with mea µ ad variace σ 2. Let M = X 1 + + X. E[M] = E[X 1] + + E[X] = µ 5

Weak law of large umbers The WLLN basically states that the sample mea of a large umber of radom variables is very close to the true mea with high probability. Cosider a sequece of i.i.d radom variables X 1,... X with mea µ ad variace σ 2. Let M = X 1 + + X. E[M] = E[X 1] + + E[X] = µ var[x var(m ) = 1 ] + + var[x] 2 = σ2 5

Weak law of large umbers The WLLN basically states that the sample mea of a large umber of radom variables is very close to the true mea with high probability. Cosider a sequece of i.i.d radom variables X 1,... X with mea µ ad variace σ 2. Let M = X 1 + + X. E[M] = E[X 1] + + E[X] = µ var[x var(m ) = 1 ] + + var[x] 2 = σ2 So P( M µ ɛ) σ2 ɛ 2 5

Weak law of large umbers The WLLN basically states that the sample mea of a large umber of radom variables is very close to the true mea with high probability. Cosider a sequece of i.i.d radom variables X 1,... X with mea µ ad variace σ 2. Let M = X 1 + + X. E[M] = E[X 1] + + E[X] = µ var[x var(m ) = 1 ] + + var[x] 2 = σ2 So P( M µ ɛ) σ2 ɛ 2 For large this probability is small. 5

Illustratio Cosider the mea of idepedet Poisso(1) radom variables. For each, we plot the distributio of the average. 6

Ca we say more? Cetral Limit Theorem Turs out that ot oly ca you say that the sample mea is close to the true mea, you ca actually predict its distributio usig the famous Cetral Limit Theorem. Cosider a sequece of i.i.d radom variables X 1,... X with mea µ ad variace σ 2. Let X = X 1 + + X. Remember E[ X] = µ ad var( X) = σ 2 / Stadardize X to get X µ σ/ X µ As gets bigger, σ/ behaves more ad more like a Normal(0, 1) radom variable. ( ) X µ P σ/ < z Φ(z) 7

Ca we say more? Cetral Limit Theorem Figure: (Courtesy: Tamara Broderick) You bet! 8

Example You have 20 idepedet Poisso(1) radom variables X 1,..., X 20. Use 20 the CLT to boud P( X i 15) i=1 9

Example You have 20 idepedet Poisso(1) radom variables X 1,..., X 20. Use 20 the CLT to boud P( X i 15) i=1 P( X i 15) = P( X i 20 5) i i = P( X 1 1/ 20.25 20) P(Z 1.18) = 0.86 9

Example You have 20 idepedet Poisso(1) radom variables X 1,..., X 20. Use 20 the CLT to boud P( X i 15) i=1 P( X i 15) = P( X i 20 5) i i = P( X 1 1/ 20.25 20) P(Z 1.18) = 0.86 How useful is this? Better tha Markov. 9

Example A astroomer is iterested i measurig the distace, i light-years, from his observatory to a distat star. Although the astroomer has a measurig techique, he kows that, because of chagig atmospheric coditios ad ormal error, each time a measuremet is made it will ot yield the exact distace, but merely a estimate. As a result, the astroomer plas to make a series of measuremets ad the use the average value of these measuremets as his estimated value of the actual distace. If the astroomer believes that the values of the measuremets are idepedet ad idetically distributed radom variables havig a commo mea d (the actual distace) ad a commo variace of 4 (light-years), how may measuremets eed he make to 95% sure that his estimated distace is accurate to withi ±.5 lightyears? 10

Example A astroomer is iterested i measurig the distace, i light-years, from his observatory to a distat star. Although the astroomer has a measurig techique, he kows that, because of chagig atmospheric coditios ad ormal error, each time a measuremet is made it will ot yield the exact distace, but merely a estimate. As a result, the astroomer plas to make a series of measuremets ad the use the average value of these measuremets as his estimated value of the actual distace. If the astroomer believes that the values of the measuremets are idepedet ad idetically distributed radom variables havig a commo mea d (the actual distace) ad a commo variace of 4 (light-years), how may measuremets eed he make to 95% sure that his estimated distace is accurate to withi ±.5 lightyears? Let X be the mea of the measuremets. How large does have to be so that P( X d.5) = 0.95 ( X d P 2/ 0.25 ) P( Z 0.25 ) = 1 2P(Z 0.25 ) = 0.95 P(Z 0.25 ) = 0.025 0.25 = 1.96 = 7.84 62 10

Normal Approximatio to Biomial The probability of sellig a umbrella is 0.5 o a raiy day. If there are 400 umbrellas i the store, whats the probability that the ower will sell at least 180? Let X be the total umber of umbrellas sold. X Biomial(400,.5) We wat P(X > 180). Crazy calculatios. 11

Normal Approximatio to Biomial The probability of sellig a umbrella is 0.5 o a raiy day. If there are 400 umbrellas i the store, whats the probability that the ower will sell at least 180? Let X be the total umber of umbrellas sold. X Biomial(400,.5) We wat P(X > 180). Crazy calculatios. But ca we approximate the distributio of X /? X / = ( i Y i )/ where E[Y i ] = 0.5 ad var(y i ) = 0.25. Sure! CLT tells us that for large, X /400 0.5 0.25/400 N(0, 1) So P(X > 180) = P((X 200)/ 100 > 2) P(Z 2) = 1 Φ( 2) = 0.97 11

Frequetist Statistics The parameter(s) θ is fixed ad ukow Data is geerated through the likelihood fuctio p(x ; θ) (if discrete) or f (X ; θ) (if cotiuous). Now we will be dealig with multiple cadidate models, oe for each value of θ We will use E θ [h(x )] to defie the expectatio of the radom variable h(x ) as a fuctio of parameter θ 12

Problems we will look at Parameter estimatio: We wat to estimate ukow parameters from data. Maximum Likelihood estimatio (sectio 9.1): Select the parameter that makes the observed data most likely. i.e. maximize the probability of obtaiig the data at had. Hypothesis testig: A ukow parameter takes a fiite umber of values. Oe wats to fid the best hypothesis based o the data. Sigificace testig: Give a hypothesis, figure out the rejectio regio ad reject the hypothesis if the observatio falls withi this regio. 13