Large Sample Theory. Convergence. Central Limit Theorems Asymptotic Distribution Delta Method. Convergence in Probability Convergence in Distribution

Similar documents
Lecture 20: Multivariate convergence and the Central Limit Theorem

Solution to Chapter 2 Analytical Exercises

An Introduction to Asymptotic Theory

Lecture 8: Convergence of transformations and law of large numbers

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

Lecture 19: Convergence

LECTURE 8: ASYMPTOTICS I

Distribution of Random Samples & Limit theorems

STA Object Data Analysis - A List of Projects. January 18, 2018

ST5215: Advanced Statistical Theory

17. Joint distributions of extreme order statistics Lehmann 5.1; Ferguson 15

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)].

Lecture 33: Bootstrap

Asymptotic Results for the Linear Regression Model

Parameter, Statistic and Random Samples

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence

MA Advanced Econometrics: Properties of Least Squares Estimators

7.1 Convergence of sequences of random variables

1 = δ2 (0, ), Y Y n nδ. , T n = Y Y n n. ( U n,k + X ) ( f U n,k + Y ) n 2n f U n,k + θ Y ) 2 E X1 2 X1

2.2. Central limit theorem.

HOMEWORK I: PREREQUISITES FROM MATH 727

ECE 330:541, Stochastic Signals and Systems Lecture Notes on Limit Theorems from Probability Fall 2002

Asymptotic distribution of products of sums of independent random variables

Convergence of random variables. (telegram style notes) P.J.C. Spreij

MIT Spring 2016

1.3 Convergence Theorems of Fourier Series. k k k k. N N k 1. With this in mind, we state (without proof) the convergence of Fourier series.

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

32 estimating the cumulative distribution function

2.1. Convergence in distribution and characteristic functions.

Statistical Inference Based on Extremum Estimators

STATISTICAL METHODS FOR BUSINESS

Lecture Chapter 6: Convergence of Random Sequences

Lecture 01: the Central Limit Theorem. 1 Central Limit Theorem for i.i.d. random variables

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett

Probability and Random Processes

62. Power series Definition 16. (Power series) Given a sequence {c n }, the series. c n x n = c 0 + c 1 x + c 2 x 2 + c 3 x 3 +

7.1 Convergence of sequences of random variables

Exercise 4.3 Use the Continuity Theorem to prove the Cramér-Wold Theorem, Theorem. (1) φ a X(1).

4. Partial Sums and the Central Limit Theorem

Introductory statistics

Entropy Rates and Asymptotic Equipartition

The Central Limit Theorem

Summary. Recap. Last Lecture. Let W n = W n (X 1,, X n ) = W n (X) be a sequence of estimators for

Asymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values

This section is optional.

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

Elements of Statistical Methods Lots of Data or Large Samples (Ch 8)

STAT331. Example of Martingale CLT with Cox s Model

Summary. Recap ... Last Lecture. Summary. Theorem

Mathematical Statistics - MS

Exponential Families and Bayesian Inference

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

Notes 19 : Martingale CLT

EFFECTIVE WLLN, SLLN, AND CLT IN STATISTICAL MODELS

Last Lecture. Biostatistics Statistical Inference Lecture 16 Evaluation of Bayes Estimator. Recap - Example. Recap - Bayes Estimator

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 6 9/23/2013. Brownian motion. Introduction

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

1 Convergence in Probability and the Weak Law of Large Numbers

Random Variables, Sampling and Estimation

Unbiased Estimation. February 7-12, 2008

Lecture 15: Density estimation

Economics 102C: Advanced Topics in Econometrics 4 - Asymptotics & Large Sample Properties of OLS

Lecture 18: Sampling distributions

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Statistical inference: example 1. Inferential Statistics

Estimation for Complete Data

Statistical Theory MT 2008 Problems 1: Solution sketches

Probability and Statistics

Lecture 3. Properties of Summary Statistics: Sampling Distribution

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Kernel density estimator

Chapter 6 Sampling Distributions

Topic 9: Sampling Distributions of Estimators

PRELIM PROBLEM SOLUTIONS

Point Estimation: properties of estimators 1 FINITE-SAMPLE PROPERTIES. finite-sample properties (CB 7.3) large-sample properties (CB 10.

Surveying the Variance Reduction Methods

Statistical Theory MT 2009 Problems 1: Solution sketches

Sieve Estimators: Consistency and Rates of Convergence

Sequences and Series of Functions

INFINITE SEQUENCES AND SERIES

(6) Fundamental Sampling Distribution and Data Discription

Stat 421-SP2012 Interval Estimation Section

Solutions to Homework 7

Lecture 23: Minimal sufficiency

Sequences A sequence of numbers is a function whose domain is the positive integers. We can see that the sequence

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

Topic 9: Sampling Distributions of Estimators

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

Stochastic Simulation

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS

11 THE GMM ESTIMATION

Introduction to Extreme Value Theory Laurens de Haan, ISM Japan, Erasmus University Rotterdam, NL University of Lisbon, PT

Limit Theorems. Convergence in Probability. Let X be the number of heads observed in n tosses. Then, E[X] = np and Var[X] = np(1-p).

Last Lecture. Wald Test

Lecture 7: Properties of Random Samples

Extreme Value Theory in Civil Engineering

Regression with an Evaporating Logarithmic Trend

Joint Probability Distributions and Random Samples. Jointly Distributed Random Variables. Chapter { }

Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Transcription:

Large Sample Theory Covergece Covergece i Probability Covergece i Distributio Cetral Limit Theorems Asymptotic Distributio Delta Method

Covergece i Probability A sequece of radom scalars {z } = (z 1,z, ) coverges i probability to z (a costat or a radom variable) if, for ay e>0, lim Prob( z - z >e) = 0. z is the probability limit of z ad is writte as: plim z = z or z p z. Extesio to a sequece of radom vectors or matrices: elemet-by-elemet covergece i probability, z p z.

Covergece i Probability A special case of covergece i probability is mea square covergece: if E(z ) = m ad Var(z ) = s such that m z ad s 0, the z coverges i mea square to z, ad z p z. What is the differece betwee E(z ) ad plim z or z p z? Mea square covergece is sufficiet (ot ecessary) for covergece i probability.

Covergece i Probability Example: Sample: x { x, x, } ~ ( ms, ) 1 1 Sample Mea: x x i 1 i Sequece of Sample Meas: { x}, E( x ) m m x m p s Var( x ) 0

Almost Sure Covergece A sequece of radom scalars {z } = (z 1,z, ) coverges almost surely to z (a costat or a radom variable) if, Prob(lim z =z) = 1. Write: z as z. If a sequece coverges almost surely, the it coverges i probability. That is, z as z z p z. Extesio to a sequece of radom vectors or matrices: elemet-by-elemet almost sure covergece. I particular, z as z z p z.

Let Laws of Large Numbers LLN cocer coditios uder which the sequece coverges i probability. Chebychev s LLN: [ lim E(z) m, lim Var(z ) 0] z p z 1 z i i 1 { z } μ

Laws of Large Numbers Kolmogorov s LLN: Let {z i } be i.i.d. with E(z i ) = m (the variace does ot eed to be fiite). The z as μ This implies z p μ

Speed of Covergece Order of a Sequece Little oh o(.): Sequece z is o( ) (order less tha ) if ad oly if - z 0. Example: z = 1.4 is o( 1.5 ) sice -1.5 z = 1 /.1 0. Big oh O(.): Sequece z is O( ) if ad oly if - z a fiite ozero costat. Example 1: z = ( + + 1) is O( ). Example : i x i is usually O( 1 ) sice this is the mea of x i ad the mea of x i geerally coverges to E[x i ], a fiite costat. What if the sequece is a radom variable? The order is i terms of the variace. Example: What is the order of the sequece i radom samplig? Because Var[ ] = σ / which is O(1/) x x

Covergece i Distributio Let {z } be a sequece of radom scalars ad F be the c.d.f. of z. {z } coverges i distributio to a radom scalar z if the c.d.f. F of z coverges to the c.d.f. F of z at every cotiuity poit of F. That is, z d z. F is the asymptotic or limitig distributio of z. z p z z d z, or z a F(z)

Covergece i Distributio The extesio to a sequece of radom vectors: z d z if the joit c.d.f. F of the radom vector z coverges to the joit c.d.f. F of z at every cotiuity poit of F. However, elemet-by-elemet covergece does ot ecessarily mea joit covergece.

Cetral Limit Theorems CLT cocer about the limitig behavior of z m Note: m blow up by E(z ). E(z i.i.d. Lideberg-Levy CLT (multivariate): Let {z i } be i.i.d. with E(z i ) = m ad Var(z i ) =. The i ) if z i d i 1 i is 1 z ( μ) ( z μ) N (0, Σ)

Cetral Limit Theorems Lideberg-Levy CLT (uivariate): If z ~ (m,s ), ad {z 1,z,...,z } are radom sample. Defie z z 1 i 1 z, the ( m) d N(0, s ) i

Cetral Limit Theorems Lideberg-Feller CLT (uivariate): If z i ~ (m i,s i ), i=1,,,. Let m 1 1 m s s s, 1 i ad i i 1 i If o sigle term domiates this average variace, the z ( m) d N(0, s )

Asymptotic Distributio A asymptotic distributio is a fiite sample approximatio to the true distributio of a radom variable that is good for large samples, but ot ecessarily for small samples. Stabilizig trasformatio to obtai a limitig distributio: Multiply radom variable x by some power, a, of such that the limitig distributio of a x has a fiite, ozero variace. x Example, has a limitig variace of zero, sice the variace is σ /. But, the variace of x is σ. However, this does ot stabilize the distributio because E( x ) m. The stabilizig trasformatio would be x ( m)

Asymptotic Distributio Obtaiig a asymptotic distributio from a limitig distributio: Obtai the limitig distributio via a stabilizig trasformatio. Assume the limitig distributio applies reasoably well i fiite samples. Ivert the stabilizig trasformatio to obtai the asymptotic distributio.

Asymptotic Distributio Example: Asymptotic ormality of a distributio. From (x m) / s N[0,1] s (x m) N[0, s ] a (x m) N[0, s / ] a x N[ m, s / ] Asymptotic distributio. a / asymptotic variace of x. d

Asymptotic Efficiecy Compariso of asymptotic variaces How to compare cosistet estimators? If both coverge to costats, both variaces go to zero. Example: Radom samplig from the ormal distributio, Sample mea is asymptotically N[μ,σ /] Media is asymptotically N[μ,(π/)σ /] Mea is asymptotically more efficiet.

Covergece: Useful Results Multivariate Covergece i Distributio Let {z } be a sequece of K-dimesioal radom vectors. The: z d z l z d l z for ay K-dimesioal vector of l real umbers.

Covergece: Useful Results Slutsky Theorem: Suppose a(.) is a scalaror vector-valued cotiuous fuctio that does ot deped o : z p a a(z ) p a(a) z d z a(z ) d a(z) x d x, y p a x +y d x+a x d x, y p 0 y x p 0

Covergece: Useful Results Slutsky results for matrices: A p A (plim A = A), B p B (plim B = B), (elemet by elemet) plim (A -1 ) = [plim A ] -1 = A -1 plim (A B ) = (plima )(plimb ) = AB

Covergece: Useful Results x d x, A p A A x d Ax I particular, if x~n(0, ), the A x d N(0,A A ) x d x, A p A x A -1 x d x A -1 x

Delta Method Suppose {x } is a sequece of K- dimesioal radom vector such that x b ad (x -b) d z. Suppose a(.): R K R r has cotiuous first derivatives with A(b) defied by a(β) A( β) β The [a(x )-a(b)] d A(b)z '

Delta Method (x -b) d N(0, ) [a(x )-a(b)] d N(0, A(b) A(b) )

Delta Method Example a x N[ m, s / ] What is the asymptotic distributio of f(x )=exp(x ) or f(x )=1/x (1) Normal sice x is asymptotically ormally distributed () Asymptotic mea is f( m)=exp( m) or 1/ m. (3) For the variace, we eed f'( m) =exp( m) or -1/ m 4 Asy.Var[f(x )]= [exp( m)] s / or [1/ m ] s /