Large Sample Properties & Simulation

Similar documents
Asymptotic Properties and simulation in gretl

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Econometrics I. Ricardo Mora

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Asymptotic Theory. L. Magee revised January 21, 2013

Applied Econometrics (QEM)

Homoskedasticity. Var (u X) = σ 2. (23)

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Graduate Econometrics I: Asymptotic Theory

Economics 583: Econometric Theory I A Primer on Asymptotics

Instrumental Variables

dqd: A command for treatment effect estimation under alternative assumptions

Introductory Econometrics

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Statistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics Summary Algebraic and Statistical Preliminaries

Applied Econometrics (QEM)

Linear Regression. Junhui Qian. October 27, 2014

Review of Econometrics

Multiple Linear Regression

ECON 3150/4150, Spring term Lecture 6

Statistical Inference with Regression Analysis

Rewrap ECON November 18, () Rewrap ECON 4135 November 18, / 35

Econometrics of Panel Data

Inference in Regression Model

ECON The Simple Regression Model

INTRODUCTION TO BASIC LINEAR REGRESSION MODEL

Final Exam. Economics 835: Econometrics. Fall 2010

An Introduction to Parameter Estimation

ECON 4160, Autumn term Lecture 1

Lecture Tricks with Random Variables: The Law of Large Numbers & The Central Limit Theorem

Motivation Non-linear Rational Expectations The Permanent Income Hypothesis The Log of Gravity Non-linear IV Estimation Summary.

Sample Problems. Note: If you find the following statements true, you should briefly prove them. If you find them false, you should correct them.

1 Motivation for Instrumental Variable (IV) Regression

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22

Recitation 1: Regression Review. Christina Patterson

Linear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons

ECON3150/4150 Spring 2015

Intermediate Econometrics

Econometrics I, Estimation

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Econometrics I Lecture 3: The Simple Linear Regression Model

Lecture 1: August 28

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Econometrics Master in Business and Quantitative Methods

Econometrics I. Introduction to Regression Analysis (Outline of the course) Marijus Radavi ius

Motivation for multiple regression

Lecture notes for probability. Math 124

MS&E 226: Small Data

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Random Variables and Expectations

The regression model with one stochastic regressor.

The Simple Regression Model. Part II. The Simple Regression Model

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Review of Statistics

Men. Women. Men. Men. Women. Women

The Simple Linear Regression Model

Bias Variance Trade-off

11. Further Issues in Using OLS with TS Data

MS&E 226: Small Data

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Econometrics A. Simple linear model (2) Keio University, Faculty of Economics. Simon Clinet (Keio University) Econometrics A October 16, / 11

Economics 241B Estimation with Instruments

A Primer on Asymptotics

Short Questions (Do two out of three) 15 points each

Linear Models in Econometrics

Note on Bivariate Regression: Connecting Practice and Theory. Konstantin Kashin

Topic 12 Overview of Estimation

A Note on the Correlated Random Coefficient Model. Christophe Kolodziejczyk

SIMULATION-BASED SENSITIVITY ANALYSIS FOR MATCHING ESTIMATORS

Regression and Statistical Inference

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

1 Appendix A: Matrix Algebra

MA 575 Linear Models: Cedric E. Ginestet, Boston University Bootstrap for Regression Week 9, Lecture 1

Lecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi)

Lecture 4: Heteroskedasticity

Heteroskedasticity and Autocorrelation

Regression #3: Properties of OLS Estimator

The Statistical Property of Ordinary Least Squares

Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators

Lecture 8 Sampling Theory

Chapter 6: Endogeneity and Instrumental Variables (IV) estimator

Applied Quantitative Methods II

ECO220Y Simple Regression: Testing the Slope

Chapter 7: Section 7-1 Probability Theory and Counting Principles

Econometrics. Week 11. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics - 30C00200

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Probit Estimation in gretl

Advanced Econometrics

MATH Notebook 5 Fall 2018/2019

The regression model with one stochastic regressor (part II)

6 The normal distribution, the central limit theorem and random samples

2. Linear regression with multiple regressors

Econometrics Lecture 1 Introduction and Review on Statistics

Intermediate Econometrics

Environmental Econometrics

Transcription:

Large Sample Properties & Simulation Quantitative Microeconomics R. Mora Department of Economics Universidad Carlos III de Madrid

Outline Large Sample Properties (W App. C3) 1 Large Sample Properties (W App. C3) 2 3

The Notion of Large Sample Properties large sample properties look at how an estimate ˆθ of a parameter θ behave as the sample size gets larger and larger: 1 how far is ˆθ from the true parameter θ as n? 2 how does the distribution of ˆθ look as n? accordingly, we look at two notions of large sample behavior: 1 convergence in probability 2 convergence in distribution

Convergence in Probability Denition As the sample size grows, any arbitrarily small dierence between ˆθ and θ becomes arbitrarily unlikely ) Technically: Pr( ˆθ θ > ε 0 as n θ is probability limit of ˆθ ˆθ converges in probability to θ plim(ˆθ) = θ

A Graphical Interpretation of Consistency Source: Wooldrigde (2003)

The Law of Large Numbers: Some Basic Info in its simplest version, rst proved by Bernoulli in 1713: it took him 20 years to get the actual proof it essentially states that the average converges in probability to the expected value the LLN is important because it guarantees stable long-term results for random events

A Law of Large Numbers For any random variable y with an expected value µ dene the average of a sample of size n as y n. Then plim (y n ) = µ Example 1 plim ( cov ˆ n (y,x)) = cov(y,x) Example 2 plim ( var ˆ n (x)) = var(x)

plim Properties Continuous Mapping Theorem For every continuous function g( ) and random variable x: plim (g(x)) = g(plim(x)) Example 1plim(x + y) = plim(x) + plim(y) ( Example 2plim x y ) = plim(x) plim(y) if plim(y) 0

Example 1: The Fundamental Theorem of Statistics Suppose that X is a random variable with CDF F (X ) and that we obtain a random sample of size n where typical element x i is an independent realization of X The empirical distribution is the discrete distribution that puts a weight of 1 n at each of the x i, i = 1,..., n The EDF is the distribution function of the empirical distribution: ˆF (x) 1 n n i=1 where I ( ) is the indicator function. The Fundamental Theorem of Statistics plim ˆF (x) = F (x) I (x i x)

14 8 Hypothesis Testing in Linear Regression Models 1.00... 0.90 0.80 0.70 0.60 0.50 n = 20 n = 100 n = 500 0. 40 0.30 0.20 0. 10 0.00 3.0 2.0 1.0 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 Figure 4.6 EDFs for several sample sizes EDFs for three samples of sizes 20, 100, and 500 drawn from three normal distributions, each with variance 1 and with means 0, 2, and 4, respectively where the x t are independent R. random Mora variables, QMicro: Large each Sample with its Properties own bounded & Simulation

Example 2: OLS under Classical Assumptions Gauss-Markov Assumptions A1: Linearity: y = β + β 1 x 1 +... + β k x k + v A2: Random Sampling A3: Conditional Mean Independence: E[y x] = β 0 + β 1 x 1 +... + β k x k A4: Invertibility of Variance-covariance Matrix A5: Homoskedasticity: Var[v x] = σ 2 Normality A6: Normality: y x N(β 0 + β 1 x 1 +... + β k x k,σ 2 )

OLS Consistency Theorem Under Gauss-Markov A1-A4, OLS is consistent Example: wages = β 0 + β 1 educ + u with cov(educ, u) = 0 ˆβ 1 = β 1 + cov(educ ˆ i,u i ) var(educ ˆ i ) ( ) plim ˆβ1 = plim (β 1 ) + plim( ˆ Since cov(educ, u) = 0 plim cov(educ i,u i )) plim( var(educ ˆ i )) ( ˆβ1 ) = β 1 = β 1 + cov(educ,u) var(educ)

An Example of Inconsistency True Model: wages = β 0 + β 1 educ + β 2 IQ + v cov(educ, v) = cov(iq, v) = 0 cov(educ, IQ) 0, β 2 0 Estimated equation by OLS: wages = ˆγ 0 + ˆγ 1 educ + û educ ˆγ 1 = ˆβ 1 + ˆβ 2 cov(educ,iq) ˆ var(educ) ˆ plim (ˆγ 1 ) β 1 if intelligence is relevant:β 2 0 plim (ˆγ 1 ) = β 1 + β 2 cov(educ,iq) var(educ) education is correlated to intelligence: cov(educ, IQ) 0

Asymptotic Normality Denition As the sample size grows, the distribution of ˆβ j gets arbitrarily close to the normal distribution Technically: Pr( ˆβ j z) Φ(z) as n

The Central Limit Theorem: Some History arguably, one of the most interesesting laws in maths, the proof is astonishing simple ( but I must admit that I cannot intuitively understand the result) the rst who thought about it was a French mathematician, de Moivre, in 1733 Pierre Simon Laplace, another French, got the simplest version right in 1812 the Russian Aleksander Liapunov was the guy who proved the general case in 1901

Rats and the Central Limit Theorem go to the sewage system in Madrid capture 100 rats, measure their tails,standardize the measures, and compute the average times square root of 100 now capture more rats, say 500 rats, do as before using the square root of 500 instead of the square root of 100 if the distribution of the second average is closer to the standard normal, that's basically it otherwise, instead of 500, try with a larger number of rats, say 10000 the point is: if you keep increasing the sample size, you will CERTAINLY get closer to the normal

Stars and the Central Limit Theorem look at the brightness of 100 stars yeah, that's it! you will CERTAINLY get as close as you want to the normal after averaging by increasing sample size what is remarkable about the CLT is that the random distributions of star's brightness and rat's tails have nothing to do with each other the crucial issue is the averaging carried out over the measures after measurement

The Central Limit Theorem For any random variable y with an expected value µ and variance σ 2 dene the average of a sample of size n as y n. Then n 1 /2 y n µ σ N(0,1)as n Under Gauss-Markov A.1 to A.5 n 1 /2 ˆβ j β j σ/a j N(0,1)as n where a j ² = plim ( 1 n i res ˆ ji ² ) Moreover: plim ( ˆσ ²) = ( ) SSR plim n k 1 = σ 2

Asymptotic Normality for OLS Estimators As sample size n increases, the OLS estimatorsconveniently scaled upget as close as we want to a normal distribution n 1 /2 ˆβj N (β j,σ²a j ²) the larger the sample, the more accurate the estimates important: the expression for the asymptotic variance depends on the homoskedasticity assumption

The t Test Large Sample Properties (W App. C3) from the CLT (and a LLN), it can also be shown that t = ˆβ j β ( ) j N (0,1) as n se ˆβj this result can be used to test whether a coecient is signicant

Small Sample Properties the idea is to undertand the behavior of an estimator for a given xed sample size n Under Gauss-Markov A.1 to A.5 AND Normality A.6 ˆβ j β j se( ˆβ j ) t n k 1 the problem is, normality is a very strong assumption

Monte Carlo Experiments: Denition and Motivation Denition Monte Carlo experiments are a type of computational algorithms that rely on repeated random sampling to compute their results Monte Carlo methods are often used in simulating physical and mathematical systems Tend to be used when it is unfeasible or impossible to compute an exact result with a deterministic algorithm We are going to use them to generate observational data to study the behaviour of econometric techniques when samples are not innite

Why Monte Carlo? the term was rst used in the 1940s by physicists working on nuclear weapons in the US they wanted to solve a radiation problem but could not do it analytically, so they decided to model the experiment using chance they codenamed the project Monte Carlo in reference to the Monte Carlo Casino in Monaco where the uncle of one of them would borrow money to gamble it has been since used in many sciences for economists, Monte Carlo techniques are important for solving typical problems (integration, optimization) and also in games and in applied statistics

Monte Carlo in Applied Statistics there are two main uses of Monte Carlo in applied statistics to compare and contrast competing statistics for small samples to study how small sample behaviour translates into asymptotic results as samples become larger Monte Carlo techniques strike a nice balance usually, they are more accurate than asymptotic results not as time consuming to obtain as are exact tests

Monte Carlo vs. Simulation a simulation is a ctitious computer representation of reality a Monte Carlo study is a technique that can be used to solve a mathematical or statistical problem A Monte Carlo simulation uses repeated sampling of simulated data to determine the properties of some phenomenon

A Simple Simulation Exercise 1 using a computer, draw a value from a pseudo-random uniform variate from the interval [0,1] 2 If the value is less than or equal to 0.50 designate the outcome as heads 3 if the value is greater than 0.50 designate the outcome as tails this is a simulation of the tossing of a coin

A Simple Monte Carlo Study the area of an irregular gure inscribed in a unit square 1 draw two values from a pseudo-random uniform variate from the interval [0,1] 2 If the point identied is within the gure, designate the outcome as success, otherwise, as failure 3 repeat steps 1 and 2 many times 4 the proportion of successes provides the area of the gure this is using simple Monte Carlo techniques to compute a complex integral

A Simple Monte Carlo Experiment 1 draw one value from a pseudo-random uniform variate from the interval [0,1] 2 If the value is less than or equal to 0.50 designate the outcome as heads, otherwise tails 3 repeat steps 1 and 2 many times 4 the proportion of heads is the Monte Carlo simulation of the probability of heads

Using Monte Carlo in Applied Statistics we assume an econometric model and simulate it may times from each simulated population, we can extract a sample of size n and estimate the parameter of interest by looking at the descriptive statistics of the estimates across all simulated realities, we estimate the properties of the estimator when the sample size is xed by increasing n and doing everything again, we estimate how the estimator behaves when the sample size increases

A Basic algorithm for a Monte Carlo experiment A Monte Carlo experiment for a xed sample of size N 1 assume values for the exogenous parts of the model or draw them from their respective distribution function 2 draw a (pseudo) random sample of size N for the error terms in the statistical model from their respective probability distribution functions 3 calculate the endogenous parts of the statistical model 4 calculate the value (e.g. the estimate) you are interested in 5 replicate step 1 to 4 R times 6 examine the empirical distribution of the R values

A simple Monte Carlo experiment log(wages) = 10 + 0.05 D + u, u N(0,1), D = 1 with prob. 0.3 1 draw N realizations of D 2 draw N realizations of u 3 compute log(wages) 4 OLS log(wages) on D and store ˆβ r 1 5 replicate step 1 to 4 R times 6 examine the empirical distribution of ˆβ r 1

Large Sample Properties (W App. C3) Large sample properties tell us how an estimator behaves as the sample size becomes arbitrarily large. Exact small sample properties are hard to get and sometimes they require strong assumptions. Monte Carlo simulations are useful in applied statistics. We can study small sample properties of estimators. We can also study how large sample properties are achieved in practice.