MA Advanced Econometrics: Properties of Least Squares Estimators

Similar documents
Asymptotic Results for the Linear Regression Model

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

Properties and Hypothesis Testing

Lecture 3. Properties of Summary Statistics: Sampling Distribution

CEU Department of Economics Econometrics 1, Problem Set 1 - Solutions

LECTURE 8: ASYMPTOTICS I

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. Comments:

Large Sample Theory. Convergence. Central Limit Theorems Asymptotic Distribution Delta Method. Convergence in Probability Convergence in Distribution

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

TAMS24: Notations and Formulas

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Distribution of Random Samples & Limit theorems

Statistical Properties of OLS estimators

Parameter, Statistic and Random Samples

Random Variables, Sampling and Estimation

Lecture 19: Convergence

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

Economics 102C: Advanced Topics in Econometrics 4 - Asymptotics & Large Sample Properties of OLS

11 Correlation and Regression

Lecture 33: Bootstrap

Lecture 01: the Central Limit Theorem. 1 Central Limit Theorem for i.i.d. random variables

Lecture 20: Multivariate convergence and the Central Limit Theorem

1 Inferential Methods for Correlation and Regression Analysis

ECE 330:541, Stochastic Signals and Systems Lecture Notes on Limit Theorems from Probability Fall 2002

Lecture Chapter 6: Convergence of Random Sequences

Exponential Families and Bayesian Inference

Estimation of the Mean and the ACVF

1 General linear Model Continued..

Kernel density estimator

(all terms are scalars).the minimization is clearer in sum notation:

Module 1 Fundamentals in statistics

Lesson 11: Simple Linear Regression

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS

Kolmogorov-Smirnov type Tests for Local Gaussianity in High-Frequency Data

Binomial Distribution

Regression with an Evaporating Logarithmic Trend

Matrix Representation of Data in Experiment

The central limit theorem for Student s distribution. Problem Karim M. Abadir and Jan R. Magnus. Econometric Theory, 19, 1195 (2003)

Lecture 12: November 13, 2018

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering

Mathematical Statistics - MS

Topic 9: Sampling Distributions of Estimators

Probability and Statistics

Notes 19 : Martingale CLT

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015

3/3/2014. CDS M Phil Econometrics. Types of Relationships. Types of Relationships. Types of Relationships. Vijayamohanan Pillai N.

SDS 321: Introduction to Probability and Statistics

Efficient GMM LECTURE 12 GMM II

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 21 11/27/2013

Lecture 8: Convergence of transformations and law of large numbers

Convergence of random variables. (telegram style notes) P.J.C. Spreij

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

In this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data.

1 Introduction to reducing variance in Monte Carlo simulations

Table 12.1: Contingency table. Feature b. 1 N 11 N 12 N 1b 2 N 21 N 22 N 2b. ... a N a1 N a2 N ab

Eco411 Lab: Central Limit Theorem, Normal Distribution, and Journey to Girl State

32 estimating the cumulative distribution function

Linear Regression Models

11 THE GMM ESTIMATION

Simple Linear Regression

Unbiased Estimation. February 7-12, 2008

Elements of Statistical Methods Lots of Data or Large Samples (Ch 8)

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)

Solution to Chapter 2 Analytical Exercises

Correlation Regression

An Introduction to Asymptotic Theory

x iu i E(x u) 0. In order to obtain a consistent estimator of β, we find the instrumental variable z which satisfies E(z u) = 0. z iu i E(z u) = 0.

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

Final Review. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech

Chapter 6 Sampling Distributions

(X i X)(Y i Y ) = 1 n

t distribution [34] : used to test a mean against an hypothesized value (H 0 : µ = µ 0 ) or the difference

II. Descriptive Statistics D. Linear Correlation and Regression. 1. Linear Correlation

Linear Regression Models, OLS, Assumptions and Properties

Lecture 11 Simple Linear Regression

Approximations and more PMFs and PDFs

ECE 901 Lecture 13: Maximum Likelihood Estimation

ARIMA Models. Dan Saunders. y t = φy t 1 + ɛ t

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator

Law of the sum of Bernoulli random variables

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 6 9/23/2013. Brownian motion. Introduction

Algebra of Least Squares

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d

Expectation and Variance of a random variable

Lecture 10 October Minimaxity and least favorable prior sequences

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett

5.4 The spatial error model Regression model with spatially autocorrelated errors

STA Object Data Analysis - A List of Projects. January 18, 2018

Hypothesis Testing. Evaluation of Performance of Learned h. Issues. Trade-off Between Bias and Variance

CLRM estimation Pietro Coretto Econometrics

Simple Linear Regression

Machine Learning Brett Bernstein

POLS, GLS, FGLS, GMM. Outline of Linear Systems of Equations. Common Coefficients, Panel Data Model. Preliminaries

This section is optional.

Open book and notes. 120 minutes. Cover page and six pages of exam. No calculators.

Transcription:

MA Advaced Ecoometrics: Properties of Least Squares Estimators Karl Whela School of Ecoomics, UCD February 5, 20 Karl Whela UCD Least Squares Estimators February 5, 20 / 5

Part I Least Squares: Some Fiite-Sample Results Karl Whela UCD Least Squares Estimators February 5, 20 2 / 5

Uivariate Regressio Model with Fixed Regressors Cosider the simple regressio model y i = βx i + ɛ i where ɛ i i =, 2,... are i.i.d. ad of mea zero ad with fiite variace. The usual textbook assumptio is that x i is a fixed determiistic regressor idepedet of all values of ɛ i. What does a fixed regressor actually mea? It meas that we are to thik of x i ot as a outcome of a radom process but merely as a fixed set of umbers. The OLS estimator is ˆβ = x iy i x 2 i = x i βx i + ɛ i x 2 i = β + x iɛ i x 2 i We ca show that this is ubiased because we ca treat the x i values as fixed costats, so xi E ˆβ = β + x i 2 E ɛ i = β 3 Karl Whela UCD Least Squares Estimators February 5, 20 3 / 5 2

Multivariate Regressio Model with Fixed Regressors This result geeralises to the multivariate model with k regressors: Y = X β + ɛ 4 where Y ad ɛ are vectors, with ɛ cotaiig i.i.d. errors with fiite variace, X is a k matrix ad β is a k vector of coefficiets. We ca show that the OLS estimators are ubiased as follows E ˆβ = E X X X Y = E X X X X β + ɛ = E β + X X X ɛ 5 6 7 = β + X X X E ɛ 8 = β 9 Karl Whela UCD Least Squares Estimators February 5, 20 4 / 5

Fixed Regressors ad Normal Errors Suppose the i.i.d. error term ɛ is ormally distributed, ɛ N 0, σ 2 I. The formula ˆβ = β + X X X ɛ 0 tells us that, whe we treat the X as a fixed set of umbers, the OLS estimator is a liear fuctio of a ormal variable. So ˆβ is also ormally distributed. We calculate the covariace of the coefficiet estimates as [ E ˆβ β ˆβ β = E X X X ɛɛ X X X ] = σ 2 [ X X X X X X ] 2 So ˆβ N β, σ 2 X X. = σ 2 X X 3 Karl Whela UCD Least Squares Estimators February 5, 20 5 / 5

Stochastic Regressors Model I reality, the data i the X matrix should usually be thought of as outcomes of radom variables, just like the Y variable. As log as the X variables are idepedet of all values i the ɛ matrix, the oe ca proceed i the same way as though the regressors are fixed. For example, oe ca demostrate ubiasedess as E ˆβ = E β + X X X ɛ 4 = β + E X X X E ɛ 5 = β 6 If the errors are ormal, the give the realised value of X oe ca say ˆβ X N β, σ 2 X X 7 This meas that, for each possible X, the distributio of β takes o a particular ormal value, so the usual t ad F statistics also work. That said, the ucoditioal distributio of β depeds upo the distributio of X ad geerally will ot be Gaussia. Karl Whela UCD Least Squares Estimators February 5, 20 6 / 5

More o the Idepedece Assumptio Oe might be tempted to thik that i the model y i = βx i + ɛ i i =, 2... 8 that oe oly eeds E x k ɛ k = 0 for OLS to be ubiased, i.e. that each error term oly eeds to be idepedet of the values of x that occur for the same observatio. However, oe eeds a stroger assumptio: x k eeds to be idepedet of all the ɛ i values for i =, 2...,. From equatio 2 we ca write E ˆβ xi = β + E x i 2 ɛ i 9 Eve if ɛ j is idepedet of x j, it may ot be idepedet of all of the x values that make up the sum x i 2. This ca iduce a correlatio betwee ad ɛ i which leads to OLS beig biased. xi x 2 i We will see that AR time series regressios are a importat example of this kid of bias. Karl Whela UCD Least Squares Estimators February 5, 20 7 / 5

Part II Asymptotic Distributios with Stochastic Regressors Karl Whela UCD Least Squares Estimators February 5, 20 8 / 5

A New Way of Lookig at OLS Estimators You kow the OLS formula i matrix form ˆβ = X X X Y. There is a useful way to restate this that allows us to make a clear coectio to the WLLN ad the CLT. Cosider the case of a regressio with 2 variables ad 3 observatios. The X matrix is thus X = x x 2 x 2 x 22 20 x 3 x 23 This meas we ca write X x x X = 2 x 3 x x 2 x x 2 x 22 x 2 x 22 2 23 x 3 x 23 x = 2 + x 2 2 + x 3 2 x x 2 + x 2 x 22 + x 3 x 23 x 2 x + x 22 x 2 + x 23 x 3 x2 2 + x 22 2 + x 23 2 22 Karl Whela UCD Least Squares Estimators February 5, 20 9 / 5

A New Way of Lookig at OLS Estimators Let x be a colum vector cotaiig the 2 datapoits o the explatory variables from the first observatio x x = 23 x 2 Multplyig this vector by its traspose, we get x x x 2 = x x 2 x 2 x x22 2 24 Goig back ad comparig this with equatio 22 from the previous slide, we get the followig ew way of describig the X X matrix X X = x i x i 25 Karl Whela UCD Least Squares Estimators February 5, 20 0 / 5

A New Formula for the OLS Estimator It follows from the previous slides that we ca re-write the matrix formula for the OLS estimator as ˆβ = X X X Y = x i x i x i y i 26 I the same way, we ca write ˆβ = β + X X X ɛ = β + x i x i x i ɛ i 27 This formula is useful because it explais how the OLS estimator depeds upo sums of radom variables. This allows us to use the Weak Law of Large Numbers ad the Cetral Limit Theorem to establish the limitig distributio of the OLS estimator. Karl Whela UCD Least Squares Estimators February 5, 20 / 5

IID Stochastic Regressors Cosider the case i which ɛ i are i.i.d. with zero mea ad the x variables are i.i.d. each period ad there exists a matrix Q xx such that x i x i p Q xx 28 This meas that x i x i p Q xx 29 We ca also say that E x i ɛ i = 0. We ca use the multivariate CLT to show that d x i ɛ i N 0, Ω 30 where Ω = E x i x i ɛ 2 i 3 Karl Whela UCD Least Squares Estimators February 5, 20 2 / 5

Limitig Distributio with IID Stochastic Regressors The OLS estimator is ˆβ = β + x i x i x i ɛ i 32 with x i x i p Q xx 33 d x i ɛ i N 0, Ω 34 Usig the Slutsky s Theorem results combiig covergece i probability with covergece i distributio, we ca say d ˆβ β N 0, Q xx ΩQxx 35 Karl Whela UCD Least Squares Estimators February 5, 20 3 / 5

Ukow Covariace Matrix Usig the termiology described i the last set of otes, we have show that ˆβ a N β, Q xx ΩQxx 36 However, we do t kow the values of the log-ru average Q xx or the covariace matrix Ω. As described i the last otes, though, we ca substitute for these with cosistet estimators. WLLN tells x i x i p Q xx 37 x i x i ɛ 2 i p Ω 38 This meas we ca substitute sample meas for the true populatio meas ad base tests o the asymptotic distributio ˆβ a N β, x i x i x i x i ɛ 2 i x i x i 39 Karl Whela UCD Least Squares Estimators February 5, 20 4 / 5

No-Idetically Distributed Stochastic Regressors I practise, it s ulikely that idepedet x i observatios will be idetically distributed. What if the x i s are idepedetly but differetly distributed? Oe ca still prove a Cetral Limit Theorem for this case. The previous CLT that we proved, for i.i.d. variables, is kow as the Lidberg-Levy CLT. Aother useful result is the Lidberg-Feller Cetral Limit Theorem: If a sequece of observatios {y, y 2,..., y } are idepedetly distributed with mea µ ad each with a differet fiite variace σi 2 the ȳ µ d N 0, σ 2. 40 where σ 2 = lim σ 2 4 σ 2 = σi 2 42 As usual, there is a atural multivariate extesio of this result i which the covariace matrix of the limitig distributio is a average of the covariace matrices of the differetly distributed x i s. Karl Whela UCD Least Squares Estimators February 5, 20 5 / 5