Interpreting Regression Results -Part II

Similar documents
Interpreting Regression Results

The Linear Regression Model

Model Mis-specification

Heteroscedasticity and Autocorrelation

Reference: Davidson and MacKinnon Ch 2. In particular page

Regression: Ordinary Least Squares

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

Economics 240A, Section 3: Short and Long Regression (Ch. 17) and the Multivariate Normal Distribution (Ch. 18)

In the bivariate regression model, the original parameterization is. Y i = β 1 + β 2 X2 + β 2 X2. + β 2 (X 2i X 2 ) + ε i (2)

The Simple Regression Model. Simple Regression Model 1

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

LECTURE 5 HYPOTHESIS TESTING

coefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1

1 Appendix A: Matrix Algebra

Stat 5101 Lecture Notes

Financial Econometrics Lecture 6: Testing the CAPM model

Lecture 6: Geometry of OLS Estimation of Linear Regession

Lecture 5. In the last lecture, we covered. This lecture introduces you to

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Essential of Simple regression

Ch 3: Multiple Linear Regression

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Answers to Problem Set #4

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Problem Set - Instrumental Variables

Part 6: Multivariate Normal and Linear Models

Econometric Methods. Prediction / Violation of A-Assumptions. Burcu Erdogan. Universität Trier WS 2011/2012

ECON 5350 Class Notes Functional Form and Structural Change

ECON3150/4150 Spring 2015

The regression model with one stochastic regressor.

Simple Linear Regression Model & Introduction to. OLS Estimation

Non-Spherical Errors

ECON3150/4150 Spring 2016

Regression Analysis. y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T,

FENG CHIA UNIVERSITY ECONOMETRICS I: HOMEWORK 4. Prof. Mei-Yuan Chen Spring 2008

Introduction to Estimation Methods for Time Series models. Lecture 1

1. The Multivariate Classical Linear Regression Model

Birkbeck Working Papers in Economics & Finance

AMS-207: Bayesian Statistics

Econ 2120: Section 2

Linear Regression with Time Series Data

18.S096 Problem Set 3 Fall 2013 Regression Analysis Due Date: 10/8/2013

Spatial Regression. 9. Specification Tests (1) Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

1. The OLS Estimator. 1.1 Population model and notation

Next is material on matrix rank. Please see the handout

The Statistical Property of Ordinary Least Squares

Multivariate Tests of the CAPM under Normality

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

ANCOVA. ANCOVA allows the inclusion of a 3rd source of variation into the F-formula (called the covariate) and changes the F-formula

Instrumental Variables and the Problem of Endogeneity

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Week 3: Simple Linear Regression

R = µ + Bf Arbitrage Pricing Model, APM

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -

Greene, Econometric Analysis (7th ed, 2012)

Lecture 5: A step back

Final Exam. Economics 835: Econometrics. Fall 2010

Homework Set 2, ECO 311, Fall 2014

Regression #2. Econ 671. Purdue University. Justin L. Tobias (Purdue) Regression #2 1 / 24

Introduction to Econometrics

We begin by thinking about population relationships.

SEM with observed variables: parameterization and identification. Psychology 588: Covariance structure and factor models

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

5. Let W follow a normal distribution with mean of μ and the variance of 1. Then, the pdf of W is

Recitation 1: Regression Review. Christina Patterson

ECON The Simple Regression Model

GMM Estimation in Stata

Linear Regression with Time Series Data

Ross (1976) introduced the Arbitrage Pricing Theory (APT) as an alternative to the CAPM.

Lecture 2: Univariate Time Series

PhD/MA Econometrics Examination. January, 2015 PART A. (Answer any TWO from Part A)

Instrumental Variables

Linear Regression. Junhui Qian. October 27, 2014

ECON 836 Midterm 2016

Classical Least Squares Theory

Hypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima

Lecture 4: Multivariate Regression, Part 2

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model

Ordinary Least Squares Regression

Simple Linear Regression: The Model

Multivariate Regression: Part I

ASSET PRICING MODELS

Lecture 8: Instrumental Variables Estimation

Generalized Method of Moments: I. Chapter 9, R. Davidson and J.G. MacKinnon, Econometric Theory and Methods, 2004, Oxford.

22s:152 Applied Linear Regression

Solutions for Econometrics I Homework No.1

The Bayesian Approach to Multi-equation Econometric Model Estimation

LECTURE 9: GENTLE INTRODUCTION TO

The Simple Regression Model. Part II. The Simple Regression Model

A Short Introduction to the Lasso Methodology

Wiley. Methods and Applications of Linear Models. Regression and the Analysis. of Variance. Third Edition. Ishpeming, Michigan RONALD R.

Multiple Regression Analysis

Motivation for multiple regression

Homework Set 2, ECO 311, Spring 2014

Econometrics of Panel Data

The regression model with one stochastic regressor (part II)

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Optimization Problems

Multiple Linear Regression CIVL 7012/8012

Transcription:

Interpreting Regression Results -Part II Carlo Favero Favero () Interpreting Regression Results -Part II / 9

The Partitioned Regression Model Given the linear model: y = Xβ + ɛ, Partition X in two blocks two blocks of dimension (Txr) and (Tx (k r)) and β in a corresponding way into β β 2. The partitioned regression model can then be written as follows y = X β + X 2 β 2 + ɛ, Favero () Interpreting Regression Results -Part II 2 / 9

The Partitioned Regression Model It is useful to derive the formula for the OLS estimator in the partitioned regression model. To obtain such results we partition the normal equations X X b β = X y as: X X 2 X X 2 b β bβ 2! = X X 2 y, or, equivalently, X X X X 2 X 2 X X 2 X 2 b β bβ 2! = X y X 2 y. () Favero () Interpreting Regression Results -Part II 3 / 9

The Partitioned Regression Model System () can be resolved in two stages by first deriving an expression β b 2 as: bβ 2 = X2 X 2 X2 y X 2 X β b, and then by substituting it in the first equation of () to obtain X X β b + X X 2 X2 X 2 X2 y X 2 X β b = X y, from which: bβ = X M 2X X M 2 y M 2 = I X 2 X2 X 2 X 2. Favero () Interpreting Regression Results -Part II 4 / 9

The Partitioned Regression Model Note that, as M 2 is idempotent, we can also write: bβ = X M 2 M 2X X M 2 M 2y, and b β can be interpreted as the vector of OLS coefficients of the regression of y on the matrix of residuals of the regression of X on X 2. Thus, an OLS regression on two regressors is equivalent to two OLS regressions on a single regressor (Frisch-Waugh theorem). Favero () Interpreting Regression Results -Part II 5 / 9

The Partitioned Regression Model Finally, consider the residuals of the partitioned model: bɛ = y X b β X 2 b β2, bɛ = y X b β X2 X 2 X 2 bɛ = M 2 y M 2 X b β X 2 y X 2 X b β, = M 2 y M 2 X X M 2X X M 2 y = M 2 M 2 X X M 2X X M 2 y, however, we already know that bɛ = My, therefore, M = M 2 M 2 X X M 2X X M 2. (2) Favero () Interpreting Regression Results -Part II 6 / 9

Testing restrictions on a subset of coefficients In the general framework to test linear restrictions we set r =, R = I r, and partition β in a corresponding way into β β 2. In this case the restriction Rβ r = is equivalent to β = in the partitioned regression model. Under H, X has no additional explicatory power for y with respect to X 2, therefore: H : y = X 2 β 2 + ɛ, (ɛ j X, X 2 ) s N, σ 2 I. Note that the statement y = X 2 γ 2 + ɛ, (ɛ j X 2 ) N, σ 2 I, is always true under our maintained hypotheses. However, in general γ 2 6= β 2. Favero () Interpreting Regression Results -Part II 7 / 9

Testing restrictions on a subset of coefficients To derive a statistic to test H remember that the general matrix R (X X) R is the upper left block of (X X), which we can now write as (X M 2X ). The statistic then takes the form bβ (X M 2X ) b β rs 2 = y M 2 X (X M 2X ) X M 2y T y My r k F (T k, r). Given (2), () can be re-written as: y M 2 y y My T y My r k F (T k, r), (3) where the denominator is the sum of the squared residuals in the unconstrained model, while the numerator is the difference between the sum of residuals in the constrained model and the sum of residuals in the unconstrained model. Favero () Interpreting Regression Results -Part II 8 / 9

Testing restrictions on a subset of coefficients Consider the limit case r = and β is a scalar. The F-statistic takes the form bβ 2 s 2 (X M 2X ) s F (T k, r), under H, where (X M 2X ) is element (, ) of the matrix (X X). Using the result on the relation between the F and the Student s t-distribution: bβ s (X M 2X ) /2 t (T k) under H. Therefore, an immediate test of significance of the coefficient can be performed, by taking the ratio of each estimated coefficient and the associated standard error. Favero () Interpreting Regression Results -Part II 9 / 9

The R-squared as a measure of relevance To illustrate the point let us consider two specific cases of applications of the CAPM: r i t r rf t r m t r rf t ui,t u m,t =.8σ m u m,t + σ i u i,t = µ m + σ m u m,t s n.i.d., µ m =.65, σ m =.54, σ =.9, σ 2 =.5 We simulate an artificial sample of 56 obs.(same length with the sample July 926-June24) observations. µ m and σ m are calibrated to match the first two moments of the market portfolio excess returns over the sample 926:7-24:7. The standard errors of the two excess returns are calibrated to deliver R 2 of respectively about.22 and.98. Favero () Interpreting Regression Results -Part II / 9

The R-squared as a measure of relevance By running the two CAPM regressions on the artificial sample: TABLE 3.: The estimation of the CAPM on artificial data Dependent Variable r t r rf t Regressor Coefficient Std. Error t-ratio Prob. r m t r rf t.875 7.48. R 2.22 S.E. of regression.76 r 2 t Dependent Variable r rf t Regressor Coefficient Std. Error t-ratio Prob. r m t r rf t.793 2.86. R 2.972 S.E. of regression. In both cases the estimated beta are statistically significant and very close to their true value of.8. Favero () Interpreting Regression Results -Part II / 9

In both experiments the conditional expectation changes of the same amount but the share of the unconditional variance of y explained by the regression function is very different, as different are the R 2 s. Favero () Interpreting Regression Results -Part II 2 / 9 The R-squared as a measure of relevance Simulate again the processes but introduce at some point a temporary shift of two per cent in the excess returns in the market portfolio..24.2.6.2.8...4 4. 8. 2. 6.3.2....2.3.4.2.5..5.. 5.. 5 Simulated Market Portfolio excess returns 95 95 952 953 954 955 956 957 958 959 96 Baseline Alternative Simulated Portfolio excess returns 95 95 952 953 954 955 956 957 958 959 96 Baseline Alternative Simulated Portfolio 2 excess returns 95 95 952 953 954 955 956 957 958 959 96 Baseline Alternative

The partial regression theorem The Frisch-Waugh Theorem described above is worth more consideration. The theorem tells us than any given regression coefficient in the model E (y j X) = Xβ can be computed in two different but exactly equivalent ways: ) by regressing y on all the columns of X, 2) by first regressing the j-th column of X on all the other columns of X, computing the residuals of this regression and then by regressing y on these residuals. This result is relevant in that it clarifies that the relationships pinned down by the estimated parameters in a linear model do not describe the connections between the regressand and each regressor but the connection between the part of each regressor that is not explained by the other ones and the regressand. Favero () Interpreting Regression Results -Part II 3 / 9

What if analysis The relevant question in this case becomes how much shall y change if I change X i? The estimation of a single equation linear model does not allow to anser that question, for a number of reasons. First, estimated parameters in a linear model can only answer the question how much shall E (y j X) if I change X? We have seen that the two questions are very different if the R 2 of the regression is low, in this case a change in E (y j X) may not effect any visible and relevant effect on y. Second, a regression model is a conditional expected value GIVEN X. In this sense there is no space for changing the value of any element in X. Favero () Interpreting Regression Results -Part II 4 / 9

What if analysis Any statement involving such a change requires some assumption on how the conditional expectation of y changes if X changes and a correct analysis of this requires an assumption on the joint distribution of y and X. Simulation might require the use of the multivariate joint model even when valid estimation can be performed concentrating only on the conditional model. Strong exogeneity is stronger than weak exogeneity for the estimation of the parameters of interest. Favero () Interpreting Regression Results -Part II 5 / 9

What if analysis Think of a linear model with know parameters y = β x + β 2 x 2 What is in this model the effect of on y of changing x by one unit while keeping x 2 constant? Easy β. Now think of the estimated linear model: y = ˆ β x + ˆ β 2 x 2 + û Now y is different from E (y j X) and the question "what is in this model the effect of on E (y j X) of changing x by one unit while keeping x 2 constant?" does not in general make sense. Favero () Interpreting Regression Results -Part II 6 / 9

What if analysis Changing x keeping x 2 unaltered implies that there is zero correlation among this variables. But the estimates β ˆ and β ˆ 2 are obained by using data in which in general there is some correlation between x and x 2. Data in which fluctuations in x do not have any effect on x 2 would have most likely generated different estimates from those obtained in the estimation sample. The only valid question that can be answered using the coefficients in linear regression is "What is the effect on E (y j X) of changing the part of each regressors that is orthogonal to the other ones". "What if" analysis requires simulation and in most cases a low level of reduction than that used for regression analysis. Favero () Interpreting Regression Results -Part II 7 / 9

The semi-partial R-squared When the columns of X are orthogonal to eache other the total R 2 can be exactly decomposed in the sum of the partial R 2 due to each regressor x i (the partial R 2 of a regressor i is defined as the R 2 of the regression of y on x i ). This is in general not the case in applications with non experimental data: columns of X are correlated and a (often large) part of the overall R 2 does depend on the joint behaviour of the columns of X. However, it is always possible to compute the marginal contribution to the overall R 2 due to each regressor x i, defined as the difference between the overall R 2 and the R 2 ot the regression that includes all columns X except x i. This is called the semi-partial R 2. Favero () Interpreting Regression Results -Part II 8 / 9

The semi-partial R-squared Interestingly, the the semi-partial R 2 is a simple tranformation of the t-ratio: spr 2 i = t2 β i R 2 (T k) This result has two interesting implications. First, a quantity which we considered as just a measure of statistical reliability, can lead to a measure of relevance when combined with the overall R 2 of the regression. Second, we can re-iterate the difference between statistical significance and relevance. Suppose you have a sample size of and you have columns in X and the t-ratio on a coefficient β i is of about 4 with an associate P-value of the order.: very statistical significant! The derivation of the semi-partial R 2 tells us that the contribution of this variable to the overall R2 is at most approximately 6/(-) that is: less than two thousands. Favero () Interpreting Regression Results -Part II 9 / 9