Econometrics Part Three

Similar documents
LECTURE 11. Introduction to Econometrics. Autocorrelation

Likely causes: The Problem. E u t 0. E u s u p 0

DEMAND ESTIMATION (PART III)

Applied Econometrics. Applied Econometrics. Applied Econometrics. Applied Econometrics. What is Autocorrelation. Applied Econometrics

Iris Wang.

Answers: Problem Set 9. Dynamic Models

CHAPTER 6: SPECIFICATION VARIABLES

ECON 497: Lecture 4 Page 1 of 1

LECTURE 10: MORE ON RANDOM PROCESSES

ECON 120C -Fall 2003 PROBLEM SET 2: Suggested Solutions. By substituting f=f0+f1*d and g=g0+... into Model 0 we obtain Model 1:

Ref.: Spring SOS3003 Applied data analysis for social science Lecture note

Eco and Bus Forecasting Fall 2016 EXERCISE 2

ECON 4230 Intermediate Econometric Theory Exam

Heteroskedasticity and Autocorrelation

FinQuiz Notes

Reading Assignment. Serial Correlation and Heteroskedasticity. Chapters 12 and 11. Kennedy: Chapter 8. AREC-ECON 535 Lec F1 1

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables.

Autocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time

Agricultural and Applied Economics 637 Applied Econometrics II

VII. Serially Correlated Residuals/ Autocorrelation

F9 F10: Autocorrelation

Diagnostics of Linear Regression

11.1 Gujarati(2003): Chapter 12

Heteroscedasticity 1

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

AUTOCORRELATION. Phung Thanh Binh

1 Graphical method of detecting autocorrelation. 2 Run test to detect autocorrelation

PBAF 528 Week 8. B. Regression Residuals These properties have implications for the residuals of the regression.

Spatial Regression. 10. Specification Tests (2) Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

ECON 497: Lecture Notes 10 Page 1 of 1

1/34 3/ Omission of a relevant variable(s) Y i = α 1 + α 2 X 1i + α 3 X 2i + u 2i

Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation

Rockefeller College University at Albany

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE

Heteroskedasticity. Occurs when the Gauss Markov assumption that the residual variance is constant across all observations in the data set

LECTURE 13: TIME SERIES I

Lecture 6: Dynamic Models

Assumptions of the error term, assumptions of the independent variables

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econ 510 B. Brown Spring 2014 Final Exam Answers

Econometrics Summary Algebraic and Statistical Preliminaries

Final Exam - Solutions

Regression Analysis. BUS 735: Business Decision Making and Research

OSU Economics 444: Elementary Econometrics. Ch.10 Heteroskedasticity

Economics 620, Lecture 13: Time Series I

Week 11 Heteroskedasticity and Autocorrelation

Final Exam. Question 1 (20 points) 2 (25 points) 3 (30 points) 4 (25 points) 5 (10 points) 6 (40 points) Total (150 points) Bonus question (10)

2 Prediction and Analysis of Variance

Finding Relationships Among Variables

Econometrics Review questions for exam

Making sense of Econometrics: Basics

Business Statistics. Lecture 10: Correlation and Linear Regression

Semester 2, 2015/2016

Heteroskedasticity. (In practice this means the spread of observations around any given value of X will not now be constant)

REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK

Freeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94

LINEAR REGRESSION ANALYSIS. MODULE XVI Lecture Exercises

Demand analysis is one of the rst topics come to in economics. Very important especially in the Keynesian paradigm.

Chapter 5. Classical linear regression model assumptions and diagnostics. Introductory Econometrics for Finance c Chris Brooks

Exercise Sheet 6: Solutions

Contest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2.

Introduction to Econometrics. Heteroskedasticity

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University

SIMULTANEOUS EQUATION MODEL

Lecture 4: Heteroskedasticity

Answers to Problem Set #4

Heteroskedasticity. Part VII. Heteroskedasticity

ECNS 561 Multiple Regression Analysis

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Økonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning

Types of economic data

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

Microeconometria Day # 5 L. Cembalo. Regressione con due variabili e ipotesi dell OLS

Lab 11 - Heteroskedasticity

Economics 308: Econometrics Professor Moody

Instrumental Variables, Simultaneous and Systems of Equations

Exercise E7. Heteroskedasticity and Autocorrelation. Pilar González and Susan Orbe. Dpt. Applied Economics III (Econometrics and Statistics)

WORKSHOP. Introductory Econometrics with EViews. Asst. Prof. Dr. Kemal Bağzıbağlı Department of Economic

Topic 7: HETEROSKEDASTICITY

Auto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e.,

Answer all questions from part I. Answer two question from part II.a, and one question from part II.b.

Linear Model Under General Variance Structure: Autocorrelation

Econometric Forecasting Overview

Spurious regressions PE The last lecture discussed the idea of non-stationarity and described how many economic time series are non-stationary.

Heteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance.

School of Mathematical Sciences. Question 1

ECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 48

df=degrees of freedom = n - 1

Bivariate Relationships Between Variables

GLS and related issues

Statistics for Managers using Microsoft Excel 6 th Edition

Modelling the Electric Power Consumption in Germany

ECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 54

CHAPTER 21: TIME SERIES ECONOMETRICS: SOME BASIC CONCEPTS

Lecture 8. Using the CLR Model. Relation between patent applications and R&D spending. Variables

Empirical Economic Research, Part II

Final Exam. 1. Definitions: Briefly Define each of the following terms as they relate to the material covered in class.

Spatial Regression. 9. Specification Tests (1) Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Transcription:

!1 I. Heteroskedasticity A. Definition 1. The variance of the error term is correlated with one of the explanatory variables 2. Example -- the variance of actual spending around the consumption line increases as income increases Consumption a)! B. Effect 1. The b's are still unbiased, but the standard errors (the estimated standard deviation for each b, or Sb) are biased (the reported value is higher or lower than the actual value). a) With positive heteroskedasticity (variance increases with X, as in the example above), the reported standard error is low, increasing the t- statistic. This may cause you to reject the hypothesis that β = 0 when it is true. b) The opposite is true with negative heteroskedasticity. C. Testing for heteroskedasticity Income 1. Graph the regression residuals on the Y axis each independent variable on the X axis and look for differences in the amount of variation in the residuals. If there is no heteroskedasticity, it will look like this: Residuals 0 Income a) b) If there is a small number of cases at one end of the X axis, the amount of variation will appear smaller because the probability of a large or small

!2 II. D. Correcting value is low in a small sample. This is not strong evidence of heteroskedasticity. 1. One possible solution is to change the dependent variable from a total to a rate. Instead of explaining the amount spent at different levels of income, explain the percentage of income spent at different levels of income. Non-normal errors A. OLS is based on the assumption that the errors are normally distributed. This means that most of the errors are close to zero, that is, most of the data points are close to the actual line. This is the rationale for estimating the line as close as possible to the data points. B. The test for statistical significance for each coefficient is based on the assumption of normally distributed errors. The estimated coefficients have a t distribution only if the errors have a normal distribution. If the errors are not normal, then the t statistic does not correctly estimate the probability that the variable has no effect. C. To test for normality of the errors, SPSS can plot the histogram for the residuals. If it is approximately normal, then OLS is appropriate to use. III. Simultaneous Equations A. Example -- supply and demand 1. Simple model a) S: Q = a + b P + ε b) D: Q = c - d P + µ 2. In this case, errors (ε and µ) represent random shifts of the supply and demand curve, respectively. Thus, the observations would be a 'cloud' of points around the stable equilibrium (where a + bp = c - dp). Fitting a regression line to these points will estimate neither the demand or supply curves. 3. If errors are measured vertically (in $s), then the estimated line will be horizontal). If the errors are measured horizontally (in units of the product, as in the two equations above), then the estimated line will be vertical. P estimated line S D a)! Q

!3 4. To estimate the supply curve, you need something (like a change in income) to shift the demand curve. a) It is important that whatever shifts the demand curve does not also shift the supply curve. P S estimated line D D' b)! 5. More complex model a) S: Q = a + b P + c P r + ε b) D: Q = d - e P + f I + µ 6. Able to estimate both S and D equations a) S is identified because it does not include I (to shift D) b) D is identified because it does not include P r (to shift S) B. In general, an equation is identified if it does not include the same variables that shift the other equations. 1. Rule is that the number of excluded variables in an equation is at least as big as the number of equations - 1. 2. If there are two equations, that means that the supply equation is identified because it does not contain the I variable, while the demand equation is identified because it does not contain the P r variable. IV. Autocorrelation A. Definition Q 1. In general, autocorrelation occurs when a variable from period t is correlated with the same variable from period t-n. Most common is first degree autocorrelation, where Z t is correlated with Z t-1 (Z in the previous period). a) Positive autocorrelation = if Z t-1 is large then Z t will also be large (same sign). b) Negative autocorrelation = if Z t-1 is large and positive then Z t will be large and negative. 2. The problem with OLS occurs when the error term (ε) is autocorrelated. a) For example ε t = ρ ε t-1 + µ t, which means that a certain fraction (ρ) of the error in the previous period (ε t-1 ) carries over to the current error (ε t ), plus an additional random effect (µ t ).

!4 B. Using OLS when the errors are autocorrelated results in 1. unbiased estimated coefficients, but 2. inefficiency = the estimated coefficients are less likely to be close to the actual coefficients (bigger standard deviations), and 3. biased standard errors = the reported standard deviation of each estimated coefficient is likely to be too high or too low, depending on the type of autocorrelation. C. Cause a) With positive autocorrelation, the reported standard errors are understated, resulting in overly large t statistics (so may reject null hypothesis when it is true). (1) Don't ask why -- there is no obvious reason! b) With negative autocorrelation, the standard errors are overstated, resulting in too small t statistics. 1. Usually due to some form of inertia on the part of the dependent variable = pure ac. If it above the predicted value in one period, it will probably still be above the predicted value next period. 2. May be due to the omission of an independent variable that is itself autocorrelated = impure ac. This is even worse than simple autocorrelation, since the coefficient estimates will also be biased and inconsistent. D. Testing for autocorrelation 1. Have the computer calculate the Durbin-Watson statistic a) If there is no autocorrelation, then d 2. b) For perfect positive autocorrelation (identical errors over time), then d 0. c) For perfect negative autocorrelation (errors have same magnitude but opposite signs), then d 4. 2. Test the null hypothesis Ho: no positive autocorrelation against the alternative hypothesis of positive autocorrelation. a) The idea is to reject the hypothesis if d is close to 0 and accept it if d is close to 2. To do this, we need to (1) decide what close means and (2) decide what to do if d is not close to 0 or 2 (like d=1). (1) For specific n and k, the Durbin-Watson tables give d u and d L for a given level of significance (usually 5% or 1%) (2) If d u < d < 2, then accept the null hypothesis (since d is "close" to 2). (3) If d < d L, then reject the null hypothesis and accept alternative hypothesis of positive autocorrelation. (4) If d L < d < d u, then the test is inconclusive

!5 3. If d > 2, so suspect negative autocorrelation, test Ho: no negative autocorrelation using 4 d u as the lower bound and 4 - d L as the upper bound (the mirror images of d u and d L around 2) 4. Warning: d is not appropriate if lagged dependent variable is used as an explanatory variable. E. Example 1. Suppose that d = 1.65 when n=50 and k=3. Using the 5% significance level, d L = 1.421 and d u = 1.674 from the DW tables. Since d is between those values, the test is inconclusive. If it had been smaller than 1.421, then would conclude that there is positive autocorrelation with 95% certainty. If it was bigger than 1.674, would conclude that there is no autocorrelation, again with 95% certainty.