Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis. 17th Class 7/1/10

Similar documents
Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation

11.1 Gujarati(2003): Chapter 12

at least 50 and preferably 100 observations should be available to build a proper model

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Regression of Time Series

FinQuiz Notes

Section 2 NABE ASTEF 65

Time Series Analysis -- An Introduction -- AMS 586

Lecture 6: Dynamic Models

Statistics 910, #5 1. Regression Methods

F9 F10: Autocorrelation

LECTURE 11. Introduction to Econometrics. Autocorrelation

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

Applied Econometrics. Applied Econometrics. Applied Econometrics. Applied Econometrics. What is Autocorrelation. Applied Econometrics

INTRODUCTORY REGRESSION ANALYSIS

Introduction to Regression Analysis. Dr. Devlina Chatterjee 11 th August, 2017

FAQ: Linear and Multiple Regression Analysis: Coefficients

FinQuiz Notes

Forecasting. Simon Shaw 2005/06 Semester II

Time Series. April, 2001 TIME SERIES ISSUES

Lecture 5: Omitted Variables, Dummy Variables and Multicollinearity

Likely causes: The Problem. E u t 0. E u s u p 0

ECON 4230 Intermediate Econometric Theory Exam

3 Time Series Regression

Reading Assignment. Serial Correlation and Heteroskedasticity. Chapters 12 and 11. Kennedy: Chapter 8. AREC-ECON 535 Lec F1 1

Time Series Methods. Sanjaya Desilva

Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis. 18th Class 7/2/10

Week 11 Heteroskedasticity and Autocorrelation

BUSI 460 Suggested Answers to Selected Review and Discussion Questions Lesson 7

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics I. Professor William Greene Stern School of Business Department of Economics 25-1/25. Part 25: Time Series

Model Mis-specification

Read Section 1.1, Examples of time series, on pages 1-8. These example introduce the book; you are not tested on them.

Iris Wang.

Economics 308: Econometrics Professor Moody

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

LECTURE 10: MORE ON RANDOM PROCESSES

Making sense of Econometrics: Basics

in the time series. The relation between y and x is contemporaneous.

Decision 411: Class 3

Statistical Methods for Forecasting

Chapter 3: Regression Methods for Trends

Auto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e.,

Autocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time

Time Series and Forecasting

Decision 411: Class 3

Reading Assignment. Distributed Lag and Autoregressive Models. Chapter 17. Kennedy: Chapters 10 and 13. AREC-ECON 535 Lec G 1

Chapter 8: Model Diagnostics

Diagnostics of Linear Regression

DEPARTMENT OF QUANTITATIVE METHODS & INFORMATION SYSTEMS

MODELLING TIME SERIES WITH CONDITIONAL HETEROSCEDASTICITY

Decision 411: Class 3

White Noise Processes (Section 6.2)

7 Introduction to Time Series

Answer all questions from part I. Answer two question from part II.a, and one question from part II.b.

Autoregressive models with distributed lags (ADL)

Christopher Dougherty London School of Economics and Political Science

Basics: Definitions and Notation. Stationarity. A More Formal Definition

Econometrics. 9) Heteroscedasticity and autocorrelation

Regression Analysis By Example

A Hybrid Method of Forecasting in the Case of the Average Daily Number of Patients

7. Forecasting with ARIMA models

Linear Regression with Time Series Data

DEMAND ESTIMATION (PART III)

A Second Course in Statistics: Regression Analysis

Nonstationary Time Series:

Analysis. Components of a Time Series

WORKSHOP. Introductory Econometrics with EViews. Asst. Prof. Dr. Kemal Bağzıbağlı Department of Economic

Chapter 9: Forecasting

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

7 Introduction to Time Series Time Series vs. Cross-Sectional Data Detrending Time Series... 15

10. Time series regression and forecasting

Econ 510 B. Brown Spring 2014 Final Exam Answers

ECONOMETRICS HONOR S EXAM REVIEW SESSION

Multiple Regression Analysis

Econ 424 Time Series Concepts

1. Fundamental concepts

Serially Correlated Regression Disturbances

Rockefeller College University at Albany

Eco and Bus Forecasting Fall 2016 EXERCISE 2

Development of Short-term Demand Forecasting Model And its Application in Analysis of Resource Adequacy. For discussion purposes only Draft

Some Time-Series Models

2. An Introduction to Moving Average Models and ARMA Models

Lecture 7: Dynamic panel models 2

Reliability of inference (1 of 2 lectures)

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Estimation and Hypothesis Testing in LAV Regression with Autocorrelated Errors: Is Correction for Autocorrelation Helpful?

Applied Time Series Topics

ACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H.

Chapter 7 Forecasting Demand

Time-Series Analysis. Dr. Seetha Bandara Dept. of Economics MA_ECON

Modeling and Forecasting Currency in Circulation in Sri Lanka

Freeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE

Practicing forecasters seek techniques

Time Series and Forecasting

ECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 54

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables.

Transcription:

Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis 17th Class 7/1/10 The only function of economic forecasting is to make astrology look respectable. --John Kenneth Galbraith show the class examples of regressions from Wesleyan faculty members [handout] going through problems with regression; differentiating problems with data from problems with technique mention multicollinearity and omitted variables as the problems we last discussed a way of dealing with the problem of omitted variables if you have panel data (a timeseries on a cross-section): first-differencing --assume for now that the error terms are serially uncorrelated If the true model is: Y = b 0 + b 1 X + b 2 Z + e then this holds true at all points in time: Y t = b 0 + b 1 X t + b 2 Z t + e t Y t-1 = b 0 + b 1 X t-1 + b 2 Z t-1 + e t-1 if data are not available on Z, then we are forced to estimate (for either time): Y = b 0 + b 1 X and if X and Z are correlated, then the estimate of b 1 will be biased but can take first differences between the true regression equations: 17-1

(Y t - Y t-1 ) = b 0 - b 0 + b 1 (X t - X t-1 ) + b 2 (Z t - Z t-1 ) + (e t - e t-1 ) using the sign to signify a change and getting rid of zeros, this reduces to: Y = b 1 X + b 2 Z and if Z does not change over time (e.g., one s sex or race), then this reduces to running the regression: Y = b 1 X and the coefficient from this regression on X will be an unbiased estimate of b 1 note in order to have a constant term in this regression there must be a time trend T, which increments by one, and the coefficient on this term will be the constant in the new equation (e.g., years of experience, or age): Y t = b 0 + b 1 X t + b 2 Z t + b 3 t + e t Y t-1 = b 0 + b 1 X t-1 + b 2 Z t-1 + b 3 (t-1) + e t first differences yields: Y = b 1 X + b 3 so far we have justified OLS as a generally sound estimation technique, and have only discussed problems (multicollinearity, omitted variables) related to the data rather than the estimation technique but consider the basic assumptions underlying regression: E(e i ) = 0 error terms have mean 0 Var(e i ) = s 2 E(e i e j ) = 0 all error terms have the same variance error terms are not correlated across observations (they are independent) when these assumptions are met, OLS yields the BLUE (best linear unbiased estimator) as evaluated by minimized MSE (mean squared error) 17-2

but when one or more of these assumptions are violated, OLS needs to be adjusted or a different method used a check for patterns in the residuals will show violations problem of nonconstant variance is called heteroscedasticity; then OLS estimates of the coefficients are still nonbiased, but do not have minimum variance --then we use a form of computation of the variance/standard errors that is robust to heteroscedasticity (White standard errors) so as to calculate the correct significance tests for the coefficients problem of correlation across residuals is called autocorrelation (refers to the commonness of this phenomenon in time-series data); then alternative estimation techniques that exploit this correlation can be used to improve prediction (Ch. 24) Ch. 24 two major categories of statistical information: cross-section and time series (also panel data) consider special characteristics of time series data on top of their having a trend (i.e., a regression line with nonzero slope): --they tend to exhibit serial correlation: each observation is statistically dependent on the previous ones (so assumption of independence of observations does not hold) consider Figure 24-1: the long-term trend does not do a good job in predicting year to year changes; a better predictor would be to forecast next year s value as being the same as this year s, but we can do even better on average than that prediction --they often have seasonal variation too consider examples for quarterly, monthly, weekly, and daily data series --retail sales would have a spike in December that would affect quarterly and monthly data --stock data: prices drop on common x-dividend days --money demand: spikes on Fridays--common paydays, particularly at middle and end of months 17-3

the general technique of regression (not necessarily the particular estimation technique of OLS) can be used to decompose the time series into its different systematic and pure noise components and to forecast future values elements include: --trend (not necessarily linear; many economic time series have exponential trend, growing at a constant rate over time, so they are linear on a logarithmic scale) --seasonal fluctuation around the trend (can handle this often with a simple set of dummy variables) --serially correlated residual, e.g. e t = ρe t-1 + v t, where ρ represents the strength of the serial correlation (and is typically less than 1 in absolute value a correlation coefficient between residual and lagged residual); the general form of this is e t = ρ 1 e t-1 + ρ 2 e t-2 + + ρ k e t-k + v t, which would be called a kth order autoregression --white, or pure noise: v t in the above equation in general, when forecasts further into the future than next period are dependent upon previous periods residuals, they will need to be updated as soon as actual figures become available to improve their accuracy Durbin-Watson test for serial correlation: tests H 0 : ρ = 0 DW = 2-2r test values) so roughly DW = 0 means positive serial correlation, 2 is no correlation, and 4 is negative serial correlation (can look up critical values in Table IX for the exact What if we want to decompose a time series into 4 components: --trend T (same as above) --seasonal variation S (same as above) --cyclical variation C (modeled as part of the error term structure) 17-4

--disturbance D (white noise) pretty easy to get rid of T and S (use a time trend t in the regression, and dummies for the seasons), then just need to model C to get the error down to D What if we want to forecast a series that has serial correlation (assume no trend or seasonal variation for simplicity)? Several possible techniques can be used. Consider the autoregressive model which predicts Y t as a linear combination of previous values of Y: Y t = β 1 Y t-1 + β 2 Y t-2 + + β p Y t-p + u t so ˆ Y t = b 1 Y t-1 + b 2 Y t-2 + + b p Y t-p 1) can use the simple technique of exponential smoothing, which assumes that the autoregression extends back infinitely with smaller and smaller weights that decline exponentially, so: b k = bλ k where k = 1, 2, and 0 < λ < 1 substituting this into the autoregressive model, we get: ˆ Y t = bλy t-1 + bλ 2 Y t-2 + bλ 3 Y t-3 + since this can be applied to any time period, we can also write: ˆ Y t-1 = bλy t-2 + bλ 2 Y t-3 + bλ 3 Y t-4 + note we can simplify the model by writing it as: ˆ Y t = bλy t-1 + λ(bλy t-2 + bλ 2 Y t-3 + ) but now the term in parentheses is ˆ Y t-1, so we can write the model as: ˆ Y t = bλy t-1 + λ ˆ Y t-1 17-5

and if we require the weights to add to 1, then this must equal: Y ˆ t = (1 - λ)y t-1 + λ Y ˆ t-1 thereby allowing us to estimate what is an infinite series as a simple equation which can be estimated by least squares and used to forecast the current Y using a weighted average of the last Y and the prediction of the last Y that we had made last period 2) can use the Box-Jenkins model, which is a freer fit of the parameters than exponential smoothing, but does not require all the parameters of a completely free fit of a long autoregression; it cuts the autoregressive model off at a small number of lags, including only the most important recent lagged values. Therefore, it does not capture all of the serial correlation, some of which remains in u t. u t could therefore be modeled as an autoregressive structure as well, but is instead modeled as a moving average of serially independent terms v t-i (white noise terms): u t = γ 0 v t + γ 1 v t-1 + + γ q v t-q the coefficients are not required to sum to 1 or to be positive. It is generally restricted slightly however, and written as: u t = v t - α 1 v t-1 - - α q v t-q Upon substituting this into the autoregressive model, we obtain a combination of an autoregression (AR) of length p and a moving average (MA) of length q, or the ARMA (p,q) model: Y t = β 1 Y t-1 + β 2 Y t-2 + + β p Y t-p + v t - α 1 v t-1 - - α q v t-q the β and α parameters can be estimated using EViews and this model used for forecasting Y both of these techniques can be revised to incorporate seasonal and trend components now consider how we might not only forecast a time series, but also understand how it is related to one or more other time series 17-6

problem with serial correlation: successive observations tend to resemble previous observations (are dependent) and hence provide less new information about trend than a set of independent observations would. Therefore, estimates are less reliable, which should be reflected in wider confidence intervals around coefficients. Consider our basic regression model: Y t = α + βx t + e t where X and Y are two time series assume that: e t = e t-1 + v t so for simplicity we are assuming that residuals are perfectly correlated; i.e., ρ = 1 then b will not be biased (it could either overestimate or underestimate β, depending on the particular pattern of the residuals), but will have a large standard error [see Figure 24-10]. Note that the true line will actually look less reliable than the OLS estimate because the observed/estimated residuals will be smaller with OLS (autocorrelated data is smooth) than the true residuals remedy for serial correlation: transform the data to a form that conforms to the assumptions underlying OLS, namely estimate the equation in changes rather than levels. Our model is: Y t = α + βx t + e t and therefore: Y t-1 = α + βx t-1 + e t-1 subtracting the second equation from the first, we get: 17-7

Y t - Y t-1 = β(x t - X t-1 ) + (e t - e t-1 ) define the differences: Y t = Y t - Y t-1 X t = X t - X t-1 then this can be written as: Y t = β X t + v t so b may be estimated validly using OLS on this equation let us now generalize this model to allow ρ < 1, so: e t = ρe t-1 + v t we can lag the model and multiply through by r: ρy t-1 = ρα + ρβx t-1 + ρe t-1 then, upon first-differencing, the model is: Y t - ρy t-1 = α(1 - ρ) + β(x t - ρx t-1 ) + (e t - ρe t-1 ) define the generalized differences: Y t = Y t - ρy t-1 X t = X t - ρx t-1 then the model can be written: Y t = α(1 - ρ) + β X t + v t so after this transformation, regress Y t on X t to estimate β. This technique is an example of generalized least squares (GLS). 17-8

Note the problem here: we don t know what ρ is. We have to estimate it somehow: 1) We can estimate it by first using OLS to fit a line to the data and generate estimates of the residuals, and then using OLS to estimate the equation: e ˆ t = r e ˆ t-1 however, this technique causes r to be a downward-biased estimate of r: the estimated residuals are less serially correlated than are the true residuals (as can be seen in Figure 24-10) 2) alternatively, we can estimate ρ by first rearranging our model: Y t - ρy t-1 = α(1 - ρ) + β(x t - ρx t-1 ) + (e t - ρe t-1 ) into the following form: Y t = α(1 - ρ) + ρy t-1 + βx t - βρx t-1 + v t which can then be estimated using OLS. Then the coefficient of Y t-1 provides us with an estimate of ρ. 3) the same as (2), except estimate ρ by dividing the coefficient of X t-1 by the coefficient of X t (and change the sign) 4) the same as (2) and (3), except take an average of the two estimates of ρ to get an even better estimate of r. Note that if the two estimates are very different, this is a warning that the model may be poorly specified. [consider problem 24-11 if there is time] 17-9