Okun's Law Testing Using Modern Statistical Data. Ekaterina Kabanova, Ilona V. Tregub

Similar documents
Making sense of Econometrics: Basics

AUTOCORRELATION. Phung Thanh Binh

Econometrics. 4) Statistical inference

11.1 Gujarati(2003): Chapter 12

Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation

Outline. 11. Time Series Analysis. Basic Regression. Differences between Time Series and Cross Section

Econometrics. 9) Heteroscedasticity and autocorrelation

LECTURE 11. Introduction to Econometrics. Autocorrelation

F9 F10: Autocorrelation

The Multiple Regression Model

Types of economic data

Multiple Regression Analysis

ECON 4230 Intermediate Econometric Theory Exam

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Econometrics. 7) Endogeneity

CHAPTER 5 FUNCTIONAL FORMS OF REGRESSION MODELS

Diagnostics of Linear Regression

Iris Wang.

Regression Models - Introduction

2 Prediction and Analysis of Variance

Section 2 NABE ASTEF 65

Chapter 14 Student Lecture Notes Department of Quantitative Methods & Information Systems. Business Statistics. Chapter 14 Multiple Regression

Microeconometria Day # 5 L. Cembalo. Regressione con due variabili e ipotesi dell OLS

DEMAND ESTIMATION (PART III)

Regression of Time Series

Autoregressive models with distributed lags (ADL)

Empirical Economic Research, Part II

Practice Questions for the Final Exam. Theoretical Part

ENERGY CONSUMPTION AND ECONOMIC GROWTH IN SWEDEN: A LEVERAGED BOOTSTRAP APPROACH, ( ) HATEMI-J, Abdulnasser * IRANDOUST, Manuchehr

Statistics for Managers using Microsoft Excel 6 th Edition

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

The Simple Linear Regression Model

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

Regression Analysis. y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T,

Testing methodology. It often the case that we try to determine the form of the model on the basis of data

Correlation Analysis

Answer all questions from part I. Answer two question from part II.a, and one question from part II.b.

Autocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time

EC312: Advanced Econometrics Problem Set 3 Solutions in Stata

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE

Inferences for Regression

Macroeconometrics. Christophe BOUCHER. Session 4 Classical linear regression model assumptions and diagnostics

Ref.: Spring SOS3003 Applied data analysis for social science Lecture note

Stat 500 Midterm 2 12 November 2009 page 0 of 11

Multiple Regression. Peerapat Wongchaiwat, Ph.D.

Econometrics I. Professor William Greene Stern School of Business Department of Economics 25-1/25. Part 25: Time Series

Ch 2: Simple Linear Regression

LINEAR REGRESSION ANALYSIS. MODULE XVI Lecture Exercises

Assumptions of classical multiple regression model

Lecture 4: Heteroskedasticity

Föreläsning /31

Chapter 5. Classical linear regression model assumptions and diagnostics. Introductory Econometrics for Finance c Chris Brooks

1/34 3/ Omission of a relevant variable(s) Y i = α 1 + α 2 X 1i + α 3 X 2i + u 2i

Heteroskedasticity and Autocorrelation

Hypothesis Testing hypothesis testing approach

Are Forecast Updates Progressive?

Økonomisk Kandidateksamen 2004 (I) Econometrics 2

Are 'unbiased' forecasts really unbiased? Another look at the Fed forecasts 1

Agricultural and Applied Economics 637 Applied Econometrics II

Chapter 8 Heteroskedasticity

Outline. Possible Reasons. Nature of Heteroscedasticity. Basic Econometrics in Transportation. Heteroscedasticity

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

The multiple regression model; Indicator variables as regressors

Analisi Statistica per le Imprese

LATVIAN GDP: TIME SERIES FORECASTING USING VECTOR AUTO REGRESSION

Are Forecast Updates Progressive?

2.1. Consider the following production function, known in the literature as the transcendental production function (TPF).

Økonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning

We like to capture and represent the relationship between a set of possible causes and their response, by using a statistical predictive model.

Modelling the Electric Power Consumption in Germany

Introduction to Econometrics

Review of Statistics

Identifying the Monetary Policy Shock Christiano et al. (1999)

Two-Variable Regression Model: The Problem of Estimation

Econometrics Part Three

Heteroskedasticity. Part VII. Heteroskedasticity

CHAPTER 6: SPECIFICATION VARIABLES

ECON 497: Lecture Notes 10 Page 1 of 1

Hypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima

Lab 07 Introduction to Econometrics

A discussion on multiple regression models

1 Quantitative Techniques in Practice

The OLS Estimation of a basic gravity model. Dr. Selim Raihan Executive Director, SANEM Professor, Department of Economics, University of Dhaka

The Dark Corners of the Labor Market

Reliability of inference (1 of 2 lectures)

Heteroscedasticity. Jamie Monogan. Intermediate Political Methodology. University of Georgia. Jamie Monogan (UGA) Heteroscedasticity POLS / 11

SMA 6304 / MIT / MIT Manufacturing Systems. Lecture 10: Data and Regression Analysis. Lecturer: Prof. Duane S. Boning

REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK

Inflation Targeting as a Tool To Control Unemployment

The One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

Econometrics Homework 1

Econometrics. 5) Dummy variables

interval forecasting

Lecture 10 Multiple Linear Regression

Inference for Regression Simple Linear Regression

Applied Econometrics (QEM)

OSU Economics 444: Elementary Econometrics. Ch.10 Heteroskedasticity

2 Regression Analysis

Chapter 16. Simple Linear Regression and Correlation

Transcription:

Okun's Law Testing Using Modern Statistical Data Ekaterina Kabanova, Ilona V. Tregub The Finance University under the Government of the Russian Federation International Finance Faculty, Moscow, Russia E-mail: ilonavl_fa@mail.ru, Fax: (495) 366 56 33 Description of the model Okun's law is an empirically observed relationship relating unemployment to losses in a country's production first quantified by Arthur M. Okun. The "gap version" states that for every 1% increase in the unemployment rate, a country's GDP will be at an additional roughly 2% lower than its potential GDP. The difference version describes the relationship between quarterly changes in unemployment and quarterly changes in real GDP. The accuracy of the law has been disputed. Okun's law is more accurately called "Okun's rule of thumb" because it is primarily an empirical observation rather than a result derived from theory. Okun's law is approximate because factors other than employment (such as productivity) affect output. In Okun's original statement of his law, a 3% increase in output corresponds to a 1% decline in the rate of unemployment; a 0.5% increase in labor force participation; a 0.5% increase in hours worked per employee; and a 1 % increase in output per hours worked (labor productivity). Okun's law states that a one point increase in the unemployment rate is associated with two percentage points of negative growth in real GDP. The relationship varies depending on the country and time period under consideration. The relationship has been tested by regressing GDP or GNP growth on change in the unemployment rate. Martin Prachowny estimated about a 3% decrease in output for every 1% increase in the unemployment rate (Prachowny 1993). The magnitude of the decrease seems to be declining over time in the United States. According to Andrew Abel and Ben Bemanke, estimates based on data from more recent years give about a 2% decrease in output for every 1% increase in unemployment (Abel and Bernanke, 2005). There are several reasons why GDP may increase or decrease more rapidly than unemployment decreases or increases. As unemployment increases: a reduction in the multiplier effect created by the circulation of money from employees unemployed persons may drop out of the labor force (stop seeking work), after which they are no longer counted in unemployment statistics employed workers may work shorter hours labor productivity may decrease, perhaps because employers retain more workers than they need

One implication of Okun's law is that an increase in labor productivity or an increase in the size of the labor force can mean that real net output grows without net unemployment rates falling (the phenomenon of "jobless growth"). Arthur Okun estimated the following relationship between the two: Y t = - 0.4 (X t - 2.5) This can also be expressed as a more traditional linear regression as: Y t = 1-0.4 X t Where Y t is the change in the unemployment rate in percentage points. X t is the percentage growth rate in real output, as measured by real GNP. So we will be estimating the model: Y t = b 1 + b 2 X t Where Y t is the change in the unemployment rate in percentage points. X t is the change in the percentage growth rate in real output, as measured by real GNP. b 1 and b 2 are the parameters we are trying to estimate. Required data for the estimation In order to analyze and test the Okun s law, we should find out some specific data: a percentage change in GNP relative to previous quarter, which we will denote y t, and change in the unemployment rate from last quarter, which we will denote as x t. We will take quarterly data from the 01.04.1948 to 01.07.2002. Also it s important to mention, that the last interval of the 01.10.2002 shouldn t be involved in our analyzing right now, we will need it later for model forecasting. For the evaluation of this model we are going to use data that was collected by U.S. government agencies and thus is in the public domain (https://fedstats.gov) The steps of an econometric model testing. Model Specification Here is a mathematical interpretation of the Okun s law. General form of the fitted line: Y i = β 0 + β 1 X i + U i, E u(t) = 0, σ u(t) = const

Where Y t is the change in the unemployment rate in percentage points., X 1t is the percentage growth rate in real output, as measured by real GNP, β 0,1 parameters (sensitivity of the explained variable to changes of the explainable variable), U i the disturbance term. Now let s turn to the precise explanation of all the steps in the Excel. Let us input the values of endogenous and exogenous variables from 1947-01-01 to 2002-07-01 into corresponding rows in the Regression, Analysis ToolPak. We did not take data of 2002 year, because We are going to use it later when checking model adequacy. Level of significance is 95%. Results of the regression statistics: Table 1 Regression Statistics Multiple R 0,711246 R Square 0,505870 Adjusted R Square 0,503583 Standard Error 0,725891 Observations 218 The results of variance analysis are generated in Table 3, which are used to the coefficient of determination R 2. Table 2 df SS MS F Significance F Regression 1 116,5189 116,5189 221,1327 0,00 Residual 216 113,8143 0,5269 Total 217 230,3333 Table 3 contains values of regression coefficients with a i (95% - in our case) confidence probability and their statistical assessment. Table 3 Coefficients Standard Error t Stat P- value Y interception 0,85 0,049 17,43 0,01 X 1-1,82 0,122 14,87 0,01 Lower 95% Upper 95% 0,76 0,95-2,06-1,58 Table 4 contains theoretical values ˆi y, computed with the help of regression formula, and residual values.

Residual values are calculated as the difference between empirical y i and theoretical y ˆi. Year of Predicted Y Residuals observation 1948-04-01 0,979 0,605 2002-07-01 0,979 0,100 Specification of estimated econometric model The specification of our model with calculated parameters: Specification of Estimated model Table 4 Y t = 0,86-1,82 X t + u t (0,049) (0,12) (0,72) F = 221,13; F crit = 3, 88; R 2 = 0,506; t crit = 3,336 Where β 0 = 0,857 with standard error of 0,04917, β 1 = (-1,826) with standard error of 0,12282, the standard error of disturbance term is 0,72589. Model testing Now let s move to another important step model testing. The calculated regression coefficients β i allow us to construct an equation of Okun s law. The equation is: Y t = 0,857-1,826Xt + ε t, where ε t is random value. R 2 Coefficient Value of multiple coefficient of determination R 2 equals to 0.506 shows that 50.6% of total deviation of y i is explained by the variation of the factors x i. Honestly speaking, such value of the R 2 is not good, as it s not so near to 1. It means that selected factors don t effect significantly to our model, which confirms the correctness of their inclusion in the estimated model. Significance F The calculated level of significance 0.0000001<0.05 (Significance F, table 2) confirms the R 2 significance.

F-test Another way of checking R 2 is based on testifying whether F (table 2) is within the interval (F crit ;+ ) or not. In our case F crit =Fраспобр(0.05;2;216)= 3,038, where 2 is the number of degrees of freedom, it equals to the number of the equation regresses m=2, and 216 is the number of degrees of freedom, it equals to n-(m+1). As our F=221,1327 and it is within the interval (3,038;+ ), the H 0 hypothesis that R 2 =0 is rejected. That means coefficient of determination R 2 is thought to be significant. Standard Error Now we are going to test the importance of regression coefficients β i. Comparing the elements of the columns Coefficients and Standard Error (Table 3), we can notice that absolute values of standard errors is less than the corresponding values of coefficients, therefore, at the first stage of analysis, all the variables should remain in the model. t-test Let us check the significance of these coefficients with the help of t-test. That is to test the inequality t tcrit, where t is the value of t-statistics (t Stat, table 3). If the inequality is right, the coefficient and its factor variable are considered to be non-significant; in the other case the coefficient and the regressor are considered to be significant. In our case the critical value can be calculated using the special tables. For the considered example, value t crit =1,971, where 0.05 is the level of significance, 216 number of observations, 1 number of factors in the regression equation, 1 number of free terms in the regression equation. All absolute values of t-statistics in table 3 are more than t crit, therefore, all the regression coefficients are significant. P-value test Another common way of regression coefficients significance testing is based on P-value indicator implementation (P-value, table 3). All p-values are less than our level of significance (α=0.05). Thus, all coefficients are being significant. After all above-listed tests, we can conclude about a good model specification and the significance of the coefficients. Goldfield-Quandt test Goldfield-Quandt test is designed to check the second assumption of Gauss- Markov theorem about homoscedasticity of random disturbances, i.e. about the following equality satisfying: 2 Var( ε1) = Var( ε2) =... = Var( εn ) = σ To implement it in the situation of the model of a linear multiple regression, it is necessary to: 1. Sort the initial data in ascending of the regressors sums of absolute values ( x1 t, x2t,..., x mt) ;

2. Split ordered data into two arrays so as (m+1)<k<n/2, where k number of observations in the first array, m- number of regressors, m+1 number of explainable regression function coefficients; 3. Assess a regression equation for each array separately, using Regression, Analysis ToolPak. This results in two models with the same equations, but with different coefficients; 4. Calculate Goldfield-Quandt GQ statistics as the ratio of the sum of squares of deviations of empirical data to the theoretical one for the first array (RSS 1 ) to the corresponding amount, calculated for the second array (RSS 2 ), GQ = RSS1 RSS2. RSS can be found in ANOVA table, SS column, Residual raw; 5. Evaluate the critical value F crit for given significance level α with ν 1 and ν 2 degrees of freedom, where ν 1 = ν 2 =k-(m+1), k number of observations in the first array, m number of regressors. Fраспобр(α;ν 1 ;ν 2 ); 6. Second assumption about homoscedasticity of random disturbances is thought to be adequate if both of the following inequalities GQ Fcrit are valid: 1/ GQ Fcrit Otherwise, conclude about heteroscedasticity of random disturbances. This leads to loss of unbiased property of estimation of the parameters of the linear regression model obtained by the method of least squares, and the accuracy inadequacy of the characteristics of these estimations. Let us consider our case. We should split ordered by the sum of x it data into two arrays. The first one contains 109 observations, while the second = 108. Using Regression, Analysis ToolPak we can see the values of RSS 1,2 and, hence, estimate GQ, 1/GQ, F crit GQ (table 5). F crit GQ is calculated by function Fраспобр(α;ν 1 ;ν 2 ), where α=0.05, ν 1 =ν 2 =k-(m+1)=107, where k number of observations in the first array, m number of factors. Thus, both inequalities are valid. Assumption about homoscedasticity of random disturbances is adequate. Goldfield-Quandt test Table 5 RSS 1 55,70 RSS 2 54,70 GQ 1,018 1/GQ 0,981 F crit GQ 1,828 Durbin-Watson test This test is designed to check a particular case of third assumption of the Gauss-Markov theorem about the absence of autocorrelation between adjacent random residuals in the model. Cov( ε, ε ) = 0 if j = i 1 i j

Using values of the residuals ε t, we can compute Durbin-Watson statistics: n n 2 2 ( t t 1) t t= 2 t= 1 DW = ε ε ε Then, we should find Durbin-Watson statistics critical values d L and d U with the help of special statistical table, where n=217 total number of observations, k=1 total number of factors. There are three possible outcomes of the test: d {(0; d );(4 d ;4)} L L => positive/negative autocorrelation of the model s residuals exists; d { ( dl; du);(4 du;4 dl) } => autocorrelation of the model s residuals is ambiguous; d ( du;4 du) => positive/negative autocorrelation of the model s residuals does not exist. Confidence interval We should estimate the lower and upper boundaries for each year. We will use the following formula: 99,9% (I had to decrease the level of significance as with the significance level of 95% not all of our data laid in our intervals) boundary= Y ˆ i+ tcriti* st. error, where t crit is calculated as it has been shown in part Model testing, section t-test and standard error = 0,725891 (Standard error of new Specification, table 1), Y ˆ - predicted value of y t (Predicted Y according to new specification). Then we should compare the empirical data for each year with resulted interval boundaries (Appendix, Table 8). Adequacy checking Let us check whether predicted, by our model, y ˆi is truly describes the empirical data correctly and, consequently, test the forecasting capabilities of our model, it does the empirical data about the real GNP (for American economy) lie within confidence interval, predicted by our model. y 1 {( ˆ. * );( ˆ n y n st error tcrit y + + n st. error * tcrit) }. Let us look at our case: Table 6 Lower Upper Empirical Empirical>Lower Empirical<Higher 95% 95% 95% 95% -1,442 3,400 1,584 True True So, our empirical for the 1.04 of 1948 data lies between upper and lower boundaries predicted by our model. We can forecast the future correctly and accurately.

PREDICTING Judging by the estimated coefficients in the model, the US real GNP is inversely proportional to its level of unemployment. The previous tests showed that Okun s law has not been useful as a stable relationship, since its parameters have varied considerably over time and over the course of the business cycle. In addition, it has not always been a reliably strong relationship, especially in quarterly data. CONCLUSION The previous tests showed that Okun s law has not always been a reliably strong relationship, especially in quarterly data. Nevertheless, the relationship between contemporaneous changes in unemployment and output growth may still be useful to policymakers and economists if they take these shortcomings into consideration. In this way our model is of a good explanatory ability and can be used for general data analysis of consumption functions.