Freeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94

Size: px
Start display at page:

Download "Freeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94"

Transcription

1 Freeing up the Classical Assumptions () Introductory Econometrics: Topic 5 1 / 94

2 The Multiple Regression Model: Freeing Up the Classical Assumptions Some or all of classical assumptions needed for derivations of Topic 3 and Topic 4 slides. Derivation of the OLS estimator only required the assumption of a linear relationship between Y and X But to show that OLS estimator had desirable properties did require assumptions Proof of the Gauss-Markov theorem required all classical assumptions except for the assumption of Normal errors Derivation of confidence intervals and hypothesis testing procedures required all classical assumptions. But what if some or all of the classical assumptions are false? This set of slides (based on Chapter 5 of textbook) will show how to relax them () Introductory Econometrics: Topic 5 2 / 94

3 Begin by discussing some general theory, before considering some special cases. Two general categories: problems which call for use of Generalized Least Squares (or GLS) Estimator Heteroskedasticity and autocorrelated errors will be discussed in this category. Second category relates to the use of the so-called Instrumental Variables (IV) Estimator. () Introductory Econometrics: Topic 5 3 / 94

4 Basic Theoretical Results Previously derived theoretical results using multiple regression model with classical assumptions: Y i = α + β 1 X 1i β k X ki + ε i 1 E (ε i ) = 0 mean zero errors. 2 var (ε i ) = E ( ε 2 ) i = σ 2 constant variance errors (homoskedasticity). 3 cov (ε i ε j ) = 0 for i = j. 4 ε i is Normally distributed 5 Explanatory variables are fixed. They are not random variables. () Introductory Econometrics: Topic 5 4 / 94

5 Remember: Assumption 1 is innocuous If error had non-zero mean could include it as part of the intercept it would have no effect on estimation of slope coeffi cients in the model). Assumption 4 can be relaxed (approximately) by using asymptotic theory (not discussed in this course, but see Appendix to Chapter 3 if you are interested) Assumption 5 we will still maintain (will discuss this later on in the context of "instrumental variables" estimation). For now focus on Assumptions 2 and 3. () Introductory Econometrics: Topic 5 5 / 94

6 Heteroskedasticity relates to Assumption 2. Autocorrelation (also called serial correlation) relates to Assumption 3 Basic ideas: Under classical assumptions, Gauss Markov theorem says "OLS is BLUE". But if Assumptions 2 and 3 are violated OLS this no longer holds (OLS is still unbiased, but is no longer "best". i.e. no longer minimum variance). Concepts/proofs/derivations often use following strategy. The model can be transformed to create a new model which does satisfy classical assumptions. We know OLS (on the transformed model) will be BLUE. (And all the theory we worked out for the OLS estimator will hold except it will hold for the transformed model). The OLS estimator using such a transformed model is called the Generalized Least Squares (GLS) estimator. () Introductory Econometrics: Topic 5 6 / 94

7 Heteroskedasticity Heteroskedasticity occurs when the error variance differs across observations. Errors reflects mis-pricing of houses If a house sells for much more (or less) than comparable houses, then it will have a big error Error variance measures the dispersion of errors/degree of possible mis-pricing houses Suppose small houses all very similar, then unlikely to be large pricing errors Suppose big houses can be very different than one another, then possible to have large pricing errors If this story is true, then expect errors for big houses to have larger variance than for small houses Statistically: heteroskedasticity is present and is associated with the size of the house () Introductory Econometrics: Topic 5 7 / 94

8 Assumption 2 replaced by var (ε i ) = σ 2 ω 2 i for i = 1,.., N. Note: ω i varies with i so error variance differs across observations Begin with some theoretical results assuming ω 2 i is known What are the properties of the OLS estimator if heteroskedasticity is present? () Introductory Econometrics: Topic 5 8 / 94

9 To make derivations easier, go back to simple regression model: Y i = βx i + ε i Classical assumptions hold, except for Assumption 2. We now have heteroskedasticity. Remember that OLS estimator can be written in various ways: β = X i Y i Xi 2 = β + X i ε i Xi 2 Under classical assumptions, we proved: ( ) σ β 2 is N β, Xi 2 We used this to derive confidence intervals and hypothesis testing procedures. () Introductory Econometrics: Topic 5 9 / 94

10 Under heteroskedasticity, most of our previous derivations still work. The error variance did not appear in our proofs for unbiasedness of OLS nor showing it had a Normal distribution. Hence, we will not repeat the derivations here but simply state the following results: Under the present assumptions ( β) (i.e. allowing for heteroskedasticity), OLS is still unbiased (i.e. E = β) and it is Normally distribution. New result: Under the present assumptions, ) var ( β = σ2 Xi 2 ω 2 i ( Xi 2 ) 2 () Introductory Econometrics: Topic 5 10 / 94

11 Proof (using properties of variance operator) ) var ( β ( = var β + ) X i ε i Xi 2 ( ) = var X i ε i Xi 2 1 = ( ) Xi 2 2 var ( ) X i ε i 1 = ( ) Xi 2 2 Xi 2 var (ε i ) = σ 2 ( X 2 i ) 2 X 2 i ω 2 i () Introductory Econometrics: Topic 5 11 / 94

12 Key Theoretical Point: If heteroskedasticity is present, variance of the OLS estimator is different than what it was under the classical assumptions. ) But var ( β enters the formula for confidence intervals and test statistics Key Points for Empirical Practice: If heteroskedasticity is present and you ignore it, simply using the OLS estimator in a software package, ) the software package will use the incorrect formula for var ( β. Software package will use formula which obtains ) under classical assumptions, where it should be using var ( β = σ2 Xi 2 ω 2 i. ( Xi 2 ) 2 Confidence intervals and test statistics will be incorrect () Introductory Econometrics: Topic 5 12 / 94

13 Summary: OLS is still unbiased if heteroskedasticity is present (so as an estimate it may be okay) But almost everything else (confidence intervals, hypothesis tests, etc.) will be incorrect. The only case where using OLS is acceptable ) is if you make sure the computer is using the correct var ( β = σ2 X 2 formula. i ω 2 i ( Xi 2 ) 2 We will return to this point later in our discussion of the heteroskedasticity consistent estimator (to be defined later). () Introductory Econometrics: Topic 5 13 / 94

14 The Generalized Least Squares Estimator under Heteroskedasticity Idea: Transform the model to create new model which does obey classical assumptions. The original regression model is: Y i = βx i + ε i (1) Consider a transformed model where we divide both sides by ω i : Y i ω i = β X i ω i + ε i ω i or (to make the notation compact): Y i = βx i + ε i (2) Can prove the transformed model given in (2) satisfies the classical assumptions. () Introductory Econometrics: Topic 5 14 / 94

15 Since ε i has mean zero, so will εi Key thing to verify is that var (εi ) is constant var (ε i ) = var ( εi ω i ) = 1 ω 2 i var (ε i ) = σ2 ω 2 i ω 2 i = σ 2 So error variances in (2) are constant. Important point: The transformed model in (2) satisfies classical assumptions. Hence, all our OLS results (using transformed model) can be used to say OLS (on transformed model) is BLUE, OLS confidence intervals (using transformed data) are correct, etc. etc. Don t have to re-do all the proofs on Topic 3 (just plug in Yi, Xi, εi instead of the Y i, X i, ε i ) () Introductory Econometrics: Topic 5 15 / 94

16 The Generalized Least Squares Estimator The previous reasoning suggests OLS using transformed model provides a good estimator: β GLS = Xi Yi Xi 2 In terms of the original data this is: β X i Y i ω GLS = 2 i X i 2 ω 2 i This is called the Generalized Least Squares (GLS) estimator I have written "GLS" as a subscript on it to make explicit it is not the same as OLS on original model. Note: I am still working with the simple regression model, but the extension to multiple regression is immediate. Simply divide every explanatory variable (and the dependent variable) by ω i and then do OLS on the transformed model. () Introductory Econometrics: Topic 5 16 / 94

17 Properties of the GLS estimator (under heteroskedasticity) Since GLS is equivalent to OLS on transformed model, can use OLS results from Topic 3 slides (and apply them to the transformed model). That is, plug in Xi and Yi instead of X i and Y i in all our old formulae. Since the transformed model satisfies the classical assumptions, draw on old results to say: ( ) σ β 2 GLS is N β, Xi 2 Thus, (under heteroskedasticity) GLS is unbiased with ) σ var ( β GLS = 2 Xi 2 = σ 2 ( Xi 2 ω 2 i () Introductory Econometrics: Topic 5 17 / 94 )

18 Gauss-Markov theorem tells us that, under the classical assumptions, OLS is BLUE. Here β GLS is equivalent to OLS estimation of a transformed model which does satisfy the classical assumptions. Hence, under heteroskedasticity, it follows immediately that β GLS is BLUE. An implication of this is that: ) ) var ( β GLS var ( β OLS Thus, GLS is a better estimator than OLS. Both are unbiased, but GLS has a smaller variance (it is more effi cient). () Introductory Econometrics: Topic 5 18 / 94

19 The fact that β GLS is N ( β, ) σ 2 Xi 2 can be used to derive confidence intervals and hypothesis tests exactly as before. We will not repeat this material. Formulae are same as before except with X y i i and y i instead of X i and () Introductory Econometrics: Topic 5 19 / 94

20 Heteroskedasticity: Estimation if Error variances are unknown Derivations above assumed that ω 2 i is known. In practice, it will usually be the case that ω 2 i is unknown. How to proceed? Either figure out what ω 2 i is or replace ω 2 i by an estimate. Or a heteroskedasticity consistent estimator (HCE) can be used. Digression: consistency is an asymptotic concept (asymptotic derivations not done in this course) Intuition 1: Consistency has some similarities to unbiasedness. Intuition 2: A consistent estimator is one which, as sample size goes to infinity, goes to true value. () Introductory Econometrics: Topic 5 20 / 94

21 Fixing up a Heteroskedasticity Problem by Logging In some cases, log linear regressions will be homoskedastic even if linear regression is heteroskedastic Note: if variables have values which are zero or negative you cannot log them. But even if you log some of your variables (or even only log the dependent variable) it is sometimes enough to fix up a heteroskedasticity problem Remember: be careful with interpretation of coeffi cients when you log variables (see Topic 4 slides) Heteroskedasticity tests (see below) can be used to see whether logging fixes us a heteroskedasticity problem Note: solving a heteroskedasticity problem by logging is not called GLS () Introductory Econometrics: Topic 5 21 / 94

22 Doing GLS by Transforming the Model In many cases, the heteroskedasticity can be related to an explanatory variable. Consider multiple regression model: Y i = α + β 1 X 1i β k X ki + ε i under the classical assumptions except that var (ε i ) = σ 2 ω i = σ 2 Z 2 i Z i is an explanatory variable (usually Z i will be one of X 2i,.., X ki ). This captures the idea "the error variances vary directly with an explanatory variable". () Introductory Econometrics: Topic 5 22 / 94

23 If you suspect "the error variances vary inversely with an explanatory variable" you could use: var (ε i ) = σ 2 1 Z 2 i Note: variances must be positive which is why I have used Z 2 i. An alternative choice is to use the exponential function (e.g. var (ε i ) = σ 2 exp (Z i )). () Introductory Econometrics: Topic 5 23 / 94

24 Remember: under heteroskedasticity, GLS says we should transform our data as: Y i = β X i + ε i ω i ω i ω i and then use OLS on transformed model. But here we have ω i = Z i. So can divide all your variables by Z i and then do OLS. Empirical tip: Experiment with different choices for Z (usually, it will be one of X 1,.., X k ) Note: cannot divide by zero and, hence, you cannot use this transformation for a variable which has Z i = 0 for any observation Cannot use of this transformation with dummy variables. If the heteroskedasticity is characterized by var (ε i ) = σ 2 exp (Z i ) then zero values of Z i are acceptable. () Introductory Econometrics: Topic 5 24 / 94

25 Previous slide has "error variances vary directly with Z" If error variances vary inversely with Z, transformed model becomes: Y i Z i = αz i + β 1 X 1i Z i β k X ki Z i + ε i Z i GLS estimator obtained by multiplying all your variables by Z i and then doing OLS with these new variables. Interpretation: from original regression β j is marginal effect of X j on Y (ceteris paribus) and this interpretation holds with transformed model But now β j appears as coeffi cient on X j Z (or X j Z ) () Introductory Econometrics: Topic 5 25 / 94

26 What if heteroskedasticity is present, but you cannot relate it to a single variable, Z? It is desirable to do GLS if you can. That is, if you can find a transformed model which does not suffer from heteroskedasticity (must check with heteroskedasticity tests) If you cannot, remember that OLS is still unbiased so is an adequate second best estimator. But the variance formula we derived under the classical assumptions not longer holds. The correct formula is: ) var ( β = σ2 Xi 2 ω 2 i ( Xi 2 ) 2 So one thing you can do is use OLS with this correct formula to calculate the variance. () Introductory Econometrics: Topic 5 26 / 94

27 Problem: we do not know σ 2 ω 2 i. Solution: Replace it with an estimate. Since can use the OLS residuals: as estimates of σ 2 ω 2 i. ) Thus, an estimate of var ( β is var (ε i ) = E ( ε 2 ) i = σ 2 ω 2 i ε 2 i ) var ( β = X i 2 ε2 i ( Xi 2 ) 2 () Introductory Econometrics: Topic 5 27 / 94

28 Summary: Use OLS to estimate β, then use var ) ( β in formulae for confidence intervals, etc. This is an example of a heteroskedasticity consistent estimator (HCE). They can be automatically calculated in Gretl Advantages: HCEs are easy to calculate and you do not need to know the form that the heteroskedasticity takes. Disadvantages: HCEs are not as effi cient as the GLS estimator (i.e. they will have larger variance). () Introductory Econometrics: Topic 5 28 / 94

29 Testing for Heteroskedasticity If heteroskedasticity is NOT present, then OLS is fine (it is BLUE). But if it is present, you should use GLS (or a HCE). Thus, it is important to know if heteroskedasticity is present. There are many tests, here I will describe two of the most popular ones: the White test and the Breusch Pagan test () Introductory Econometrics: Topic 5 29 / 94

30 Suppose you think the error variance might depend on any or all of the variables Z 1,.., Z p (which may be the same as the explanatory variables in the regression itself). Breusch-Pagan test involves the following steps: Run OLS on the original regression (ignoring heteroskedasticity) and obtain the residuals, ε i, and using the residuals calculate: σ 2 = ε2 i N Run a second regression of the equation: ε 2 i σ 2 = γ 0 + γ 1 Z 1i γ p Z pi + v i Calculate the Breusch-Pagan test statistic using the regression sum of squares (RSS) from this second regression: BP = RSS 2 This test statistic has a χ 2 (p) distribution, which can be used to get a critical value from. () Introductory Econometrics: Topic 5 30 / 94

31 The White Test for Heteroskedasticity Similar to the Breusch-Pagan test White test involves the following steps: Run OLS on the original regression (ignoring heteroskedasticity) and obtain the residuals, ε i. Run a second regression of the equation: ε 2 i = γ 0 + γ 1 Z 1i γ p Z pi + v i and obtain the R 2 from this regression. Calculate the White test statistic: W = NR 2 This test statistic has a χ 2 (p) distribution which can be used to get a critical value from. () Introductory Econometrics: Topic 5 31 / 94

32 An advantage of the White and Breusch Pagan tests is that you just need to choose Z 1,.., Z p Usually these are just the explanatory variables in the original regression, although with the White test researchers sometimes also include squares and cross-products of the explanatory variables. A disadvantage is that, if the tests indicate that heteroskedasticity is present, they do not offer much guidance on how you should try and transform the model to do GLS. All you know is that heteroskedasticity is present and is related to one (or several) of the variables Z 1,.., Z p. () Introductory Econometrics: Topic 5 32 / 94

33 Recommendations for Empirical Practice If you think you might have a heteroskedasticity problem, begin by doing a heteroskedasticity test. If your tests indicate heteroskedasticity is present, then do some experimentation to see if you can solve the heteroskedasticity problem Sometimes logging your dependent variable (and some or all of the explanatory variables) will be enough to fix the problem. Sometimes multiply/dividing all your explanatory variables by some variable (Z j ) is enough to fix the problem Try different choices for Z j Note: Every time you try such a transformation you must do a heteroskedasticity test to check if it has fixed the problem. () Introductory Econometrics: Topic 5 33 / 94

34 If you cannot find a transformation which fixes the heteroskedasticity problem, then use a HCE. Remember: if heteroskedasticity is present, then hypothesis tests involving β s will be incorrect. So wait until after you have corrected the problem (or are using an HCE) before doing hypothesis testing (e.g. to find out which of your explanatory variables are insignificant). Remember to be careful with your interpretation of the marginal effects of coeffi cients If logging solves heteroskedasticity problem, see page 111 of textbook for interpretation of logged coeffi cients If multiply/dividing Z j solves the problem, see page 129 for interpretation of coeffi cients () Introductory Econometrics: Topic 5 34 / 94

35 Example: Explaining House Prices Data set used before, here is a reminder: N = 546 houses sold in Windsor, Canada. The dependent variable, Y, is the sales price of the house in Canadian dollars. The explanatory variables are: X 1 = the lot size of the property (in square feet) X 2 = the number of bedrooms X 3 = the number of bathrooms X 4 = the number of storeys (excluding the basement) () Introductory Econometrics: Topic 5 35 / 94

36 D 1 = 1 if the house has a driveway (= 0 if it does not) D 2 = 1 if the house has a recreation room (= 0 if not) D 3 = 1 if the house has a basement (= 0 if not) D 4 = 1 if the house has gas central heating (= 0 if not) D 5 = 1 if the house has air conditioning (= 0 if not). () Introductory Econometrics: Topic 5 36 / 94

37 Breusch-Pagan and White tests require the selection of variables which might be related to the error variance I set these variables to be the explanatory variables in the original regression I find BP = and W = Critical values for both of these tests are taken from the χ 2 (9) distribution. The 5% critical value is Since both test statistics are greater than critical value, both of these tests indicate that heteroskedasticity is present Thus, although the OLS estimates presented in Topic 3 slides for this data set were unbiased, the confidence intervals and hypothesis tests were incorrect. () Introductory Econometrics: Topic 5 37 / 94

38 I next experimented with different transformations to try and eliminate the heteroskedasticity problem until I found one that eliminated heteroskedasticity The one that worked was to take logs of all variables except for the dummies (since the log of zero is undefined, you cannot take the log of a dummy variable) In this log-linear regression, the values for the two heteroskedasticity tests are BP = and W = which are both less than the 5% critical value Hence, both tests fail to reject the hypothesis of homoskedasticity. Thus, the log transformation has solved the heteroskedasticity problem in this example. Remember: β j in log linear regression is interpreted as an elasticity () Introductory Econometrics: Topic 5 38 / 94

39 Autocorrelation We will continue our discussion of problems which call for the use of the Generalized Least Squares Estimator by considering autocorrelation. This is used with time series data, so we will use t = 1,.., T to denote observations (rather than i = 1,.., N) Reminder of basic theoretical Results Previously derived theoretical results using multiple regression model with classical assumptions: 1 E (ε i ) = 0 mean zero errors. 2 var (ε i ) = E ( ε 2 ) i = σ 2 constant variance errors (homoskedasticity). 3 cov (ε i ε j ) = 0 for i = j. 4 ε i is Normally distributed 5 Explanatory variables are fixed. They are not random variables. () Introductory Econometrics: Topic 5 39 / 94

40 Remember: Assumption 1 is innocuous (if the error had a non-zero mean we could include it as part of the intercept it would have no effect on estimation of slope coeffi cients in the model). Assumption 4 can be relaxed (approximately) by using asymptotic theory. Assumption 5 we will still maintain. Autocorrelation (also called serial correlation) relates to Assumption 3. () Introductory Econometrics: Topic 5 40 / 94

41 Basic ideas: Under classical assumptions, Gauss Markov theorem says "OLS is BLUE". But if Assumptions 2 and 3 are violated OLS this no longer holds (OLS is still unbiased, but is no longer "best". i.e. no longer minimum variance). Concepts/proofs/derivations use following strategy. The model can be transformed to create a new model which does satisfy classical assumptions. We know OLS (on the transformed model) will be BLUE. (And all the theory we worked out for the OLS estimator will hold except it will hold for the transformed model). The OLS estimator using such a transformed model is called the Generalized Least Squares (GLS) estimator. () Introductory Econometrics: Topic 5 41 / 94

42 Autocorrelated Errors Use multiple regression model under the classical assumptions, with the exception that the errors follow an autoregressive process of order 1 (AR(1)): ε t = ρε t 1 + u t where it is u t which satisfies classical assumptions. So E (u t ) = 0, var (u t ) = σ 2 and cov (u t, u s ) = 0 (for t = s). We also assume 1 < ρ < 1. To preview later material, this restriction ensure stationarity and means you do not have to worry about problems relating to unit roots and cointegration (definitions will be provided to you later on). Focus on the AR(1) cases, but note that the AR(p) errors case is a simple extension: ε t = ρ 1 ε t 1 + ρ 2 ε t ρ p ε t p + u t () Introductory Econometrics: Topic 5 42 / 94

43 Variances and Covariances of Regression Error Assumptions above specified properties of u t, but we need to know properties of ε t. Notation: σ 2 ε = var (ε t ) = E ( ε 2 ) t where last equal sign follows since errors have mean zero. Derivation of variance of regression errors (textbook does derivation in different way): σ 2 ε = var (ρε t 1 + u t ) = ρ 2 var (ε t 1 ) + var (u t ) = ρ 2 σ 2 ε + σ 2 = σ 2 1 ρ 2 In previous derivations used properties of variance operator, the fact that ε t 1 and u t are independent of one another and that ε t is homoskedastic. () Introductory Econometrics: Topic 5 43 / 94

44 The derivation of covariance between different regression errors will be done in Problem Sheet 4: cov (ε t, ε t 1 ) = ρσ 2 ε For errors more than one period apart, can show: cov (ε t, ε t s ) = ρ s σ 2 ε Thus, the regression model with autocorrelated errors violates assumption 3. Regression errors are NOT uncorrelated with one another. Hence, we need to work with a GLS estimator. () Introductory Econometrics: Topic 5 44 / 94

45 The GLS Estimator for the Autocorrelated Errors Case Remember: GLS can be interpreted as OLS on a suitably transformed model. In this case, the appropriate transformation is referred to as "quasi-differencing". To explain what this is, consider the regression model: Y t = α + β 1 X 1t β k X kt + ε t Holds for every time period so we can take it at period t 1 and multiply both sides of the equation by ρ: ρy t 1 = ρα + ρβ 1 X 1t ρβ k X kt 1 + ρε t 1 Subtract this equation from the original regression equation: Y t ρy t 1 = α ρα + β 1 (X 1t ρx 1t 1 ) β k (X kt ρx kt 1 ) + ε t ρε t 1 () Introductory Econometrics: Topic 5 45 / 94

46 or Y t = α + β 1 X 1t β k X kt + u t But u t satisfies the classical assumptions so OLS on this transformed model will be GLS (which will be BLUE). Note: transformed variables are "quasi-differenced" = Y t ρy t 1 X1t = (X 1t ρx 1t 1 ) etc. Y t The case with ρ = 1 is called "differenced" this is not quite the same so we say "quasi" differenced. () Introductory Econometrics: Topic 5 46 / 94

47 One (relatively minor) issue: if our original data is from t = 1,.., T then Y 1 = Y 1 ρy 0 will involve Y 0 (and same issue for explanatory variables). But we do not observe such "initial conditions". There are many ways of treating initial conditions. Simplest is to work with data from t = 2,.., T (and use t = 1 values for variables as initial conditions). Summary: If we knew ρ, then we could quasi-difference the data and do OLS using the transformed data (which is equivalent to GLS). In practice, we rarely (if ever) know ρ. Hence, replace ρ by an estimate: ρ. This is what Cochrane-Orcutt procedure does () Introductory Econometrics: Topic 5 47 / 94

48 The Cochrane-Orcutt Procedure Remember: with autocorrelated errors, GLS is BLUE. However, OLS (on original data) is still unbiased. Cochrane-Orcutt procedure begins with OLS and then uses OLS residuals to estimate ρ. Cochrane-Orcutt procedure goes through following steps: () Introductory Econometrics: Topic 5 48 / 94

49 1 Do OLS regression of Y t on intercept, X 1t,.., X kt and produce the OLS residuals, ε t. 2 Do OLS regression of ε t on ε t 1 which will provide a ρ. 3 Quasi-difference all variables to produce = Y t ρy t 1 X1t = (X 1t ρx 1t 1 ) etc. Y t 4 Do OLS regression of Yt on intercept, X1t,.., X kt, thus producing GLS estimates of the coeffi cients. () Introductory Econometrics: Topic 5 49 / 94

50 Autocorrelation Consistent Estimators Remember: with heteroskedasticity we discussed heteroskedasticity consistent estimator (HCE). Less effi cient than GLS, but is a correct second-best solution when GLS diffi cult to implement. Similar issues hold autocorrelated errors. There exist autocorrelation consistent estimators which allow for the correct use of OLS methods when you have autocorrelated errors. We will not explain these, but many popular econometrics software packages include them. The most popular is the Newey-West estimator (which corrects for heteroskedasticity as well). () Introductory Econometrics: Topic 5 50 / 94

51 Testing for Autocorrelated Errors If ρ = 0 then doing OLS on the original data is fine (OLS is BLUE). However, if ρ = 0, then a GLS estimator such as the Cochrane-Orcutt estimator is better. This motivates testing H 0 : ρ = 0 against H 1 : ρ = 0. There are several such tests, here we describe some of the most popular. () Introductory Econometrics: Topic 5 51 / 94

52 Breusch-Godfrey Test This is a test of H 0 : ρ 1 = 0, ρ 2 = 0,.., ρ p = 0 in the regression model with AR(p) errors: ε t = ρ 1 ε t 1 + ρ 2 ε t ρ p ε t p + u t Breusch-Godfrey test involves the following steps: 1 Run a regression of Y t on an intercept, X 1,.., X k using OLS and produce the residuals, ε t. 2 Run second regression of ε t on intercept, X 1,.., X k, ε t 1,.., ε t p using OLS and produce the R 2. 3 Calculate the test statistic: LM = TR 2 If H 0 is true, then LM has an (approximate) χ 2 (p) distribution. Thus, critical value taken from statistical tables for the Chi-square distribution. () Introductory Econometrics: Topic 5 52 / 94

53 The Box-Pierce and Ljung Tests These test H 0 : ρ 1 = 0, ρ 2 = 0,.., ρ p = 0 Both based on idea that, if the errors are not autocorrelated, then the correlations between different errors should be zero. Replace errors by residuals. ε t are residuals from OLS regression of Y on an intercept, X 1,.., X k, Correlations between ε t and ε t s are: r s = T t=s+1 ε t ε t s T t=s+1 ε2 t () Introductory Econometrics: Topic 5 53 / 94

54 Box-Pierce test statistic (sometimes called the Q test statistic) is: Q = T p rj 2 j=1 The p means that AR(p) errors are being tested for. The Ljung test statistic is: Q = T (T + 2) p j=1 r 2 j T j Critical values for both taken from χ 2 (p) tables. Econometrics software packages such as Gretl will calculate test statistics () Introductory Econometrics: Topic 5 54 / 94

55 Warning: in some cases, one of the explanatory variables will be the dependent variable from a previous period ("lagged dependent variable"). For instance: Y t = α + δy t 1 + βx t + ε t The Box-Pierce and Ljung tests are not appropriate in this case. The Breusch-Godfrey test, however, is still appropriate. The textbook discusses two other approaches: the Durbin-Watson statistic and Durbin s h-test. () Introductory Econometrics: Topic 5 55 / 94

56 Instrumental Variable Methods Overview: Under the classical assumptions, OLS is BLUE. When we relax Assumptions 2 or 3 (e.g. to allow for heteroskedasticity or autocorrelated errors), then OLS is no longer BLUE but it is still unbiased (although GLS is better) However, in the case we are about to consider, OLS will be biased and an entirely different estimator will be called for the instrumental variables (IV) estimator. Here we relax assumption that explanatory variables are not random variables. For simplicity, use simple regression model, but results generalize to the case of multiple regression. () Introductory Econometrics: Topic 5 56 / 94

57 Theory Motivating the IV Estimator Reminder of basic theoretical Results Previously derived theoretical results using multiple regression model with classical assumptions: 1 E (ε i ) = 0 mean zero errors. 2 var (ε i ) = E ( ε 2 ) i = σ 2 constant variance errors (homoskedasticity). 3 cov (ε i ε j ) = 0 for i = j. 4 ε i is Normally distributed 5 Explanatory variables are fixed. They are not random variables. () Introductory Econometrics: Topic 5 57 / 94

58 Remember: Assumption 1 is innocuous. Assumption 4 can be relaxed (approximately) by using asymptotic theory. Assumptions 2 and 3 were covered above (heteroskedasticity and autocorrelated errors). Now we will focus on relaxing Assumption 5. Note: When explanatory variables are be random, many derivations we did before with expected value and variance operators become much more diffi cult/impossible. For this reason, most relevant results are asymptotic. But asymptotic methods not covered in course (see appendix to Chapter 5 if you are interested) These lecture slides provide some intuition, hints at derivations and discussion of things relevant for empirical practice. () Introductory Econometrics: Topic 5 58 / 94

59 Case 1: Explanatory Variable is Random But is Uncorrelated with Error If X i is now a random variable, we have to make some assumptions about its distribution. Assume X i are independent and identically distributed (i.i.d.) random variables with: E (X i ) = µ X var (X i ) = σ 2 X In Case 1 we will assume explanatory variable and errors are uncorrelated with one another: cov (X i, ε i ) = E (X i ε i ) = 0 () Introductory Econometrics: Topic 5 59 / 94

60 Remember, under classical assumptions: ( ) σ β 2 is N β, Xi 2 This result can still be shown to hold approximately in this case (I will not provide details, some given in textbook) Bottom line: If we relax the assumptions of Normality and fixed explanatory variables we get exactly the same results as for OLS under the classical assumptions (but here they hold approximately), provided explanatory variables are uncorrelated with the error term. () Introductory Econometrics: Topic 5 60 / 94

61 Case 2: Explanatory Variable is Correlated with the Error Term Assume X i are i.i.d. random variables with: E (X i ) = µ X var (X i ) = σ 2 X In Case 2 we will assume explanatory variable and errors are correlated with one another: cov (X i, ε i ) = E (X i ε i ) = 0 It turns out that, in this case, OLS is biased and a new estimator is called for: the instrumental variables (IV) estimator. Why is this? We will not provide proof, but outline the basic idea () Introductory Econometrics: Topic 5 61 / 94

62 The proof that OLS is biased begins in the same manner as the proof of unbiasedness under classical assumptions (see Topic 3 slides or page 68 of textbook). We can get up to the following stage in the proof: ) ( β) E = β + E ( X i ε i Xi 2 But at this stage we can go ( no farther ) other than to note that there is no reason to think that E X i ε i = 0 and, in fact, it is not. X i 2 Intuition: (ignoring Xi 2 in the denominator), the numerator is E ( ) X i ε i = E (Xi ε i ) = cov (X i ε i ) = 0 Important point: if the error and explanatory variable are correlated, then OLS is biased and should be avoided. Soon will offer some explanation for why this might occur, but first introduce new estimator to handle this case. () Introductory Econometrics: Topic 5 62 / 94

63 The Instrumental Variables Estimator An instrumental variable (IV), Z i, is a random variable which is uncorrelated with the error but is correlated with the explanatory variable. Formally, an instrumental variable is assumed to satisfy the following assumptions: E (Z i ) = µ Z var (Z i ) = σ 2 Z cov (Z i, ε i ) = E (Z i ε i ) = 0 cov (X i, Z i ) = E (X i Z i ) µ Z µ X = σ XZ = 0 Assuming an instrumental variable exists (something we will return to later), IV estimator: β IV = N i=1 Z i Y i N X i Z i i=1 () Introductory Econometrics: Topic 5 63 / 94

64 The asymptotic derivations in the textbook (not covered in this course) imply (approximately): ( ( σ 2 β IV is N β, Z + µ 2 ) ) Z σ 2 N (σ XZ + µ X µ Z ) 2 ) Note: this implies E ( β IV = β so unbiased (approximately) This formula can be used to calculate confidence intervals, hypothesis tests, etc. (comparable to Topic 3 slide derivations) In practice, the unknown means and variances can be replaced by their sample counterparts. Thus, µ X can be replaced by X, σ 2 Z by the sample variance of (Z i Z) 2 N 1, etc. No additional details of how this is done, but note that econometrics software packages do IV estimation. Note: this is sometimes called the two stage least squares estimator (and Gretl uses this term) () Introductory Econometrics: Topic 5 64 / 94

65 Using the IV Estimator in Practice what if you have a multiple regression model involving more than one explanatory variable? Answer: you need at least one instrumental variable for each explanatory variable that is correlated with the error. Note: if even only one of the explanatory variables in a multiple regression model is correlated with the errors, OLS estimates of all coeffi cients can be biased. what if you have more instrumental variables than you need? Use the generalized instrumental variables estimator (GIVE). () Introductory Econometrics: Topic 5 65 / 94

66 The Generalized Instrumental Variables Estimator Illustrate in simple regression model with one explanatory variable, X. X is correlated with error, so IV estimator is needed Suppose two instrumental variables exist: Z 1 and Z 2. GIVE involves running an initial regression of the explanatory variable on the instruments: X i = γ 0 + γ 1 Z 1i + γ 2 Z 2i + u i OLS estimation of this initial regression provides fitted values: X i = γ 0 + γ 1 Z 1i + γ 2 Z 2i. () Introductory Econometrics: Topic 5 66 / 94

67 It can be shown that X is uncorrelated with the errors in the original regression and, hence, that it is a suitable instrument. This is what GIVE uses as an instrument. Thus, the GIVE is given by: β GIVE = N i=1 X i Y i N X i X i i=1 () Introductory Econometrics: Topic 5 67 / 94

68 The Hausman Test Hausman test used to see if IV estimator is necessary If X correlated with error, then OLS is biased and you should use an IV estimator. However, if X uncorrelated with error, then (under the classical assumptions) OLS is BLUE and, hence, is more effi cient than IV. Basic idea of Hasuman test (Gretl will calculate it for you): Let H 0 be the hypothesis that explanatory variables are uncorrelated with error. If H 0 is true, then both OLS and IV are both acceptable estimators and should give roughly the same result. However, if H 0 is false, then OLS in not acceptable, but IV is and they can be quite different. Hausman test based measuring the difference between β and β IV. () Introductory Econometrics: Topic 5 68 / 94

69 It can be shown that Hausman test can be done using OLS methods Example: simple regression with one instrument. Original regression model of interest is: Y i = α + βx i + ε i Hausman test uses this regression but adds the instrumental variable to it: Y i = α + βx i + γz i + ε i Can be shown that Hausman test is equivalent to t-test of H 0 : γ = 0. If coeffi cient on Z is significant, you reject H 0 and use the IV estimator. If coeffi cient on Z is insignificant, do OLS on original regression model () Introductory Econometrics: Topic 5 69 / 94

70 Example: Hausman test in multiple regression model Suppose we have 3 explanatory variables (X 1, X 2 and X 3 ), 2 of which might be correlated with the error (X 2 and X 3 ). For X 2 we have two possible instrumental variables (Z 1 and Z 2 ), For X 3 we have three possible instrumental variables (Z 3, Z 4 and Z 5 ) Since potentially more instruments than explanatory variables, might want to use GIVE Hausman test runs initial GIVE regressions (see the discussion of the GIVE) to get fitted values, X 2 and X 3 Then adds these fitted values to original regression and tests if they are significant Precise steps given on next slide () Introductory Econometrics: Topic 5 70 / 94

71 Run an OLS regression of X 2 on an intercept, Z 1 and Z 2. Obtain the fitted values, X 2. Run an OLS regression of X 3 on an intercept Z 3, Z 4 and Z 5. Obtain the fitted values, X 3. Run an OLS regression of Y on an intercept X 1, X 2, X 3, X 2 and X 3. Do an F-test of the hypothesis that the coeffi cients on X 2 and X 3 are jointly equal to zero. If the F-test rejects, then proceed with GIVE, if not use OLS on the original regression. () Introductory Econometrics: Topic 5 71 / 94

72 The Sargan Test Hausman test tells you whether it is necessary to do instrumental variables estimation, if we have valid instrumental variables But how do we know that Z truly is a valid instrument? Remember to be a valid instrument, Z must be correlated with X but uncorrelated with regression error. The latter is not easy to check, since the regression error is not observed. The issue of testing whether instruments truly are valid ones is diffi cult If you have only one potential instrument for every explanatory variable, there is no way of testing instrument validity Next slide illustrates why this is so () Introductory Econometrics: Topic 5 72 / 94

73 Suppose one potential instrument, Z for a simple regression model: Y i = βx i + ε i For Z to be instrument must have cov (Z i, ε i ) = 0. ε i is unobserved, but why not create a test statistic based on cov (Z i, ε i ) using residuals But which residuals should you use? IV residuals (i.e. ε IV i = Y i β IV X i ) sound plausible, but will be unacceptable if Z is not a valid instrument OLS residuals are inappropriate since they could very well be biased (and you cannot use a Hausman test to check this, since you are not sure that Z is a valid instrument). In fact, there is no way of testing whether the instruments are valid in such a case. () Introductory Econometrics: Topic 5 73 / 94

74 But if more potential instruments than explanatory variables, there is a test for whether instruments are valid: the Sargan test Suppose you have k explanatory variables which all are potentially correlated with the error. Have r instrumental variables where r > k. The Sargan test involves the following steps: 1 Run regression of Y on X 1,.., X k using GIVE and obtain IV residuals, ε IV i. 2 Run OLS regression of ε IV i the R 2. on all the instruments, Z 1,.., Z r and obtain 3 Sargan test statistic is NR 2 and critical value can be obtained from χ 2 (r k) 4 If test statistic less than critical value conclude they are valid instruments, else they are not. () Introductory Econometrics: Topic 5 74 / 94

75 Why Might the Explanatory Variable Be Correlated with Error? There are many different reasons why the explanatory variables might be correlated with the errors. In these slides we discuss two reasons and give an example Errors in Variables problem Simultaneous equations model () Introductory Econometrics: Topic 5 75 / 94

76 Errors in Variables What if you want to run the regression: Y i = βx i + ε i Suppose this regression satisfies the classical assumptions but you do not observe x i, but instead observe: X i = X i + v i where v i is i.i.d. with mean zero, variance σ 2 ν and is independent of ε i. () Introductory Econometrics: Topic 5 76 / 94

77 In other words, X is observed with error. Replacing X i in the original regression yields a new regression: Y i = β (Xi v i ) + ε i = βxi + εi where εi = ε i βv i What is the covariance between the explanatory variable, Xi, and the error, εi, in this new regression? cov (Xi, εi ) = E [(X i + v i ) (ε i βv i )] = βσ 2 ν = 0 Hence measurement error in explanatory variables (but not dependent variable), causes them to be correlated with the regression error. () Introductory Econometrics: Topic 5 77 / 94

78 The Simultaneous Equations Model Example: supply and demand model Demand curve: Supply curve Q D = β D P + ε D Q S = β S P + ε S Q D is quantity demanded, Q S is quantity supplied, P is price In equilibrium, Q D = Q S and this condition is used to solve for equilibrium price and quantity. Price and quantity are determined by this model so they are both endogenous variables. These two equations are a simultaneous equations model. () Introductory Econometrics: Topic 5 78 / 94

79 Suppose econometrician collects data on price and quantity and runs a regression of quantity on price. This will yield an estimate β. But would it be an estimate of β D or β S? There is no way of knowing. OLS will probably neither estimate the supply curve nor estimate the demand curve. () Introductory Econometrics: Topic 5 79 / 94

80 How might we solve this problem? Include exogenous variable (not determined by the model). E.g. suppose Q D depends also on the income, I : Q D = β D P + γi + ε D Notation: original supply and demand model is structural form. When we solve model so that only exogenous variables on right hand side, we get reduced form () Introductory Econometrics: Topic 5 80 / 94

81 Setting Q D = Q S and solving for equilibrium P and Q Set right-hand sides of the demand and supply equations equal to each other, and rearrange: P = γ β D β S I + ε S ε D β D β S = π 1 I + ε 1 Substitute this expression for P into the supply equation: Q = β S (π 1 I + ε 1 ) + ε S = β S π 1 I + β S ε 1 + ε S = π 2 I + ε 2 γ β D β S Note: I am using π 1 = and π 2 = β S π 1 as notation for reduced form coeffi cients ε 1 and ε 2 are reduced form errors which are functions of the structural form errors This example shows what reduced and structural form are, but what are the implications for the econometrics? () Introductory Econometrics: Topic 5 81 / 94

82 Directly Estimating Structural Equations using OLS Leads to Biased Estimates Why not run a regression based on the demand curve? Regression of Q on P and I? Can show explanatory variable P is correlated with the error. As shown in this section of lecture slides, OLS is not appropriate To prove this, assume ε S and ε D are uncorrelated and homoskedastic var (ε D ) = σ 2 D var (ε S ) = σ 2 S Assume exogenous variable I is not random () Introductory Econometrics: Topic 5 82 / 94

83 cov (P, ε D ) = E (Pε D ) = E [(π 1 I + ε 1 ) ε D ] [( ) ] εs ε = D E ε D β D β S σ 2 D = β D β S = 0, () Introductory Econometrics: Topic 5 83 / 94

84 OLS Estimation of Reduced Form Equations is Fine It is not hard to show that reduced form errors do satisfy the classical assumptions. Thus, OLS estimation of reduced form equations is BLUE So π 1 and π 2 are BLUE, but what are they estimates of? For instance, π 1 = demand curve. γ β D β S which is not the slope of supply curve or But here we can estimate slope of supply curve using π 1 and π 2 We have π 2 = β S π 1 and, thus, β S = π 2 π 1 Such a strategy is called indirect least squares. () Introductory Econometrics: Topic 5 84 / 94

85 How does all this relate to instrumental variables? Easy to show that indirect least squares is equivalent to instrumental variable estimation of supply curve using I as an instrument This example sheds insight on how to choose instruments Note: here we are estimating the supply curve, but there is no way of estimating the demand curve Why is this? Key thing is that income does not appear in supply curve. A general rule is: if you have an exogenous variable which is excluded from an equation, then it can be used as an instrumental variable for that equation. This occurs for supply curve, but not demand curve () Introductory Econometrics: Topic 5 85 / 94

86 An example where the explanatory variable could be correlated with the error Suppose interested in estimating returns to schooling and have data from a survey of many individuals on: The dependent variable: Y = income The explanatory variable: X = years of schooling And other explanatory variables like experience, age, occupation, etc.. which we will ignore here to simplify exposition. My contention is that, in such a regression it probably is the case that X is correlated with the error and, thus, OLS will be inappropriate. () Introductory Econometrics: Topic 5 86 / 94

87 To understand why, first think of how errors are interpreted in this regression. An individual with a positive error is earning an unusually high level of income. That is, his/her income is more than his/her education would suggest. An individual with a negative error is earning an unusually low level of income. That is, his/her income is less than his/her education would suggest. What might be correlated with this error? Perhaps each individual has some underlying quality (e.g. intelligence, ambition, drive, talent, luck or even family encouragement). () Introductory Econometrics: Topic 5 87 / 94

88 This quality would like be associated with the error (e.g. individuals with more drive tend to achieve unusually high incomes). But this quality would also effect the schooling choice of the individual. For instance, ambitious students would be more likely to go to university. Summary: Ambitious, intelligent, driven individuals would both tend to have more schooling and more income (i.e. positive errors). So both the error and the explanatory variable would be influenced by this quality. Error and explanatory variable probably would be correlated with one another. () Introductory Econometrics: Topic 5 88 / 94

89 How do you choose instrumental variables? There is a lot of discussion in the literature how to do this. But this is too extensive and complicated for this course, so we offer a few practical thoughts. An instrumental variable should be correlated with explanatory variable, but not with error. Sometimes economic theory (or common sense) suggests variables with this property. In our example, we want a variable which is correlated with the schooling decision, but is unrelated to error (i.e. factors which might explain why individuals have unusually high/low incomes) An alternative way of saying this: we want to find a variable which affects schooling choice, but has no direct effect on income. () Introductory Econometrics: Topic 5 89 / 94

90 Characteristics of parents or older siblings have been used as instruments. Justification: if either of your parents had a university degree, then you probably come from a family where education is valued (increase the chances you go to university). However, your employer will not care that your parents went to university (so no direct effect on your income). Other researchers have used geographical location variables as instruments. Justification: if you live in a community where a university/college is you are more likely to go to university. However, your employer will not care where you lived so location variable will have no direct effect on your income. () Introductory Econometrics: Topic 5 90 / 94

91 Advice for Empirical Practice Divide your variables into the dependent variable, the ones(s) which may be endogenous explanatory variables, the ones which are exogenous explanatory variables and the ones which might be instruments Economic theory and common sense can help you make an initial classification Previous example illustrates this Then use Hausman and Sargan tests to try and see if your classification is a good one and, if not, experiment with adjusting classifications Also: a valid instrument should not have explanatory power for y (after controlling for effect of exogenous explanatory variables) Hence, can experiment with regressions including instruments: they should not be significant (in a regression which passes the Hausman test) () Introductory Econometrics: Topic 5 91 / 94

92 Chapter Summary Chapter discusses violations of classical assumptions and breaks into a "GLS" part and an "IV" part If errors either have different variances (heteroskedasticity) or are correlated with one another, then OLS is unbiased, but is no longer the best estimator. The best estimator is GLS. If heteroskedasticity is present, then the GLS estimator can be calculated using OLS on a transformed model. If suitable transformation cannot be found, then heteroskedasticity consistent estimator should be used. There are many tests for heteroskedasticity, including the Breusch-Pagan and the White test. () Introductory Econometrics: Topic 5 92 / 94

93 If errors are autocorrelated, GLS estimator is OLS on a transformed model. The required transformation involves quasi-differencing each variable. The Cochrane-Orcutt procedure is a popular way of implementing the GLS estimator. There are many tests for autocorrelated errors, including the Breusch-Godfrey test, the Box-Pierce test and the Ljung test. In many applications, it is implausible to treat the explanatory variables and fixed. Hence, it is important to allow for them to be random variables. If explanatory variables are random and all of them are uncorrelated with the regression error, then standard methods associated with OLS still work. If explanatory variables are random and some of them are correlated with the regression error, then OLS is biased. The instrumental variables estimator is not. () Introductory Econometrics: Topic 5 93 / 94

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

1 The Multiple Regression Model: Freeing Up the Classical Assumptions 1 The Multiple Regression Model: Freeing Up the Classical Assumptions Some or all of classical assumptions were crucial for many of the derivations of the previous chapters. Derivation of the OLS estimator

More information

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43 Panel Data March 2, 212 () Applied Economoetrics: Topic March 2, 212 1 / 43 Overview Many economic applications involve panel data. Panel data has both cross-sectional and time series aspects. Regression

More information

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Econometrics Week 4 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 23 Recommended Reading For the today Serial correlation and heteroskedasticity in

More information

1 A Non-technical Introduction to Regression

1 A Non-technical Introduction to Regression 1 A Non-technical Introduction to Regression Chapters 1 and Chapter 2 of the textbook are reviews of material you should know from your previous study (e.g. in your second year course). They cover, in

More information

2. Linear regression with multiple regressors

2. Linear regression with multiple regressors 2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions

More information

Introduction to Econometrics. Heteroskedasticity

Introduction to Econometrics. Heteroskedasticity Introduction to Econometrics Introduction Heteroskedasticity When the variance of the errors changes across segments of the population, where the segments are determined by different values for the explanatory

More information

1 Motivation for Instrumental Variable (IV) Regression

1 Motivation for Instrumental Variable (IV) Regression ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data

More information

F9 F10: Autocorrelation

F9 F10: Autocorrelation F9 F10: Autocorrelation Feng Li Department of Statistics, Stockholm University Introduction In the classic regression model we assume cov(u i, u j x i, x k ) = E(u i, u j ) = 0 What if we break the assumption?

More information

Econometrics. 9) Heteroscedasticity and autocorrelation

Econometrics. 9) Heteroscedasticity and autocorrelation 30C00200 Econometrics 9) Heteroscedasticity and autocorrelation Timo Kuosmanen Professor, Ph.D. http://nomepre.net/index.php/timokuosmanen Today s topics Heteroscedasticity Possible causes Testing for

More information

Reading Assignment. Serial Correlation and Heteroskedasticity. Chapters 12 and 11. Kennedy: Chapter 8. AREC-ECON 535 Lec F1 1

Reading Assignment. Serial Correlation and Heteroskedasticity. Chapters 12 and 11. Kennedy: Chapter 8. AREC-ECON 535 Lec F1 1 Reading Assignment Serial Correlation and Heteroskedasticity Chapters 1 and 11. Kennedy: Chapter 8. AREC-ECON 535 Lec F1 1 Serial Correlation or Autocorrelation y t = β 0 + β 1 x 1t + β x t +... + β k

More information

Economics 308: Econometrics Professor Moody

Economics 308: Econometrics Professor Moody Economics 308: Econometrics Professor Moody References on reserve: Text Moody, Basic Econometrics with Stata (BES) Pindyck and Rubinfeld, Econometric Models and Economic Forecasts (PR) Wooldridge, Jeffrey

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

The regression model with one stochastic regressor (part II)

The regression model with one stochastic regressor (part II) The regression model with one stochastic regressor (part II) 3150/4150 Lecture 7 Ragnar Nymoen 6 Feb 2012 We will finish Lecture topic 4: The regression model with stochastic regressor We will first look

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Lecture 4: Heteroskedasticity

Lecture 4: Heteroskedasticity Lecture 4: Heteroskedasticity Econometric Methods Warsaw School of Economics (4) Heteroskedasticity 1 / 24 Outline 1 What is heteroskedasticity? 2 Testing for heteroskedasticity White Goldfeld-Quandt Breusch-Pagan

More information

LECTURE 10: MORE ON RANDOM PROCESSES

LECTURE 10: MORE ON RANDOM PROCESSES LECTURE 10: MORE ON RANDOM PROCESSES AND SERIAL CORRELATION 2 Classification of random processes (cont d) stationary vs. non-stationary processes stationary = distribution does not change over time more

More information

Linear Regression with Time Series Data

Linear Regression with Time Series Data Econometrics 2 Linear Regression with Time Series Data Heino Bohn Nielsen 1of21 Outline (1) The linear regression model, identification and estimation. (2) Assumptions and results: (a) Consistency. (b)

More information

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Econometrics Week 8 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 25 Recommended Reading For the today Instrumental Variables Estimation and Two Stage

More information

EC4051 Project and Introductory Econometrics

EC4051 Project and Introductory Econometrics EC4051 Project and Introductory Econometrics Dudley Cooke Trinity College Dublin Dudley Cooke (Trinity College Dublin) Intro to Econometrics 1 / 23 Project Guidelines Each student is required to undertake

More information

LECTURE 11. Introduction to Econometrics. Autocorrelation

LECTURE 11. Introduction to Econometrics. Autocorrelation LECTURE 11 Introduction to Econometrics Autocorrelation November 29, 2016 1 / 24 ON PREVIOUS LECTURES We discussed the specification of a regression equation Specification consists of choosing: 1. correct

More information

Environmental Econometrics

Environmental Econometrics Environmental Econometrics Syngjoo Choi Fall 2008 Environmental Econometrics (GR03) Fall 2008 1 / 37 Syllabus I This is an introductory econometrics course which assumes no prior knowledge on econometrics;

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory

More information

ECNS 561 Multiple Regression Analysis

ECNS 561 Multiple Regression Analysis ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking

More information

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) 1 2 Model Consider a system of two regressions y 1 = β 1 y 2 + u 1 (1) y 2 = β 2 y 1 + u 2 (2) This is a simultaneous equation model

More information

AUTOCORRELATION. Phung Thanh Binh

AUTOCORRELATION. Phung Thanh Binh AUTOCORRELATION Phung Thanh Binh OUTLINE Time series Gauss-Markov conditions The nature of autocorrelation Causes of autocorrelation Consequences of autocorrelation Detecting autocorrelation Remedial measures

More information

Types of economic data

Types of economic data Types of economic data Time series data Cross-sectional data Panel data 1 1-2 1-3 1-4 1-5 The distinction between qualitative and quantitative data The previous data sets can be used to illustrate an important

More information

Econometrics - 30C00200

Econometrics - 30C00200 Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business

More information

Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation

Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation 1/30 Outline Basic Econometrics in Transportation Autocorrelation Amir Samimi What is the nature of autocorrelation? What are the theoretical and practical consequences of autocorrelation? Since the assumption

More information

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)

More information

Linear Regression with Time Series Data

Linear Regression with Time Series Data u n i v e r s i t y o f c o p e n h a g e n d e p a r t m e n t o f e c o n o m i c s Econometrics II Linear Regression with Time Series Data Morten Nyboe Tabor u n i v e r s i t y o f c o p e n h a g

More information

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1 Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Many economic models involve endogeneity: that is, a theoretical relationship does not fit

More information

Lectures 5 & 6: Hypothesis Testing

Lectures 5 & 6: Hypothesis Testing Lectures 5 & 6: Hypothesis Testing in which you learn to apply the concept of statistical significance to OLS estimates, learn the concept of t values, how to use them in regression work and come across

More information

11.1 Gujarati(2003): Chapter 12

11.1 Gujarati(2003): Chapter 12 11.1 Gujarati(2003): Chapter 12 Time Series Data 11.2 Time series process of economic variables e.g., GDP, M1, interest rate, echange rate, imports, eports, inflation rate, etc. Realization An observed

More information

An overview of applied econometrics

An overview of applied econometrics An overview of applied econometrics Jo Thori Lind September 4, 2011 1 Introduction This note is intended as a brief overview of what is necessary to read and understand journal articles with empirical

More information

Multiple Regression Analysis: Heteroskedasticity

Multiple Regression Analysis: Heteroskedasticity Multiple Regression Analysis: Heteroskedasticity y = β 0 + β 1 x 1 + β x +... β k x k + u Read chapter 8. EE45 -Chaiyuth Punyasavatsut 1 topics 8.1 Heteroskedasticity and OLS 8. Robust estimation 8.3 Testing

More information

Week 11 Heteroskedasticity and Autocorrelation

Week 11 Heteroskedasticity and Autocorrelation Week 11 Heteroskedasticity and Autocorrelation İnsan TUNALI Econ 511 Econometrics I Koç University 27 November 2018 Lecture outline 1. OLS and assumptions on V(ε) 2. Violations of V(ε) σ 2 I: 1. Heteroskedasticity

More information

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE 1. You have data on years of work experience, EXPER, its square, EXPER, years of education, EDUC, and the log of hourly wages, LWAGE You estimate the following regressions: (1) LWAGE =.00 + 0.05*EDUC +

More information

Applied Quantitative Methods II

Applied Quantitative Methods II Applied Quantitative Methods II Lecture 4: OLS and Statistics revision Klára Kaĺıšková Klára Kaĺıšková AQM II - Lecture 4 VŠE, SS 2016/17 1 / 68 Outline 1 Econometric analysis Properties of an estimator

More information

Iris Wang.

Iris Wang. Chapter 10: Multicollinearity Iris Wang iris.wang@kau.se Econometric problems Multicollinearity What does it mean? A high degree of correlation amongst the explanatory variables What are its consequences?

More information

A Course on Advanced Econometrics

A Course on Advanced Econometrics A Course on Advanced Econometrics Yongmiao Hong The Ernest S. Liu Professor of Economics & International Studies Cornell University Course Introduction: Modern economies are full of uncertainties and risk.

More information

Regression with time series

Regression with time series Regression with time series Class Notes Manuel Arellano February 22, 2018 1 Classical regression model with time series Model and assumptions The basic assumption is E y t x 1,, x T = E y t x t = x tβ

More information

ECON 4551 Econometrics II Memorial University of Newfoundland. Panel Data Models. Adapted from Vera Tabakova s notes

ECON 4551 Econometrics II Memorial University of Newfoundland. Panel Data Models. Adapted from Vera Tabakova s notes ECON 4551 Econometrics II Memorial University of Newfoundland Panel Data Models Adapted from Vera Tabakova s notes 15.1 Grunfeld s Investment Data 15.2 Sets of Regression Equations 15.3 Seemingly Unrelated

More information

Instrumental Variables

Instrumental Variables Instrumental Variables Department of Economics University of Wisconsin-Madison September 27, 2016 Treatment Effects Throughout the course we will focus on the Treatment Effect Model For now take that to

More information

Chapter 8 Heteroskedasticity

Chapter 8 Heteroskedasticity Chapter 8 Walter R. Paczkowski Rutgers University Page 1 Chapter Contents 8.1 The Nature of 8. Detecting 8.3 -Consistent Standard Errors 8.4 Generalized Least Squares: Known Form of Variance 8.5 Generalized

More information

Christopher Dougherty London School of Economics and Political Science

Christopher Dougherty London School of Economics and Political Science Introduction to Econometrics FIFTH EDITION Christopher Dougherty London School of Economics and Political Science OXFORD UNIVERSITY PRESS Contents INTRODU CTION 1 Why study econometrics? 1 Aim of this

More information

Topic 7: Heteroskedasticity

Topic 7: Heteroskedasticity Topic 7: Heteroskedasticity Advanced Econometrics (I Dong Chen School of Economics, Peking University Introduction If the disturbance variance is not constant across observations, the regression is heteroskedastic

More information

ECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 48

ECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 48 ECON2228 Notes 10 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 10 2014 2015 1 / 48 Serial correlation and heteroskedasticity in time series regressions Chapter 12:

More information

Linear Models in Econometrics

Linear Models in Econometrics Linear Models in Econometrics Nicky Grant At the most fundamental level econometrics is the development of statistical techniques suited primarily to answering economic questions and testing economic theories.

More information

Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data

Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data Panel data Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data - possible to control for some unobserved heterogeneity - possible

More information

Unit Root and Cointegration

Unit Root and Cointegration Unit Root and Cointegration Carlos Hurtado Department of Economics University of Illinois at Urbana-Champaign hrtdmrt@illinois.edu Oct 7th, 016 C. Hurtado (UIUC - Economics) Applied Econometrics On the

More information

Reliability of inference (1 of 2 lectures)

Reliability of inference (1 of 2 lectures) Reliability of inference (1 of 2 lectures) Ragnar Nymoen University of Oslo 5 March 2013 1 / 19 This lecture (#13 and 14): I The optimality of the OLS estimators and tests depend on the assumptions of

More information

EC402 - Problem Set 3

EC402 - Problem Set 3 EC402 - Problem Set 3 Konrad Burchardi 11th of February 2009 Introduction Today we will - briefly talk about the Conditional Expectation Function and - lengthily talk about Fixed Effects: How do we calculate

More information

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 7 Introduction to Specification Testing in Dynamic Econometric Models In this lecture I want to briefly describe

More information

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han Econometrics Honor s Exam Review Session Spring 2012 Eunice Han Topics 1. OLS The Assumptions Omitted Variable Bias Conditional Mean Independence Hypothesis Testing and Confidence Intervals Homoskedasticity

More information

1 Correlation between an independent variable and the error

1 Correlation between an independent variable and the error Chapter 7 outline, Econometrics Instrumental variables and model estimation 1 Correlation between an independent variable and the error Recall that one of the assumptions that we make when proving the

More information

Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis. 17th Class 7/1/10

Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis. 17th Class 7/1/10 Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis 17th Class 7/1/10 The only function of economic forecasting is to make astrology look respectable. --John Kenneth Galbraith show

More information

Linear Regression with Time Series Data

Linear Regression with Time Series Data u n i v e r s i t y o f c o p e n h a g e n d e p a r t m e n t o f e c o n o m i c s Econometrics II Linear Regression with Time Series Data Morten Nyboe Tabor u n i v e r s i t y o f c o p e n h a g

More information

Econ 510 B. Brown Spring 2014 Final Exam Answers

Econ 510 B. Brown Spring 2014 Final Exam Answers Econ 510 B. Brown Spring 2014 Final Exam Answers Answer five of the following questions. You must answer question 7. The question are weighted equally. You have 2.5 hours. You may use a calculator. Brevity

More information

08 Endogenous Right-Hand-Side Variables. Andrius Buteikis,

08 Endogenous Right-Hand-Side Variables. Andrius Buteikis, 08 Endogenous Right-Hand-Side Variables Andrius Buteikis, andrius.buteikis@mif.vu.lt http://web.vu.lt/mif/a.buteikis/ Introduction Consider a simple regression model: Y t = α + βx t + u t Under the classical

More information

Graduate Econometrics Lecture 4: Heteroskedasticity

Graduate Econometrics Lecture 4: Heteroskedasticity Graduate Econometrics Lecture 4: Heteroskedasticity Department of Economics University of Gothenburg November 30, 2014 1/43 and Autocorrelation Consequences for OLS Estimator Begin from the linear model

More information

CHAPTER 6: SPECIFICATION VARIABLES

CHAPTER 6: SPECIFICATION VARIABLES Recall, we had the following six assumptions required for the Gauss-Markov Theorem: 1. The regression model is linear, correctly specified, and has an additive error term. 2. The error term has a zero

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the

More information

ACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H.

ACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H. ACE 564 Spring 2006 Lecture 8 Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information by Professor Scott H. Irwin Readings: Griffiths, Hill and Judge. "Collinear Economic Variables,

More information

ECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 54

ECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 54 ECON2228 Notes 10 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 10 2014 2015 1 / 54 erial correlation and heteroskedasticity in time series regressions Chapter 12:

More information

Chapter 2: simple regression model

Chapter 2: simple regression model Chapter 2: simple regression model Goal: understand how to estimate and more importantly interpret the simple regression Reading: chapter 2 of the textbook Advice: this chapter is foundation of econometrics.

More information

Wooldridge, Introductory Econometrics, 2d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of

Wooldridge, Introductory Econometrics, 2d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of Wooldridge, Introductory Econometrics, d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of homoskedasticity of the regression error term: that its

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS Page 1 MSR = Mean Regression Sum of Squares MSE = Mean Squared Error RSS = Regression Sum of Squares SSE = Sum of Squared Errors/Residuals α = Level

More information

1. How can you tell if there is serial correlation? 2. AR to model serial correlation. 3. Ignoring serial correlation. 4. GLS. 5. Projects.

1. How can you tell if there is serial correlation? 2. AR to model serial correlation. 3. Ignoring serial correlation. 4. GLS. 5. Projects. 1. How can you tell if there is serial correlation? 2. AR to model serial correlation. 3. Ignoring serial correlation. 4. GLS. 5. Projects. 1) Identifying serial correlation. Plot Y t versus Y t 1. See

More information

STOCKHOLM UNIVERSITY Department of Economics Course name: Empirical Methods Course code: EC40 Examiner: Lena Nekby Number of credits: 7,5 credits Date of exam: Friday, June 5, 009 Examination time: 3 hours

More information

Heteroskedasticity and Autocorrelation

Heteroskedasticity and Autocorrelation Lesson 7 Heteroskedasticity and Autocorrelation Pilar González and Susan Orbe Dpt. Applied Economics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 7. Heteroskedasticity

More information

Motivation for multiple regression

Motivation for multiple regression Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope

More information

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication G. S. Maddala Kajal Lahiri WILEY A John Wiley and Sons, Ltd., Publication TEMT Foreword Preface to the Fourth Edition xvii xix Part I Introduction and the Linear Regression Model 1 CHAPTER 1 What is Econometrics?

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple

More information

1 Regression with Time Series Variables

1 Regression with Time Series Variables 1 Regression with Time Series Variables With time series regression, Y might not only depend on X, but also lags of Y and lags of X Autoregressive Distributed lag (or ADL(p; q)) model has these features:

More information

ECON 4230 Intermediate Econometric Theory Exam

ECON 4230 Intermediate Econometric Theory Exam ECON 4230 Intermediate Econometric Theory Exam Multiple Choice (20 pts). Circle the best answer. 1. The Classical assumption of mean zero errors is satisfied if the regression model a) is linear in the

More information

The Simple Linear Regression Model

The Simple Linear Regression Model The Simple Linear Regression Model Lesson 3 Ryan Safner 1 1 Department of Economics Hood College ECON 480 - Econometrics Fall 2017 Ryan Safner (Hood College) ECON 480 - Lesson 3 Fall 2017 1 / 77 Bivariate

More information

Heteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance.

Heteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance. Heteroskedasticity y i = β + β x i + β x i +... + β k x ki + e i where E(e i ) σ, non-constant variance. Common problem with samples over individuals. ê i e ˆi x k x k AREC-ECON 535 Lec F Suppose y i =

More information

Environmental Econometrics

Environmental Econometrics Environmental Econometrics Jérôme Adda j.adda@ucl.ac.uk Office # 203 EEC. I Syllabus Course Description: This course is an introductory econometrics course. There will be 2 hours of lectures per week and

More information

Økonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning

Økonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning Økonomisk Kandidateksamen 2004 (I) Econometrics 2 Rettevejledning This is a closed-book exam (uden hjælpemidler). Answer all questions! The group of questions 1 to 4 have equal weight. Within each group,

More information

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47 ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with

More information

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity LECTURE 10 Introduction to Econometrics Multicollinearity & Heteroskedasticity November 22, 2016 1 / 23 ON PREVIOUS LECTURES We discussed the specification of a regression equation Specification consists

More information

Introduction to Econometrics

Introduction to Econometrics Introduction to Econometrics T H I R D E D I T I O N Global Edition James H. Stock Harvard University Mark W. Watson Princeton University Boston Columbus Indianapolis New York San Francisco Upper Saddle

More information

Section 2 NABE ASTEF 65

Section 2 NABE ASTEF 65 Section 2 NABE ASTEF 65 Econometric (Structural) Models 66 67 The Multiple Regression Model 68 69 Assumptions 70 Components of Model Endogenous variables -- Dependent variables, values of which are determined

More information

Econometrics Homework 4 Solutions

Econometrics Homework 4 Solutions Econometrics Homework 4 Solutions Question 1 (a) General sources of problem: measurement error in regressors, omitted variables that are correlated to the regressors, and simultaneous equation (reverse

More information

Covers Chapter 10-12, some of 16, some of 18 in Wooldridge. Regression Analysis with Time Series Data

Covers Chapter 10-12, some of 16, some of 18 in Wooldridge. Regression Analysis with Time Series Data Covers Chapter 10-12, some of 16, some of 18 in Wooldridge Regression Analysis with Time Series Data Obviously time series data different from cross section in terms of source of variation in x and y temporal

More information

ECONOMETRICS HONOR S EXAM REVIEW SESSION

ECONOMETRICS HONOR S EXAM REVIEW SESSION ECONOMETRICS HONOR S EXAM REVIEW SESSION Eunice Han ehan@fas.harvard.edu March 26 th, 2013 Harvard University Information 2 Exam: April 3 rd 3-6pm @ Emerson 105 Bring a calculator and extra pens. Notes

More information

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Estimation - Theory Department of Economics University of Gothenburg December 4, 2014 1/28 Why IV estimation? So far, in OLS, we assumed independence.

More information

FinQuiz Notes

FinQuiz Notes Reading 10 Multiple Regression and Issues in Regression Analysis 2. MULTIPLE LINEAR REGRESSION Multiple linear regression is a method used to model the linear relationship between a dependent variable

More information

Multiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C =

Multiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C = Economics 130 Lecture 6 Midterm Review Next Steps for the Class Multiple Regression Review & Issues Model Specification Issues Launching the Projects!!!!! Midterm results: AVG = 26.5 (88%) A = 27+ B =

More information

8. Instrumental variables regression

8. Instrumental variables regression 8. Instrumental variables regression Recall: In Section 5 we analyzed five sources of estimation bias arising because the regressor is correlated with the error term Violation of the first OLS assumption

More information

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables.

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables. Regression Analysis BUS 735: Business Decision Making and Research 1 Goals of this section Specific goals Learn how to detect relationships between ordinal and categorical variables. Learn how to estimate

More information

GMM, HAC estimators, & Standard Errors for Business Cycle Statistics

GMM, HAC estimators, & Standard Errors for Business Cycle Statistics GMM, HAC estimators, & Standard Errors for Business Cycle Statistics Wouter J. Den Haan London School of Economics c Wouter J. Den Haan Overview Generic GMM problem Estimation Heteroskedastic and Autocorrelation

More information

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors: Wooldridge, Introductory Econometrics, d ed. Chapter 3: Multiple regression analysis: Estimation In multiple regression analysis, we extend the simple (two-variable) regression model to consider the possibility

More information

Introductory Econometrics

Introductory Econometrics Introductory Econometrics Violation of basic assumptions Heteroskedasticity Barbara Pertold-Gebicka CERGE-EI 16 November 010 OLS assumptions 1. Disturbances are random variables drawn from a normal distribution.

More information

Econometrics Homework 1

Econometrics Homework 1 Econometrics Homework Due Date: March, 24. by This problem set includes questions for Lecture -4 covered before midterm exam. Question Let z be a random column vector of size 3 : z = @ (a) Write out z

More information

INTRODUCTORY REGRESSION ANALYSIS

INTRODUCTORY REGRESSION ANALYSIS ;»»>? INTRODUCTORY REGRESSION ANALYSIS With Computer Application for Business and Economics Allen Webster Routledge Taylor & Francis Croup NEW YORK AND LONDON TABLE OF CONTENT IN DETAIL INTRODUCTORY REGRESSION

More information

Course information EC2020 Elements of econometrics

Course information EC2020 Elements of econometrics Course information 2015 16 EC2020 Elements of econometrics Econometrics is the application of statistical methods to the quantification and critical assessment of hypothetical economic relationships using

More information

Auto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e.,

Auto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e., 1 Motivation Auto correlation 2 Autocorrelation occurs when what happens today has an impact on what happens tomorrow, and perhaps further into the future This is a phenomena mainly found in time-series

More information