Extensions to the Basic Framework I
|
|
- Marianna Warren
- 5 years ago
- Views:
Transcription
1 Topic 5 Extensions to the Basic Framework I ARE/ECN 240 A Graduate Econometrics Professor: Òscar Jordà
2 Outline of this topic Heteroskedasticity: reminder of OLS results and White (1980) corrected standard errors, derivation of GLS/GMM results. Testing Autocorrelation: brief introduction to time series, derivation of AR(1) GLS correction, Newey-West standard errors 2
3 Recall: Asymptotic Distribution of the OLS Estimator t under Homoscedasticity it Recall: ^ =(X 0 X) 1 X 0 ² We know: Where: p n( ^ ) = μ X 0 X n plim X0 X n By Slutsky s theorem 1 μ X 0 ² p n p! Q; and X0 ² d p! N(0; ) n =V (X 0 ²)=E E(X 0 ²² 0 XjX) = ¾ 2 E((X 0 X)) = ¾ 2 Q p n( ^ ) d! N(0;¾ 2 (X 0 X) 1 ) 3
4 Remarks: 4 To derive this fundamental result we have assumed: We have an i.i.d. sample. However, we will see that this can be relaxed in several ways: 1. To data that is not identically distributed (but independent): useful when heteroskedasticity is suspected 2. To data that is identical but not independent (useful for time series data but we need to limit the amount of dependence) 3. To data that is neither identical nor independent (which is the safest assumption with time series data We used the assumption that =V(X 0 ²)=E E(X 0 ²² 0 XjX) = ¾ 2 E((X 0 X)) = ¾ 2 Q
5 A More Formal Statement for Asymptotic Normality with Ht Heteroskedastic kd ti Errors Assume: 1. X contains an intercept 2. E(y 2 ) < E(x 2 j ) < 1 for j =1; :::; K Q = E(X 0 X)isinvertible Then as 4 4 E(² 4 i ) < 1 and E(x 4 j) < 1 for j =1; :::; K fy i ;x 1i ; :::; x Ki g n i=1 is i:i:d: n!1 p n( ^ ) d! N(0;V)whereV = Q 1 ÐQ 1 5
6 Specifics Using slightly more detailed notation than before, notice that: p n( ^ ) = Ã 1 n! 1 Ã nx XiX 0 1 i p n i=1 Notice that tx i isa1 K vector for i = 1,,n observations! nx Xi² 0 i We had established that under the assumptions, a LLN applies to 1 X n X 0 X p n ix i! E(X 0 i X i ) = Q i=1 Next we use a CLT and Slutsky s i=1 6
7 Specifics (cont.) Under the assumptions, the CLT means that 1 p n nx Xi² 0 d i! N(0; Ð) i=1 where: Ð=E(XiX 0 i ² 2 i ). Under homoskedasticity, we saw that t when X d are independent, d then X i and ² i Ð=E(X 0 ix i )E(² 2 i )=Q¾ 2 V = Q 1 ÐQ 1 = Q 1 Q¾ 2 Q 1 = ¾ 2 Q 1 But in general, this will not hold 7
8 Heteroskedasticity Instead, White (1980) proposed his very famous estimator nx X 0 ^Ð = 1 n n i=1 ix i^² 2 i So that the estimate of the variance-covariance matrix of the ^ is the sandwich estimator ^V = ^Q 1 ^Ð ^Q 1 Formally, one would still need to show that Q ^Q! p Q; Ð ^Ð! p Ð; and hence V ^V! p V 8
9 Heteroskedasticity: Where Does It Come From? Think of the relation between Food Expenditures and dincome. Wealthier individuals do not necessarily consume more calories, on average, than poorer folk. But they can certainly spend more on how they obtain those calories: think of a Big Mac versus Filet Mignon. However, not all wealthy individuals develop a taste for expensive food (think of a computer programmer for Google recently graduated from college and pasta with Ragu sauce and pizza) 9
10 A Picture Food Expenditures Total Income 10
11 Also, recall our own data This assumption is often violated. However, it is easy to relax. It will not affect parameter estimates but it will affect how their standard errors are calculated (i.e., the efficiency of the estimator).. sum testscr if str <= 17 Variable Obs Mean Std. Dev. Min Max testscr sum testscr if str >17 & str <=20 Variable Obs Mean Std. Dev. Min Max testscr sum testscr if str >= 22.8 Variable Obs Mean Std. Dev. Min Max testscr sum testscr if str < 22.8 & str >= 19.8 Variable Obs Mean Std. Dev. Min Max testscr
12 Recap White s (1980) correction gives appropriate standard errors but, is there a way to improve efficiency? i In order to discuss how to improve the efficiency of regression estimators, it is helpful to remember the theory of extremum estimation. Remember that extremum estimator theory encompasses the majority of estimation procedures that are commonly used in econometrics (OLS, MLE, and GMM) Here are some basic results we have already seen and which help motivate GLS 12
13 Basic Theorem for Extremum Eti Estimatorst Consider extremum estimator ^μ obtained from an objective function ^Q n (μ) such that: max ^ ^ The Method of Moments: ^Q n (μ) = " 1 n 0 nx g(y i ;X i ; )# ^W i=1 " 1 n μ2 Q n(μ) =μ # nx g(y i ;X i ; ) where E(g(y i ;X i ; )) = 0 and is a weighting matrix. For example, in deriving the estimator for linear regression we used E(X 0 ²) = E(X 0 (y X )) = 0 hence g(:) ) = X 0 (y X ) and W ^W = I ^W i=1 13
14 Basic Asymptotic ti Normality Theorem ^Q n (μ) =^μ; ^μ p! μ 0,and(i)μ 0 2 interior( ); Suppose: max μ2 (ii) ^Q n (μ) is twice continuously differentiable in a p neighborhood of ; ;(iii) nr ^Qn (μ 0 )! d N(0; Ð) ; (iv) there is H(μ) that is continuous at and sup μ2n jjr μμ ^Q n (μ) H(μ)jj p! 0 with (v) H = H(μ 0 ) nonsingular. Then: p n(^μ μ0 )! d N(0;H 1 ÐH 1 ) 14
15 Sketch of the Proof Conditions (i) (iii) imply that r ^Q n (^μ) =0w.p.1 Expand this condition around and use mean- value theorem We know: Since r ^Q n (^μ) =r ^Q n (μ 0 )+ ^H( μ)(^μ ¹ μ 0 ) ^ ^ ^ rq n (μ) = 0 and by (iii) p nrq n (μ 0 )! d N(0; Ð) ¹μ 2 [^μ; μ 0 ]and^μ p! μ 0 then ¹ μ p! μ 0 By (iv) we can show that (after a few intermediate steps): jj ^H( ¹ p μ) H(μ 0 )jj! 0 p Hence: n(^μ ( μ0 0) ) = ^H( μ) ¹ ) 1 rq ^Q (μ! H(μ 1 ^Q n 0 ) ( 0 ) rq n (μ 0 ) 15
16 An Application to GMM for Linear Regression: A Preview of GLS Use the moment condition: E(X 0 ")=0 Hence: g(yi ;X i ; ) = 1 n nx Xi" 0 i = 1 n i=1 First order conditions: nx Xi(y 0 i X i ) i=1 bq n ( ) = g(y i ;X i ; ) 0 c Wg(yi ;X i ; ) G(y i ;X i ; ^ ) 0 Wg(yi c ;X i ; ^ ) =0 Ã! Ã! 1 nx X 0 1 nx n ix i cw Xi 0 ^" i =0 n i=1 i=1 16
17 GMM (cont.) Take a mean-value expansion: We know p n G 0 Wg( 0) = G 0 W p 1 n Hence: 0= G( 0) 0 cwg( 0) G( 0) 0 cwg( ) ³ b 0 p n ³ b 0 p ³ n b 0 G 0 p 0! G W G n G W g( 0) nx Xi" 0 i! N(0;G 0 W ÐWG); Ð = E(XiX 0 i " 2 i ) i=1! N(0; (G 0 WG) 1 G 0 W ÐWG(G 0 WG) 1 ) 17
18 Optimal Weighting Recall: ³ p n ³ b 0! N(0; (G 0 WG) 1 G 0 W ÐWG(G 0 WG) 1 ) Choose: W =Ð 1. Then, things simplify a lot! Recall: p ³ n b 0! N(0; (G 0 Ð 1 G) 1 ) Ã G = ^G = 1 n! nx XiX 0 i i=1 And with White s estimator: nx X 0 bð = 1 n n i=1 ix i^" 2 i 18
19 Optimal Weighting (cont.) Or, if homoscedasticty holds, then: bð = b¾2 n nx XiX 0 i i=1 And hence, we recover the usual OLS result 0 Ã n! 1 1 p ³ n b 0! b¾ 2 1 nx X 0 n ix i A GLS consists in choosing the weighting matrix optimally when we cannot make these simplifying assumptions i=1 19
20 Remarks In general settings, the optimal weighting involves parameter estimates. No problem, since non-optimaloptimal weights still produces consistent parameter estimates, we can go in stages. However, under homoscedasticity, this is not a problem since the optimal weights do not involve parameter estimates and hence we recover the usual OLS formulas. Notice that for the more general case, the GMM returns a GLS procedure where observations are weighed by the inverse of their variances why does this make sense? Intuition? i Be careful keeping track of population versus sample quantities in the derivations 20
21 A Preliminary Example Suppose you want to simulate an n-dimensional, normally distributed random vector with mean and variance using a random number generator that produces n-dimensional N(0, I) random variates. How would you do it? Any symmetric, positive-definite matrix A with real entries can be decomposed as: where L is a lower triangular matrix with strictly positive diagonal entries. This decomposition is unique although there are other square-root matrices if lower-triangularity is not required of L 21
22 Example (cont.) Let z» N(0;I) and we want to generate μμ μ y» N ; First notice that μ μ μ = Hence μ μ μμ μ y = z +» N ;
23 Sample GAUSS Program z = rndn(100000,2); // generate obs of 2 N(0,1) rvs mean = 1~2; // Desired mean var = (1 1)~(1 2); // desired variance L = chol(var); // Choleski decompostion of var // Notice GAUSS produces an UPPER triangular matrix // Generate y = z*l + mean; " "; " Verify the mean of y is correct"; " "; "Original Mean"$~"Sample Mean"; mean'~meanc(y); " "; " Verify the Co-variance Matrix is correct"; " "; "Desired Co-variance matrix"; var; "Sample Co-variance matrix"; vcx(y); 23
24 Output» run C:\Docs\teaching\240A\STATA\corrnormal.g; Verify the mean of y is correct Original Mean Sample Mean Verify the Co-variance Matrix is correct Desired Co-variance matrix Sample Co-variance matrix
25 Why Does this Matter? Suppose you could calculate the variancecovariance matrix for the residuals of each observation in a linear regression model: Ð n n Because it is symmetric and positive definite it has an inverse that is also symmetric and positive definite and hence admits a Choleski decomposition, for example: Ð 1 = PP 0 Pre-multiplying the regression by P on both sides P 0 y = P 0 X + P 0 ² 25
26 The GLS Estimator Now, using the OLS formulas on the modified regression P 0 y = P 0 X + P 0 ² we have: ^ GLS =(X 0 PP 0 X) 1 (X 0 PP 0 y)=(x 0 Ð 1 X) 1 (X 0 Ð 1 y) u Notice then that E(P 0 ² 0 P ) = P 0 E(²² 0 )P = P 0 Ð 1 1 P = :::=? I E u0 ) (u ²0 P The residuals u are homoscedastic! 26
27 The GLS Estimator (cont.) The variance of the OLS estimator is, using the usual formula (you should be able to show this): V ( ^ GLS ) = (X 0 Ð 1 X) 1 In fact, the GLS estimator should not be a surprise. We had already derived it using GMM using as the optimal weighting matrix ^W^W Ð 1 27
28 Practical Observations Notice that I have assumed that Ð is available n n but in practice, it would have to be estimated. Moreover, clearly we need to impose some structure since we would be estimating n(n 1)/2 terms with n observations, clearly infeasible. But even if we had all these terms available, we would be inverting a fairly large matrix, which could be computationally problematic. So next we look at simple structures that we can impose on Ð to obtain reasonable estimates. n n 28
29 Feasible GLS Let s think back at the food expenditures example, where residuals across individuals are probably not correlated but they do vary with the income of the individual, that is: E(² 2 i jx i )=¾ 2 (X i ) For example, we could conjecture E(² 2 i jx i )=exp(x i ) so as to ensure positivity. Using method of moments arguments and using the residuals from a first stage 29
30 Feasible GLS (cont.) we could consider fitting the auxiliary regression: 30 log ^² i = X i + v i The v are funny looking residuals but centered at zero, which is what matters to obtain consistent estimates of the These could be used to construct: ^w i =(exp(x i^ )) 1 and hence the weighting g matrix: 0 1 ^w 1 0 ::: 0 0 ^w ^W = 2 ::: 0 W B C ^ FGLS =(X 0 ^W X) 1 (X 0 ^W A 0 0 ::: ^w n
31 Remarks Reasons FGLS may not work in practice: The auxiliary regression, sometimes called the skedastic regression, may be misspecified. The OLS estimator is more robust as it can be justified on projection arguments. The FGLS estimator relies on the correct specification of the conditional mean. The skedastic regression lends itself to test for heteroskedasticity as we could use a typical F-test to assess the null that H 0 : 2 = :::= K =0(i.e. excluding the constant). 31
32 Recap So far I have focused on situations where I essentially rule out covariation across units, i.e., heteroskedasticity is all about the terms in the diagonal of the covariance matrix not been equal. Hence there are n different terms so with a sample of size n, in principle we cannot estimate consistently any of them. But with FGLS, if we specify a conditional o variance a function, we can reduce dimensionality considerably. Think of White s (1980) correction as a nonparametric correction for heteroskedasticity: the correction does not require consistent estimates but by the same token, it is not as efficient as when one can come up with a valid FGLS specification. 32
33 Serial Correlation There is a natural environment where the residuals may be correlated across units and that is in time series contexts. So far we have assumed i.i.d. samples, time series clearly violates the independence assumption: GDP today depends on GDP yesterday We will investigate time series in more detail. Here we focus on serial correlation in the residuals, i.e., if the conditional mean were well specified, the residuals would be white noise (uncorrelated over time). However, some correlation in the residuals may still generate consistent estimates of parameters of interest we just have to fix the standard errors! 33
34 Dependency What is the magic of the i.i.d. sample? Recall, a random sample fz 1 ; :::; z n g n i=1 has a joint density f(z 1 ;:::;z n ; μ) This density, under iid i.i.d., can be factored as: f(z 1 ;:::;z n ; μ) =f(z 1 ; μ):::f(z n ; μ) However, if the data are dependent d f(z 1 ;:::;z t ;:::;z n ; μ) =f(z 1 ; μ)f(z 2 jz 1 ; μ)::: f(z t jz t 1 ;:::;z 1 ; μ):::f(z n jz n 1 ;:::;z 1 ; μ) So we need a way to limit dependence on the past 34
35 AR(1) One way to limit dependency is with the autoregressive model of order 1: AR(1) Let s focus on the residuals of an OLS regression, an AR(1) structure looks like this: u t = ½u t 1 + ² t ;² t» D(0;¾ 2 ); j½j < 1 Remarks: Notice that by recursive substitution: u t = ² t + ½² t 1 + ½ 2 ² t 2 + ::: hence j½j < 1 ensures dependency on the past diesoff quickly 35
36 Remarks (cont.) Notice that: ¾ ¾2 2 u = ¾² 2 + ½ 2 ¾² 2 + ½ 4 ¾² 2 + ::: = ¾2 ² 1 ½ 2 ; j½j < 1 And Cov(u t ;u t j )=E(u t ;u t j )=½ j ¾ u 2 Under the assumption j½j < 1, these infinite sums converge and are independent d of t. When the mean, variance and covariances are independent from t, such a process is known as covariancestationary. 2 36
37 Covariance Residual Matrix Using the formula of the variance and the covariances, we can show that 0 1 ½ ½ 2 ::: ½ n 11 Ð u = ¾2 B ² ½ 1 ½ ::: ½ n 2 B@ C 1 ½ 2... :::. A ½ n 1 ½ n 2 ½ n 3 ::: 1 So although Ð u has in principle n(n-1)/2 elements, they are all determined by two parameters, ¾ 2 ² and ½ Let s see how the FGLS estimator looks like in this case 37
38 FGLS with AR(1) residuals 1. Estimate the regression model as usual. Obtain consistent estimates of ^ and more importantly ^u t 2. Estimate the following regression by OLS, for example: ^u t = ½^u t 1 + ² t 3. Given ½ ^½ and ¾ ^¾ 2 ² then construct 0 1 ^½ ^½ 2 ::: ^½ n 1 1 ^¾ 2 ^½ ^½ ^½ n 2 ^W 1 = ^Ð u = ¾ B ² ½ 1 ½ ::: ½ B@ C 1 ^½ 2... :::. A ½ ^½ n 1 ½ ^½ n 2 ½ ^½ n 3 ::: 1 38
39 A More Direct Approach However, notice that we can correct serial correlation in one fell swoop. Notice: hence u t = y t X t u t 1 = y t 1 X t 1 u t = ½u t 1 + ² t becomes: y t = ½y t 1 + X t ½X t 1 + ² t The model can be estimated with NL maximization (e.g. NLS, GMM) and a test of the null H 0 : ½ =0 as a test of serial correlation. 39
40 Higher Order Serial Correlation A more general model would allow for more serial dependence, such as u t = ½ 1 u t 1 + ½ 2 u t 2 + ::: + ½ p u t p + ² t from which h we could do as we did with the AR(1) model. Is there something similar to White s (1980) correction for heteroskedasticity for serial correlation? Yes, it is called HAC covariance matrix estimators, the most common one is the Newey-West estimator but it is beyond the scope of this class. 40
41 A STATA example Suppose I want to estimate Okun s Law, which says: Y Y = c k U where Y is real GDP and U is the unemployment rate. I prefer to look at year-on-year changes to smooth out the data a bit. The file okun.do contains a brief STATA file to estimate this regression and adjusts standard errors with Newey West 41
42 Basic Commands in Okun.do log using okun.log, replace insheet using "C:\Docs\teaching\240A\STATA\okun.csv", comma gen qtr = q(1948q1) + _n-1 tsset qtr,q gen dy = 100*(gdp - l4.gdp)/l4.gdp gen du = u - l4.u reg dy du // OLS reg dy du, vce(robust) // OLS with White SE newey dy du, lag(4) log close exit 42
43 Output. reg dy du // OLS Source SS df MS Number of obs = F( 1, 241) = Model Prob > F = Residual R-squared = Adj R-squared = Total Root MSE = dy Coef. Std. Err. t P> t [95% Conf. Interval] du _cons reg dy du, vce(robust) // OLS with White corrected SE Regression with Newey-West standard errors Number of obs = 243 maximum lag: 4 F( 1, 241) = Prob > F = Newey-West dy Coef. Std. Err. t P> t [95% Conf. Interval] du _cons Linear regression Number of obs = 243 F( 1, 241) = Prob > F = R-squared = Root MSE = Robust dy Coef. Std. Err. t P> t [95% Conf. Interval] du _cons
Multivariate Regression: Part I
Topic 1 Multivariate Regression: Part I ARE/ECN 240 A Graduate Econometrics Professor: Òscar Jordà Outline of this topic Statement of the objective: we want to explain the behavior of one variable as a
More informationGraduate Econometrics Lecture 4: Heteroskedasticity
Graduate Econometrics Lecture 4: Heteroskedasticity Department of Economics University of Gothenburg November 30, 2014 1/43 and Autocorrelation Consequences for OLS Estimator Begin from the linear model
More informationExtensions to the Basic Framework II
Topic 7 Extensions to the Basic Framework II ARE/ECN 240 A Graduate Econometrics Professor: Òscar Jordà Outline of this topic Nonlinear regression Limited Dependent Variable regression Applications of
More informationEconometrics Midterm Examination Answers
Econometrics Midterm Examination Answers March 4, 204. Question (35 points) Answer the following short questions. (i) De ne what is an unbiased estimator. Show that X is an unbiased estimator for E(X i
More informationLab 07 Introduction to Econometrics
Lab 07 Introduction to Econometrics Learning outcomes for this lab: Introduce the different typologies of data and the econometric models that can be used Understand the rationale behind econometrics Understand
More informationEconometrics - 30C00200
Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business
More informationRegression with a Single Regressor: Hypothesis Tests and Confidence Intervals
Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression
More informationApplied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple
More informationthe error term could vary over the observations, in ways that are related
Heteroskedasticity We now consider the implications of relaxing the assumption that the conditional variance Var(u i x i ) = σ 2 is common to all observations i = 1,..., n In many applications, we may
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationHeteroskedasticity. Occurs when the Gauss Markov assumption that the residual variance is constant across all observations in the data set
Heteroskedasticity Occurs when the Gauss Markov assumption that the residual variance is constant across all observations in the data set Heteroskedasticity Occurs when the Gauss Markov assumption that
More informationECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests
ECON4150 - Introductory Econometrics Lecture 5: OLS with One Regressor: Hypothesis Tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 5 Lecture outline 2 Testing Hypotheses about one
More informationECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 48
ECON2228 Notes 10 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 10 2014 2015 1 / 48 Serial correlation and heteroskedasticity in time series regressions Chapter 12:
More informationEconometrics. 9) Heteroscedasticity and autocorrelation
30C00200 Econometrics 9) Heteroscedasticity and autocorrelation Timo Kuosmanen Professor, Ph.D. http://nomepre.net/index.php/timokuosmanen Today s topics Heteroscedasticity Possible causes Testing for
More informationSection I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -
First Exam: Economics 388, Econometrics Spring 006 in R. Butler s class YOUR NAME: Section I (30 points) Questions 1-10 (3 points each) Section II (40 points) Questions 11-15 (10 points each) Section III
More informationHeteroskedasticity. We now consider the implications of relaxing the assumption that the conditional
Heteroskedasticity We now consider the implications of relaxing the assumption that the conditional variance V (u i x i ) = σ 2 is common to all observations i = 1,..., In many applications, we may suspect
More informationAutocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time
Autocorrelation Given the model Y t = b 0 + b 1 X t + u t Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time This could be caused
More informationMultiple Regression Analysis
Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,
More informationHeteroskedasticity. (In practice this means the spread of observations around any given value of X will not now be constant)
Heteroskedasticity Occurs when the Gauss Markov assumption that the residual variance is constant across all observations in the data set so that E(u 2 i /X i ) σ 2 i (In practice this means the spread
More informationECON2228 Notes 10. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 54
ECON2228 Notes 10 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 10 2014 2015 1 / 54 erial correlation and heteroskedasticity in time series regressions Chapter 12:
More informationChapter 7. Hypothesis Tests and Confidence Intervals in Multiple Regression
Chapter 7 Hypothesis Tests and Confidence Intervals in Multiple Regression Outline 1. Hypothesis tests and confidence intervals for a single coefficie. Joint hypothesis tests on multiple coefficients 3.
More informationStatistical Inference with Regression Analysis
Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Steven Buck Lecture #13 Statistical Inference with Regression Analysis Next we turn to calculating confidence intervals and hypothesis testing
More informationEconometrics Homework 1
Econometrics Homework Due Date: March, 24. by This problem set includes questions for Lecture -4 covered before midterm exam. Question Let z be a random column vector of size 3 : z = @ (a) Write out z
More informationAUTOCORRELATION. Phung Thanh Binh
AUTOCORRELATION Phung Thanh Binh OUTLINE Time series Gauss-Markov conditions The nature of autocorrelation Causes of autocorrelation Consequences of autocorrelation Detecting autocorrelation Remedial measures
More information9) Time series econometrics
30C00200 Econometrics 9) Time series econometrics Timo Kuosmanen Professor Management Science http://nomepre.net/index.php/timokuosmanen 1 Macroeconomic data: GDP Inflation rate Examples of time series
More informationNonlinear Regression Functions
Nonlinear Regression Functions (SW Chapter 8) Outline 1. Nonlinear regression functions general comments 2. Nonlinear functions of one variable 3. Nonlinear functions of two variables: interactions 4.
More informationWeek 11 Heteroskedasticity and Autocorrelation
Week 11 Heteroskedasticity and Autocorrelation İnsan TUNALI Econ 511 Econometrics I Koç University 27 November 2018 Lecture outline 1. OLS and assumptions on V(ε) 2. Violations of V(ε) σ 2 I: 1. Heteroskedasticity
More informationAnswer all questions from part I. Answer two question from part II.a, and one question from part II.b.
B203: Quantitative Methods Answer all questions from part I. Answer two question from part II.a, and one question from part II.b. Part I: Compulsory Questions. Answer all questions. Each question carries
More information10) Time series econometrics
30C00200 Econometrics 10) Time series econometrics Timo Kuosmanen Professor, Ph.D. 1 Topics today Static vs. dynamic time series model Suprious regression Stationary and nonstationary time series Unit
More informationECON Introductory Econometrics. Lecture 4: Linear Regression with One Regressor
ECON4150 - Introductory Econometrics Lecture 4: Linear Regression with One Regressor Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 4 Lecture outline 2 The OLS estimators The effect of
More information13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity
Outline: Further Issues in Using OLS with Time Series Data 13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process I. Stationary and Weakly Dependent Time Series III. Highly Persistent
More informationIntroduction to Econometrics. Multiple Regression (2016/2017)
Introduction to Econometrics STAT-S-301 Multiple Regression (016/017) Lecturer: Yves Dominicy Teaching Assistant: Elise Petit 1 OLS estimate of the TS/STR relation: OLS estimate of the Test Score/STR relation:
More information1 Outline. 1. Motivation. 2. SUR model. 3. Simultaneous equations. 4. Estimation
1 Outline. 1. Motivation 2. SUR model 3. Simultaneous equations 4. Estimation 2 Motivation. In this chapter, we will study simultaneous systems of econometric equations. Systems of simultaneous equations
More informationECON3150/4150 Spring 2015
ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2
More information7 Introduction to Time Series
Econ 495 - Econometric Review 1 7 Introduction to Time Series 7.1 Time Series vs. Cross-Sectional Data Time series data has a temporal ordering, unlike cross-section data, we will need to changes some
More informationEcon 423 Lecture Notes
Econ 423 Lecture Notes (hese notes are modified versions of lecture notes provided by Stock and Watson, 2007. hey are for instructional purposes only and are not to be distributed outside of the classroom.)
More informationHypothesis Tests and Confidence Intervals. in Multiple Regression
ECON4135, LN6 Hypothesis Tests and Confidence Intervals Outline 1. Why multipple regression? in Multiple Regression (SW Chapter 7) 2. Simpson s paradox (omitted variables bias) 3. Hypothesis tests and
More informationLongitudinal Data Analysis Using Stata Paul D. Allison, Ph.D. Upcoming Seminar: May 18-19, 2017, Chicago, Illinois
Longitudinal Data Analysis Using Stata Paul D. Allison, Ph.D. Upcoming Seminar: May 18-19, 217, Chicago, Illinois Outline 1. Opportunities and challenges of panel data. a. Data requirements b. Control
More informationIntroduction to Econometrics. Multiple Regression
Introduction to Econometrics The statistical analysis of economic (and related) data STATS301 Multiple Regression Titulaire: Christopher Bruffaerts Assistant: Lorenzo Ricci 1 OLS estimate of the TS/STR
More informationEmpirical Macroeconomics
Empirical Macroeconomics Francesco Franco Nova SBE April 5, 2016 Francesco Franco Empirical Macroeconomics 1/39 Growth and Fluctuations Supply and Demand Figure : US dynamics Francesco Franco Empirical
More informationECON2228 Notes 7. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 41
ECON2228 Notes 7 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 6 2014 2015 1 / 41 Chapter 8: Heteroskedasticity In laying out the standard regression model, we made
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More information4 Instrumental Variables Single endogenous variable One continuous instrument. 2
Econ 495 - Econometric Review 1 Contents 4 Instrumental Variables 2 4.1 Single endogenous variable One continuous instrument. 2 4.2 Single endogenous variable more than one continuous instrument..........................
More informationWeek 3: Simple Linear Regression
Week 3: Simple Linear Regression Marcelo Coca Perraillon University of Colorado Anschutz Medical Campus Health Services Research Methods I HSMP 7607 2017 c 2017 PERRAILLON ALL RIGHTS RESERVED 1 Outline
More informationEmpirical Application of Simple Regression (Chapter 2)
Empirical Application of Simple Regression (Chapter 2) 1. The data file is House Data, which can be downloaded from my webpage. 2. Use stata menu File Import Excel Spreadsheet to read the data. Don t forget
More informationEmpirical Macroeconomics
Empirical Macroeconomics Francesco Franco Nova SBE April 21, 2015 Francesco Franco Empirical Macroeconomics 1/33 Growth and Fluctuations Supply and Demand Figure : US dynamics Francesco Franco Empirical
More information7 Introduction to Time Series Time Series vs. Cross-Sectional Data Detrending Time Series... 15
Econ 495 - Econometric Review 1 Contents 7 Introduction to Time Series 3 7.1 Time Series vs. Cross-Sectional Data............ 3 7.2 Detrending Time Series................... 15 7.3 Types of Stochastic
More informationWarwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation
Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation Michele Aquaro University of Warwick This version: July 21, 2016 1 / 31 Reading material Textbook: Introductory
More informationEconometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague
Econometrics Week 4 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 23 Recommended Reading For the today Serial correlation and heteroskedasticity in
More information1 The Multiple Regression Model: Freeing Up the Classical Assumptions
1 The Multiple Regression Model: Freeing Up the Classical Assumptions Some or all of classical assumptions were crucial for many of the derivations of the previous chapters. Derivation of the OLS estimator
More informationDynamic Panels. Chapter Introduction Autoregressive Model
Chapter 11 Dynamic Panels This chapter covers the econometrics methods to estimate dynamic panel data models, and presents examples in Stata to illustrate the use of these procedures. The topics in this
More informationECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors
ECON4150 - Introductory Econometrics Lecture 6: OLS with Multiple Regressors Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 6 Lecture outline 2 Violation of first Least Squares assumption
More informationECONOMICS AND ECONOMIC METHODS PRELIM EXAM Statistics and Econometrics August 2013
ECONOMICS AND ECONOMIC METHODS PRELIM EXAM Statistics and Econometrics August 2013 Instructions: Answer all six (6) questions. Point totals for each question are given in parentheses. The parts within
More informationInstrumental Variables, Simultaneous and Systems of Equations
Chapter 6 Instrumental Variables, Simultaneous and Systems of Equations 61 Instrumental variables In the linear regression model y i = x iβ + ε i (61) we have been assuming that bf x i and ε i are uncorrelated
More informationControl Function and Related Methods: Nonlinear Models
Control Function and Related Methods: Nonlinear Models Jeff Wooldridge Michigan State University Programme Evaluation for Policy Analysis Institute for Fiscal Studies June 2012 1. General Approach 2. Nonlinear
More information11. Further Issues in Using OLS with TS Data
11. Further Issues in Using OLS with TS Data With TS, including lags of the dependent variable often allow us to fit much better the variation in y Exact distribution theory is rarely available in TS applications,
More informationAuto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e.,
1 Motivation Auto correlation 2 Autocorrelation occurs when what happens today has an impact on what happens tomorrow, and perhaps further into the future This is a phenomena mainly found in time-series
More informationRecent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data
Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationECON Introductory Econometrics. Lecture 16: Instrumental variables
ECON4150 - Introductory Econometrics Lecture 16: Instrumental variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 12 Lecture outline 2 OLS assumptions and when they are violated Instrumental
More informationEcon 836 Final Exam. 2 w N 2 u N 2. 2 v N
1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.
More information4 Instrumental Variables Single endogenous variable One continuous instrument. 2
Econ 495 - Econometric Review 1 Contents 4 Instrumental Variables 2 4.1 Single endogenous variable One continuous instrument. 2 4.2 Single endogenous variable more than one continuous instrument..........................
More informationMonday 7 th Febraury 2005
Monday 7 th Febraury 2 Analysis of Pigs data Data: Body weights of 48 pigs at 9 successive follow-up visits. This is an equally spaced data. It is always a good habit to reshape the data, so we can easily
More informationHypothesis Tests and Confidence Intervals in Multiple Regression
Hypothesis Tests and Confidence Intervals in Multiple Regression (SW Chapter 7) Outline 1. Hypothesis tests and confidence intervals for one coefficient. Joint hypothesis tests on multiple coefficients
More informationGLS. Miguel Sarzosa. Econ626: Empirical Microeconomics, Department of Economics University of Maryland
GLS Miguel Sarzosa Department of Economics University of Maryland Econ626: Empirical Microeconomics, 2012 1 When any of the i s fail 2 Feasibility 3 Now we go to Stata! GLS Fixes i s Failure Remember that
More informationEssential of Simple regression
Essential of Simple regression We use simple regression when we are interested in the relationship between two variables (e.g., x is class size, and y is student s GPA). For simplicity we assume the relationship
More informationMeasurement Error. Often a data set will contain imperfect measures of the data we would ideally like.
Measurement Error Often a data set will contain imperfect measures of the data we would ideally like. Aggregate Data: (GDP, Consumption, Investment are only best guesses of theoretical counterparts and
More informationLecture 5. In the last lecture, we covered. This lecture introduces you to
Lecture 5 In the last lecture, we covered. homework 2. The linear regression model (4.) 3. Estimating the coefficients (4.2) This lecture introduces you to. Measures of Fit (4.3) 2. The Least Square Assumptions
More informationMultivariate Time Series
Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form
More informationEcon 2120: Section 2
Econ 2120: Section 2 Part I - Linear Predictor Loose Ends Ashesh Rambachan Fall 2018 Outline Big Picture Matrix Version of the Linear Predictor and Least Squares Fit Linear Predictor Least Squares Omitted
More informationHeteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD
Heteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD Introduction For pedagogical reasons, OLS is presented initially under strong simplifying assumptions. One of these is homoskedastic errors,
More informationECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47
ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with
More informationUnemployment Rate Example
Unemployment Rate Example Find unemployment rates for men and women in your age bracket Go to FRED Categories/Population/Current Population Survey/Unemployment Rate Release Tables/Selected unemployment
More informationInterpreting coefficients for transformed variables
Interpreting coefficients for transformed variables! Recall that when both independent and dependent variables are untransformed, an estimated coefficient represents the change in the dependent variable
More informationECON Introductory Econometrics. Lecture 13: Internal and external validity
ECON4150 - Introductory Econometrics Lecture 13: Internal and external validity Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 9 Lecture outline 2 Definitions of internal and external
More informationProblem Set 1 ANSWERS
Economics 20 Prof. Patricia M. Anderson Problem Set 1 ANSWERS Part I. Multiple Choice Problems 1. If X and Z are two random variables, then E[X-Z] is d. E[X] E[Z] This is just a simple application of one
More informationAutoregressive models with distributed lags (ADL)
Autoregressive models with distributed lags (ADL) It often happens than including the lagged dependent variable in the model results in model which is better fitted and needs less parameters. It can be
More informationECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests
ECON4150 - Introductory Econometrics Lecture 7: OLS with Multiple Regressors Hypotheses tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 7 Lecture outline 2 Hypothesis test for single
More informationEconometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018
Econometrics I KS Module 1: Bivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: March 12, 2018 Alexander Ahammer (JKU) Module 1: Bivariate
More information****Lab 4, Feb 4: EDA and OLS and WLS
****Lab 4, Feb 4: EDA and OLS and WLS ------- log: C:\Documents and Settings\Default\Desktop\LDA\Data\cows_Lab4.log log type: text opened on: 4 Feb 2004, 09:26:19. use use "Z:\LDA\DataLDA\cowsP.dta", clear.
More informationECO375 Tutorial 7 Heteroscedasticity
ECO375 Tutorial 7 Heteroscedasticity Matt Tudball University of Toronto Mississauga November 9, 2017 Matt Tudball (University of Toronto) ECO375H5 November 9, 2017 1 / 24 Review: Heteroscedasticity Consider
More informationECON Introductory Econometrics. Lecture 11: Binary dependent variables
ECON4150 - Introductory Econometrics Lecture 11: Binary dependent variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 11 Lecture Outline 2 The linear probability model Nonlinear probability
More informationLinear Regression with Multiple Regressors
Linear Regression with Multiple Regressors (SW Chapter 6) Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS 4. Measures of fit 5. Sampling distribution
More informationApplied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 7 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 68 Outline of Lecture 7 1 Empirical example: Italian labor force
More information5. Let W follow a normal distribution with mean of μ and the variance of 1. Then, the pdf of W is
Practice Final Exam Last Name:, First Name:. Please write LEGIBLY. Answer all questions on this exam in the space provided (you may use the back of any page if you need more space). Show all work but do
More informationCourse Econometrics I
Course Econometrics I 4. Heteroskedasticity Martin Halla Johannes Kepler University of Linz Department of Economics Last update: May 6, 2014 Martin Halla CS Econometrics I 4 1/31 Our agenda for today Consequences
More informationEconometrics. 8) Instrumental variables
30C00200 Econometrics 8) Instrumental variables Timo Kuosmanen Professor, Ph.D. http://nomepre.net/index.php/timokuosmanen Today s topics Thery of IV regression Overidentification Two-stage least squates
More informationIntroduction to Econometrics. Review of Probability & Statistics
1 Introduction to Econometrics Review of Probability & Statistics Peerapat Wongchaiwat, Ph.D. wongchaiwat@hotmail.com Introduction 2 What is Econometrics? Econometrics consists of the application of mathematical
More informationTitle. Description. var intro Introduction to vector autoregressive models
Title var intro Introduction to vector autoregressive models Description Stata has a suite of commands for fitting, forecasting, interpreting, and performing inference on vector autoregressive (VAR) models
More informationHandout 12. Endogeneity & Simultaneous Equation Models
Handout 12. Endogeneity & Simultaneous Equation Models In which you learn about another potential source of endogeneity caused by the simultaneous determination of economic variables, and learn how to
More informationProblem Set 5 ANSWERS
Economics 20 Problem Set 5 ANSWERS Prof. Patricia M. Anderson 1, 2 and 3 Suppose that Vermont has passed a law requiring employers to provide 6 months of paid maternity leave. You are concerned that women
More informationMicroeconometrics: Clustering. Ethan Kaplan
Microeconometrics: Clustering Ethan Kaplan Gauss Markov ssumptions OLS is minimum variance unbiased (MVUE) if Linear Model: Y i = X i + i E ( i jx i ) = V ( i jx i ) = 2 < cov i ; j = Normally distributed
More informationApplied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 5 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 44 Outline of Lecture 5 Now that we know the sampling distribution
More informationLinear Regression with Time Series Data
Econometrics 2 Linear Regression with Time Series Data Heino Bohn Nielsen 1of21 Outline (1) The linear regression model, identification and estimation. (2) Assumptions and results: (a) Consistency. (b)
More informationLecture 8: Instrumental Variables Estimation
Lecture Notes on Advanced Econometrics Lecture 8: Instrumental Variables Estimation Endogenous Variables Consider a population model: y α y + β + β x + β x +... + β x + u i i i i k ik i Takashi Yamano
More informationThe Classical Linear Regression Model
The Classical Linear Regression Model ME104: Linear Regression Analysis Kenneth Benoit August 14, 2012 CLRM: Basic Assumptions 1. Specification: Relationship between X and Y in the population is linear:
More informationIntroduction to Econometrics. Heteroskedasticity
Introduction to Econometrics Introduction Heteroskedasticity When the variance of the errors changes across segments of the population, where the segments are determined by different values for the explanatory
More informationSection Least Squares Regression
Section 2.3 - Least Squares Regression Statistics 104 Autumn 2004 Copyright c 2004 by Mark E. Irwin Regression Correlation gives us a strength of a linear relationship is, but it doesn t tell us what it
More informationESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics
ESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics July 2009 1. Introduction 2. The Sharp RD Design 3.
More informationHeteroskedasticity Example
ECON 761: Heteroskedasticity Example L Magee November, 2007 This example uses the fertility data set from assignment 2 The observations are based on the responses of 4361 women in Botswana s 1988 Demographic
More information