Problem Set 2

Size: px
Start display at page:

Download "Problem Set 2"

Transcription

1 Problem Set 2 Paul Schrimpf Due: Friday, March 14, 2008 Problem 1 IV Basics Consider the model: y = Xβ + ɛ where X is T K. You are concerned that E[ɛ X] 0. Luckily, you have M > K variables, Z, which you think might serve as instruments. Solution to this question was written by Konrad last year. (a) What properties are required for Z to be a valid instrument? What is the exclusion restric- Question: tion? Answer: Z must satisfy two conditions: it must (1) have sufficient covariation with X in order to satisfy the rank condition ( rank plim 1 ) T Z X = rank(q ZX ) = k and (2) be uncorrelated with the disturbance ε, i.e. (and this is the exclusion restriction) (b) E[Z ε] = 0 Question: What is the IV estimator using Z as an instrument? What is its asymptotic distribution? Assume E[ɛɛ Z] = Σ. Answer In order to estimate β we construct instruments W = ZA T, where A T is a full-rank m k matrix that converges in probability to a limit A (I m allowing A T to be stochastic in order to nest the 2SLS case in the next part of the problem). Then we can form the usual IV estimator using W as instruments, ˆβ AT = (W X) 1 W y = (A TZ X) 1 A TZ y = β + (A TZ X) 1 A TZ ε From the exclusion restriction, plim ˆβ A = β for any A, and by the usual arguments, plim 1 T A TZ X = A Q ZX 1

2 e.g. from the Chebychev LLN. From an appropriate CLT (e.g. Lindeberg-Levy if we only have heteroskedasticity but individuals are i.i.d. draws from one stable distribution, or a CLT for mixing processes if we have (not too much) autocorrelation) and the Cramér-Wold device where by the law of iterated expectations T V 1 2 T Z ε = N (0, I k ) V T = 1 T E[Z εε Z] = 1 T E[Z E[εε Z]Z] = 1 T E[Z ΣZ] Since A T A, we can apply the continuous mapping theorem to the asymptotic variance of the IV estimator in order to obtain ( plimt Var( ˆβ A AT ) = plim T Z ) 1 X A ( T Z ΣZA T X ) 1 ZA T = (A Q ZX) 1 A Q ZΣZA(Q T T T ZXA) 1 where Q ZΣZ := plim 1 T E[Z ΣZ]. (c) Question: Suppose we form an instrument, Ŵ = ZÂ, where  = (Z Z) 1 Z X, so Ŵ is T K. Show that ˆβ 2SLS = (Ŵ X) 1 Ŵ y is consistent. Calculate the asymptotic variance of ˆβ 2SLS. Answer: 2SLS is a special case of the IV estimator discussed above, and therefore it is consistent whenever the rank condition is satisfied, and by plugging in, the variance becomes (d) ( T Var( ˆβ X Z(Z Z) 1 Z ) 1 X Z ( ΣZ X Z(Z Z) 1 Z X 2SLS ) = T T T Question: Suppose that ˆΩ is a consistent estimator of the limit of Z ΣZ/T. Consider the following minimum chi-squared estimator: ˆβ χ = arg min(y Xβ) Z ˆΩ 1 Z (y Xβ) β Show that ˆβ χ = (à Z X) 1 (à Z y) (so ˆβ χ is an IV estimator) for some Ã. ) 1 Answer: The first-order condition for this minimization problem is X Z ˆΩ 1 Z (y Xβ) = 0 ˆβ IV = (X Z ˆΩ 1 ZX) 1 X Z ˆΩ 1 Z y which fits into the framework in (b) by choosing à = ˆΩ 1 ZX, which corresponds to 2SLS after a GLS transformation of the data. (e) Question: Show that the à is optimal in minimizing the asymptotic variance of IV. What does this imply about the asymptotic efficiency of ˆβ IV relative to 2SLS? (Hint: you can think of this as a linear model of the form Z y = Z Xβ + Z u and apply the Gauss-Markov theorem asymptotically.) 2

3 Answer: After the GLS transformation of the data by premultiplying all variables (including the constant) with (ˆΩ 1 ) 1 2, the linear model Z y = Z Xβ 0 + Z ε satisfies the Gauss-Markov assumptions, and therefore in the limit, 2SLS is BLUE (only asymptotically, there s still bias in finite samples!). Problem 2 Measurement Error This question comes from last year s midterm. Consider the simple linear model y i = β 1 + β 2 x i + ɛ i i = 1,..., n, where x i is unobserved but a measurement of x i of x i is observed with x i = x i + v i, where E(x i v i) = 0 and E(v i ɛ i ) = 0, E(x i ɛ i) = 0. (a) Question: What is the probability limit of the least squares estimator of β? How big would V ar(v i ) have to be relative to V ar(x i ) have to be for the plim of the least square estimator to be half the size of the true parameter. Answer: Let X i = x i x. The limit of ˆβ 2 is plim ˆβ 2 =β 2 + (plim 1 n X X) 1 (plim 1 n X (ɛ β 2 v)) ( ) =β 2 1 σ2 v σx 2 + σv 2 where σ 2 x = Var(x i ) and σ2 v = Var(v i ). Generally, we don t care very much about the constant in a regression, so it would be probably fine to answer this question without saying anything about ˆβ 1, but for completeness: ˆβ 1 =(1 nm x 1 n ) 1 (1 nm x y) where 1 n is an n 1 vector of ones, and M x is the orthogonal projection to x. Expanding gives: ˆβ 1 =β nm x (β 2 x + ɛ) n 1 nx(x x) 1 x 1 n 2 x +σ2 x ) p β 1 + β 2µ x µxβ 2(µ σx 2+µ2 x p β 1 1 µ2 x µ 2 x +σ2 x so ˆβ 1 is consistent. The reason is that even though x is measured with error, the constant is uncorrelated with x, so it s coefficient remains consistent. (b) Question: Suppose you have an instrumental variable z i for x i. What properties does z i need to have for consistent estimation in part (i)? Could your instrument be measured with error and still have a consistent estimator? 3

4 Answer: Like in problem 1, z must be correlated with x i, but uncorrelated with the error, ɛ i β 2 v i. This is possible even if z is measured with error, as long as the measurement error in z is uncorrelated with ɛ and the measurement error in x. (c) Question: You are worried that your instrumental variable may have a little bit of correlation with v. However, you are quite confident that Corr(x i, v i ) > Corr(z i, v i ). Can you make any conclusion that the IV estimator will be less inconsistent than the OLS estimator? Answer: From part (a), we know that ( ) plim ˆβ 2 O LS = β 2 1 σ2 v σx 2 + σv 2 = β(1 ρ xv ) where ρ xv is the correlation between x and v. Let s look at the plim of IV. Remove the mean from z and x. Since z and x are scalars, the choice of A in our IV estimator does not matter, so let s just take A = 1. Then, ˆβ IV =(Z X) 1 (Z y) =β 2 + (Z X) 1 (Z (ɛ β 2 v)) p β 2 (1 σ zv σ zx + σ zv ) p ρ zv β 2 (1 ) ρ zx + ρ zv So, the bias of IV depends on both the correlation of z and v, and the correlation of z and x. Regardless of ρ zv, if ρ zx is small, then ˆβ IV can be more biased than OLS. (d) Question: Suppose you have a nonclassical measurement error (E[x i v i] 0) but you know that Cov(x i, ɛ i ) = 0. What will be the plim of the least squares estimator of β? Show that you will have attenuation bias if Cov(x i, x i ) > 0 and V ar(x i) > V ar(x i ). Answer: The plim of ˆβ 2 is: plim ˆβ 2 =(plim 1 n X X) 1 (plim 1 n X X β 2 ) =β 2 cov(x, x ) Var(x) In general, the Cauchy-Schwarz inequality tells us that cov(x, y) 2 Var(x)Var(y). Then, since Var(x ) Var(x), we have cov(x,x ) Var(x) < 1, so there is attenuation bias. Problem 3 Misspecified First Stage Suppose you re interested in the effect of an educational policy on child outcomes. covariates, you have a model that looks like: After partialing out outcome i = β 1 policylevel i + ɛ i 4

5 where outcome is something like a standardized test score, and policylevel is something like the studentteacher ratio. For this exercise, think of the policy as a continuous choice variable. You re worried that policy level is endogenous. A school that has a lower student-teacher ratio is also likely to be higher quality in other, unobserved dimensions. You think that prior test scores are a valid instrument for the policy level. You think that prior test scores are related to the policy level through: policylevel i = α 1 priortest i + v i Sadly, you re wrong about the functional form of this equation. The truth is that policylevel i = α 0 priortest i p + v i that is, the policy level is actually a function of how far prior test scores are from some cutoff level. This question explores how misspecifying the first stage affects IV. Throughout, you may assume that E[ɛ priortest] = 0 and that ɛ i is i.i.d. The very last part of this question is especially difficult. You probably have not yet seen all the tools required to answer it, so don t worry if you can t. However, if you can answer the last question and prove your answer, you can skip all the other parts of this question. (a) Question: Suppose you had correctly specified the first stage and know p. You estimate β 1 using Z = priortest i p as an instrument. ˆβ IV = (A Z X) 1 (A Z y) where X = policylevel and y = outcome. Is this estimator consistent? What is its asymptotic variance? (You can just state your answer, no proof necessary.) Answer: Yes, this estimator is consistent because we know that E[ɛZ] = 0. As in problem 1, its asymptotic variance is: V ( ˆβ IV ) = (A Q ZX ) 1 A Q ZΣZ A(Q ZXA) 1 In this case, assuming that we have partialed out a constant so all variables are mean zero Q ZX = plim 1 N Z X =α 0 Var( priortest i p ) Q ZΣZ = plim 1 N Z ɛɛ Z =σ 2 Var( priortest i p ) so, V ( ˆβ IV ) = σ 2 α 2 0 Var( priortest i p ) (b) Question: estimate Now, suppose you choose the wrong first stage and use Z = priortest i as an instrument to β IV = (Ã Z X) 1 (Ã Z y) Is this estimator consistent? What is its asymptotic variance? 5

6 Answer: Yes, this estimator is consistent, provided that cov(x, Z) 0. As above, its asymptotic variance is given by the formula in question 1. As in part (a), if we substitute for Q ZX and Q ZΣZ we get V ( β IV ) = σ 2 α 2 1 Var(priortest i) (c) (i) Finite Sample Properties Question: How would you expect β IV to perform relative to ˆβ IV in small samples? Is either estimator unbiased? Would you prefer one over the other? Give some intuition for your answer. Answer: Neither IV estimator is unbiased. The reason is that E[ ˆβ IV β] = E [ (A Z X) 1 (A Z ɛ) ] Since X is correlated with ɛ, this expectation is generally non-zero. After spring break, we ll look at the finite sample bias of IV in greater detail. If we get the functional form wrong, the first stage may be substantially weaker (the denominator of the bias smaller), and as you will see this typically increases the finite-sample bias. (ii) Let s simulate the model to compare the two estimators. Let: x i = α z i z + u i + v i y i =βx i + u i + e i where α = β = 1. Let z, u, v, and e be i.i.d, N(0, 1) variables. Set z = 0.5. Simulate the model many times and compute ˆβ OLS, ˆβ2SLS = (X P Z X) 1 (X P Z y), and β 2SLS = (X P ZX) 1 (X P Zy). Report the MSE, median, and probability of reject H 0 : β = 1 for each of these estimators. Examine how the results depend on the sample size. Answer: From Konrad last year: here are my simulation results: OLS coefficient (median) IV coefficient, (median) OLS coefficient (iqr) IV coefficient, (iqr) OLS Rejection Rates IV Rejection Rates Here you can again see that the bias of OLS is unaffected by sample size, whereas the substantial median bias in the IV coefficient for N = 20 goes away as the sample size increases. Looking at the inter-quartile range, the IV estimator seems to have a standard error about 5 times the size of that of OLS, which is much more than in problem 5 even though most other magnitudes were comparable, and this is in fact a consequence of the bad approximation of the functional form of the first stage. In fact, if z were zero, this strategy breaks down completely: beta = OLS coefficient (median) IV coefficient, (median)

7 OLS coefficient (iqr) IV coefficient, (iqr) OLS Rejection Rates IV Rejection Rates You can see that IV has about the same bias as OLS, and furthermore, the interquartile range doesn t decrase at all even though from the first to the third column, we increased the sample size by a factor of 100! This is a consequence of having no first stage anymore, and again this should give you a taste for what the second half of the course is going to be about. (d) (i) Asymptotically Optimal Instruments Question: What choices of A and à minimize the asymptotic variances of ˆβ IV and β IV? Which estimator has a smaller asymptotic variance with this A and Ã? Answer: The easiest answer is that A does not matter, since in both cases, Z and X are scalars. If this were not the case, A would matter, but we already know from problem 1 that the optimal A is A = Ω 1 Z X. Here, since ɛ is iid, Ω = Z Zσ 2, so the optimal A is just (Z Z) 1 Z X, which gives us 2SLS. Below, we ll show that Z E[x z] yields the minimum asymptotic variance among instruments of the form Z = g(z). Applying this result, it must be that ˆβ IV has smaller variance. (ii) Question: Consider a more general setup where instead of the first stage given above you have E[X Z] = f(z) for some function. Among IV estimators of the form: what choice of g() and A is asymptotically efficient? ˆβ IV (g(z), A) =(A g(z) X) 1 (A g(z) y) Answer: For a given g(), we know that A = (g(z) g(z)) 1 g(z) X is optimal for the same reason as in part (d)(i). Alternatively, we know that A doesn t matter when x is a scalar and g(z) is scalar valued. Sticking to the scalar case and assuming all variables have been de-meaned, the asymptotic variance for ˆβ using g(z) as an instrument is: V (β g ) = σ 2 Var(g(z)) cov(x, g(z)) 2 We can rewrite cov(x, g(z)) as: cov(x, g(z)) =E[xg(z)] =E[E[x z]g(z)] =E[f(z), g(z)] = cov(f(z), g(z)) Now comes the tricky part. Observe that the set of functions that map z to R, equipped with innner product f, g = cov(f(z), g(z)) forms an inner product space. The Cauchy-Schwarz inequality tells us that: f, g 2 f, f g, g Therefor, cov(f(z), g(z)) 2 Var(f(z))Var(g(z)) and in particular, V (β g ) = σ 2 Var(g(z)) cov(f(z), g(z)) 2 V (β f ) = σ 2 1 Var(f(z)) 7

8 (iii) Question: Suppose you don t know f(z) = E[X Z] exactly, but know it up to a finite dimensional parameter, θ. Can you construct a feasible estimator that has the same asymptotic variance as the efficient estimator when f(z) is known? Suppose you don t know anything about f(z), except that it s a wellbehaved function. Is there a feasible estimator with the same asymptotic variance as the efficient estimator when f(z) is known? Answer: The answer is yes, you can construct a feasible efficient estimator in both cases. Newey (1990) showed this result in a slightly more general setting when f(z) is completely unknown. To begin with, let s suppose we have some estimate of f(z), ˆf(z), and look at the asymptotic distribution of ˆβ ˆf. For simplicity, we will stick with the case where x, z, and f(z) are all scalars. We have: ˆβ ˆf =( ˆf X) 1 ˆf y =β + ( ˆf X) 1 ˆf ɛ where ˆf is short for ˆf(z). Notice that we can always write x = E[x z] + v = f(z) + v where E[v z] = 0. Thus, we have ˆβ ˆf β = ˆf ɛ ˆf (f + v) f ˆβ ɛ + ( ˆf β = ˆf f) ɛ f (f + v) + ( ˆf f) (f + v) n( ˆβ ˆf β) = 1 n f ɛ + 1 n ( ˆf f) ɛ 1 n f (f + v) + 1 n ( ˆf f) (f + v) The first terms in the numerator and denominator are exactly what we would get if f were known. Therefore, ˆβ ˆf has the same distribution as ˆβ f if and only if and plim 1 n ( ˆf f) ɛ = 0 (1) plim 1 n ( ˆf f) (f + v) = 0 (2) Equation (2) will be satisfied, if ˆf converges to f in probability uniformly in z. That is, if for any ɛ > 0, there s an N such that P ( ˆf(z) f(z) > ɛ) < ɛ z Z where Z is the domain of z. When f is known up to a finite dimensional parameter, then we know that we can estimate θ consistently by doing nonlinear-least squares. In this case, ˆf(z) f(z) =f(z, ˆθ) f(z, θ) = f(z, θ) (ˆθ θ) θ so ˆf p f uniformly if f θ is bounded. More generally, in this case plim 1 n ( ˆf f) (f + v) = plim(ˆθ θ) 1 n f(z i, θ) (f(z i ) + v i ) n θ i=1 [ ] f(zi, θ) =0 E (f(z i ) + v i ) θ 8

9 [ ] so we just need E f(zi,θ) 1 θ (f(z i ) + v i ) to be finite. A similar argument shows that if n f θ ɛ converges in distribution, then equation (1) is satisfied. Therefore, in the parametric case (when f is known up a finite number of parameters) a sufficient condition (in addition to the usual conditions needed for IV) for ˆβ ˆf to be [ ( ) ] 2 asymptotically efficient is that E f θ is finite. When f is completely unknown, we must estimate it nonparametrically. The details of showing that (1) and (2) are satisfied depend on the particular nonparametric estimator we use. Problem 4 Inference with Difference in Differences In this question, you ll reproduce some of the results from Bertrand, Duflo, and Mullainathan (2004). You ll then explore how the bias and size corrected test of Hausman and Kuersteiner (2007) performs in this setup. On the course website, I posted data from the CPS merged outgoing rotation groups. I originally obtained the data from I merged the CPS-MORG data for all the years and kept a few of the variables. If you re curious, I posted the do-file I used. This is the same data as used by Bertrand, Duflo, and Mullainathan (2004). The variables in the data are: ˆ state state census code ˆ year ˆ age ˆ female 1 = female, 0 = male ˆ yrsedu years of education ˆ earnwke weekly earnings ˆ employed whether employed ˆ white 1=white, 0=non-white ˆ black 1=black, 0=non-black The data set is rather large. If it is annoyingly slow to work with, you might want to limit the sample in some way. Bertrand, Duflo, and Mullainathan just use prime age white women. They also aggregate the data to state-year means for most of their simulations (the ones where the first column of the table is labeled CPS agg). You will probably also want to aggregate the data to save time. You will need to aggregate the data to impliment Hausman and Kuersteiner s estimator. (a) Reproducing Bertrand, Duflo, and Mullainathan Question: Perform similar simulations as Bertrand, Duflo, and Mullainathan (2004), and reproduce their key results. More specifically, randomly simulate policy changes and use difference in differences to estimate the effect of these changes. For each simulation, first sample 50 states with replacement from the data. Then, simulate the policy changes by randomly picking a year between 1985 and Randomly choose half of the states to have been affected. Then, estimate y ist = α s + λ t + γt st + x ist β + ɛ ist Estimate γ using OLS with various types of standard errors. Specifically, consider the usual, homoskedastic, iid OLS standard errors, standard errors clustered at state and year, and standard errors clustered at state. (If you want, you can also try some of the other things Bertrand, Duflo, and Mullainathan did). Report the fraction of simulations where you reject H 0 : γ = 0. Do this under both γ = 0 and γ = Explore how the results vary with N (the number of states) and T (the number of years). 9

10 Answer: See the code I posted. Table 1 shows the results of the simulations. The results are similar to Bertrand, Duflo, and Mullainathan. OLS and OLS state-year clustered over-reject for large N and T. OLS state clustered works well for N=50 and 20, but less well for smaller N. (b) FGLS Question: Perform the same simulations, but estimate γ using FGLS. Compare the results to OLS. As in Hausman and Kuersteiner (or section IV.D of Bertrand, Duflo, and Mullainathan), assume that the autocorrelation process is the same in all states and there is no heteroskedasticity. Again, explore how the results depend on N and T. Answer: The presence of fixed effects makes doing FGLS somewhat tricky. In particular, the natural estimate of the variance matrix, S ˆΣ = ê s ê s will be rank T 1 instead of rank T. You can proceed as in Kiefer (1980). Specifically: s=1 1. Aggregate the data to state-year averages. As in Bertrand, Duflo, and Mullainathan, first regress the outcome on any demographics you want to control for. Then, take the mean of the residuals within each state and year. Call this y s,t. Using OLS, etimate: y s,t = α s + λ t + γt s,t + e s,t 2. Form ˆΣ = S s=1 êsê s where ê s is the T 1 vector of residuals from state s. ˆΣ is singular, so we need to take a generalized inverse instead of its inverse. Following Kiefer (1980), let S be the T 1 T 1 matrix consisting of the first T 1 rows and columns of ˆΣ. Let S = S Estimate ˆβ F GLS = (X (I s M 1t S 1 M 1t )X) 1 (X (I s M 1t S 1 M 1t )y), where M 1t = I t 1 T 1 T 1 T is the matrix that projects out fixed effects, and X consists of time fixed effects and the policy indicator, and y = [y 11, y 12,..., y 1T,..., y ST ] and X is similarly arranged. See the code for more details. The results are in Table 1. FGLS does not perform very well. It only has nearly correct size for N=50 and T =10. However, in these situations FGLS is dominated in terms of power by OLS state clustered. (c) Bias-corrected FGLS Since T is small, and the regression includes state fixed effects, estimates of the variance matrix will be biased. Hausman and Kuersteiner show how to correct this bias (page 7). Implement this correction and repeat the simulations in part (b). How much does this bias correction help? How large does T need to be for the bias to not matter? Hint: there were two things that tripped me up when I was trying to do this. First, Hausman and Kuersteiner sort the variables as time 1 state 1, time 1 state 2,... time 1 state S, time 2 state 1, etc. This is the opposite of what I had been doing. You ll need to either reorder your variables, or reverse the order of the kronecker products in Hausman and Kuersteiner s formulas. The other thing that was difficult was figuring out what was meant by V. V s is supposed to be a vector of all right hand side variables at all times, except for state fixed effects. Since, after aggregating, the only variables we have are a policy indicator and time fixed effects, if you naively include all of these, V will not be full. 10

11 rank. In our setup, V s should just be two variables, a constant, and an indicator for whether the policy was ever impleented. Once you figure out the two things mentioned in the hint, implementing the bias correction is just a matter of following Hausman and Kuersteiner s formulas. See the code for details. In the table, we that bias-corrected FGLS almost always does better than plain FGLS. (d) Size-corrected Tests Question: Hausman and Kuersteiner (2007) also propose a size correction for testing H 0 : γ = γ 0 for when N is small. 1 Repeat the simulations in the previous sub-question using size corrected tests. Discuss the results. Answer: Following the theorem 3.1 of Hausman and Kuersteiner, the size-corrected tests, replaces the usual t = 1.96 critical value with ( t sc = t 1 + (1 + ) t2 ) + 2(T 2) 2S Since t sc > t, we know that size and bias corrected FGLS will always have a smaller size and power than bias corrected FGLS. However, this is a good thing, since bias corrected FGLS tends to over-rejct. From the table, we see that the size correction works fairly well, except for N=6, T=5. I am fairly impressed by how well the size correction works. I expected the results to be much worse. One method of inference in difference in differences setups, which we have not talked about, is the synthetic control method of Abadie, Diamond, and Hainmueller (2007). Abadie presented this paper at the econometrics seminar recently, and it appears quite promising. If you ever estimate a difference in differences in your research, the Abadie paper is definitely worth reading. Table 1: Simulation Results Estimator N T Rejection Rate No Effect Rejection Rate 2% Effect OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected Notice that this setup falls under the special case in section 3 of the paper, so you can use the simpler formulas for the test given in theorem

12 FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected

13 OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected Based on 100 simulations 13

1 Motivation for Instrumental Variable (IV) Regression

1 Motivation for Instrumental Variable (IV) Regression ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data

More information

1 Introduction to Generalized Least Squares

1 Introduction to Generalized Least Squares ECONOMICS 7344, Spring 2017 Bent E. Sørensen April 12, 2017 1 Introduction to Generalized Least Squares Consider the model Y = Xβ + ɛ, where the N K matrix of regressors X is fixed, independent of the

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

Final Exam. Economics 835: Econometrics. Fall 2010

Final Exam. Economics 835: Econometrics. Fall 2010 Final Exam Economics 835: Econometrics Fall 2010 Please answer the question I ask - no more and no less - and remember that the correct answer is often short and simple. 1 Some short questions a) For each

More information

Casuality and Programme Evaluation

Casuality and Programme Evaluation Casuality and Programme Evaluation Lecture V: Difference-in-Differences II Dr Martin Karlsson University of Duisburg-Essen Summer Semester 2017 M Karlsson (University of Duisburg-Essen) Casuality and Programme

More information

GLS and FGLS. Econ 671. Purdue University. Justin L. Tobias (Purdue) GLS and FGLS 1 / 22

GLS and FGLS. Econ 671. Purdue University. Justin L. Tobias (Purdue) GLS and FGLS 1 / 22 GLS and FGLS Econ 671 Purdue University Justin L. Tobias (Purdue) GLS and FGLS 1 / 22 In this lecture we continue to discuss properties associated with the GLS estimator. In addition we discuss the practical

More information

The Simple Linear Regression Model

The Simple Linear Regression Model The Simple Linear Regression Model Lesson 3 Ryan Safner 1 1 Department of Economics Hood College ECON 480 - Econometrics Fall 2017 Ryan Safner (Hood College) ECON 480 - Lesson 3 Fall 2017 1 / 77 Bivariate

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Multiple Equation GMM with Common Coefficients: Panel Data

Multiple Equation GMM with Common Coefficients: Panel Data Multiple Equation GMM with Common Coefficients: Panel Data Eric Zivot Winter 2013 Multi-equation GMM with common coefficients Example (panel wage equation) 69 = + 69 + + 69 + 1 80 = + 80 + + 80 + 2 Note:

More information

Non-Spherical Errors

Non-Spherical Errors Non-Spherical Errors Krishna Pendakur February 15, 2016 1 Efficient OLS 1. Consider the model Y = Xβ + ε E [X ε = 0 K E [εε = Ω = σ 2 I N. 2. Consider the estimated OLS parameter vector ˆβ OLS = (X X)

More information

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Econometrics Week 8 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 25 Recommended Reading For the today Instrumental Variables Estimation and Two Stage

More information

IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors

IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors Laura Mayoral IAE, Barcelona GSE and University of Gothenburg Gothenburg, May 2015 Roadmap Deviations from the standard

More information

The BLP Method of Demand Curve Estimation in Industrial Organization

The BLP Method of Demand Curve Estimation in Industrial Organization The BLP Method of Demand Curve Estimation in Industrial Organization 9 March 2006 Eric Rasmusen 1 IDEAS USED 1. Instrumental variables. We use instruments to correct for the endogeneity of prices, the

More information

Economics 583: Econometric Theory I A Primer on Asymptotics

Economics 583: Econometric Theory I A Primer on Asymptotics Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:

More information

Economics 582 Random Effects Estimation

Economics 582 Random Effects Estimation Economics 582 Random Effects Estimation Eric Zivot May 29, 2013 Random Effects Model Hence, the model can be re-written as = x 0 β + + [x ] = 0 (no endogeneity) [ x ] = = + x 0 β + + [x ] = 0 [ x ] = 0

More information

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012 Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.

More information

1 Appendix A: Matrix Algebra

1 Appendix A: Matrix Algebra Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix

More information

Introduction to Econometrics. Heteroskedasticity

Introduction to Econometrics. Heteroskedasticity Introduction to Econometrics Introduction Heteroskedasticity When the variance of the errors changes across segments of the population, where the segments are determined by different values for the explanatory

More information

Econometrics - 30C00200

Econometrics - 30C00200 Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business

More information

Instrumental Variables

Instrumental Variables Instrumental Variables Department of Economics University of Wisconsin-Madison September 27, 2016 Treatment Effects Throughout the course we will focus on the Treatment Effect Model For now take that to

More information

Panel Data Models. James L. Powell Department of Economics University of California, Berkeley

Panel Data Models. James L. Powell Department of Economics University of California, Berkeley Panel Data Models James L. Powell Department of Economics University of California, Berkeley Overview Like Zellner s seemingly unrelated regression models, the dependent and explanatory variables for panel

More information

Ordinary Least Squares Regression

Ordinary Least Squares Regression Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section

More information

Spatial Regression. 3. Review - OLS and 2SLS. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Spatial Regression. 3. Review - OLS and 2SLS. Luc Anselin.   Copyright 2017 by Luc Anselin, All Rights Reserved Spatial Regression 3. Review - OLS and 2SLS Luc Anselin http://spatial.uchicago.edu OLS estimation (recap) non-spatial regression diagnostics endogeneity - IV and 2SLS OLS Estimation (recap) Linear Regression

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 6: Bias and variance (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 49 Our plan today We saw in last lecture that model scoring methods seem to be trading off two different

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the

More information

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) 1 2 Model Consider a system of two regressions y 1 = β 1 y 2 + u 1 (1) y 2 = β 2 y 1 + u 2 (2) This is a simultaneous equation model

More information

F9 F10: Autocorrelation

F9 F10: Autocorrelation F9 F10: Autocorrelation Feng Li Department of Statistics, Stockholm University Introduction In the classic regression model we assume cov(u i, u j x i, x k ) = E(u i, u j ) = 0 What if we break the assumption?

More information

Econometrics Master in Business and Quantitative Methods

Econometrics Master in Business and Quantitative Methods Econometrics Master in Business and Quantitative Methods Helena Veiga Universidad Carlos III de Madrid Models with discrete dependent variables and applications of panel data methods in all fields of economics

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

ECONOMICS 210C / ECONOMICS 236A MONETARY HISTORY

ECONOMICS 210C / ECONOMICS 236A MONETARY HISTORY Fall 018 University of California, Berkeley Christina Romer David Romer ECONOMICS 10C / ECONOMICS 36A MONETARY HISTORY A Little about GLS, Heteroskedasticity-Consistent Standard Errors, Clustering, and

More information

ECO375 Tutorial 8 Instrumental Variables

ECO375 Tutorial 8 Instrumental Variables ECO375 Tutorial 8 Instrumental Variables Matt Tudball University of Toronto Mississauga November 16, 2017 Matt Tudball (University of Toronto) ECO375H5 November 16, 2017 1 / 22 Review: Endogeneity Instrumental

More information

Topic 6: Non-Spherical Disturbances

Topic 6: Non-Spherical Disturbances Topic 6: Non-Spherical Disturbances Our basic linear regression model is y = Xβ + ε ; ε ~ N[0, σ 2 I n ] Now we ll generalize the specification of the error term in the model: E[ε] = 0 ; E[εε ] = Σ = σ

More information

EC402 - Problem Set 3

EC402 - Problem Set 3 EC402 - Problem Set 3 Konrad Burchardi 11th of February 2009 Introduction Today we will - briefly talk about the Conditional Expectation Function and - lengthily talk about Fixed Effects: How do we calculate

More information

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)

More information

Introduction to Econometrics Midterm Examination Fall 2005 Answer Key

Introduction to Econometrics Midterm Examination Fall 2005 Answer Key Introduction to Econometrics Midterm Examination Fall 2005 Answer Key Please answer all of the questions and show your work Clearly indicate your final answer to each question If you think a question is

More information

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Please answer all of the questions and show your work. If you think a question is ambiguous, clearly state how you interpret it before

More information

Stat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS

Stat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS Stat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS 1a) The model is cw i = β 0 + β 1 el i + ɛ i, where cw i is the weight of the ith chick, el i the length of the egg from which it hatched, and ɛ i

More information

Linear Regression. Junhui Qian. October 27, 2014

Linear Regression. Junhui Qian. October 27, 2014 Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency

More information

Econometrics II - EXAM Answer each question in separate sheets in three hours

Econometrics II - EXAM Answer each question in separate sheets in three hours Econometrics II - EXAM Answer each question in separate sheets in three hours. Let u and u be jointly Gaussian and independent of z in all the equations. a Investigate the identification of the following

More information

The returns to schooling, ability bias, and regression

The returns to schooling, ability bias, and regression The returns to schooling, ability bias, and regression Jörn-Steffen Pischke LSE October 4, 2016 Pischke (LSE) Griliches 1977 October 4, 2016 1 / 44 Counterfactual outcomes Scholing for individual i is

More information

A Primer on Asymptotics

A Primer on Asymptotics A Primer on Asymptotics Eric Zivot Department of Economics University of Washington September 30, 2003 Revised: October 7, 2009 Introduction The two main concepts in asymptotic theory covered in these

More information

New Developments in Econometrics Lecture 11: Difference-in-Differences Estimation

New Developments in Econometrics Lecture 11: Difference-in-Differences Estimation New Developments in Econometrics Lecture 11: Difference-in-Differences Estimation Jeff Wooldridge Cemmap Lectures, UCL, June 2009 1. The Basic Methodology 2. How Should We View Uncertainty in DD Settings?

More information

Review of Econometrics

Review of Econometrics Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,

More information

Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares

Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares L Magee Fall, 2008 1 Consider a regression model y = Xβ +ɛ, where it is assumed that E(ɛ X) = 0 and E(ɛɛ X) =

More information

Short Questions (Do two out of three) 15 points each

Short Questions (Do two out of three) 15 points each Econometrics Short Questions Do two out of three) 5 points each ) Let y = Xβ + u and Z be a set of instruments for X When we estimate β with OLS we project y onto the space spanned by X along a path orthogonal

More information

Regression #3: Properties of OLS Estimator

Regression #3: Properties of OLS Estimator Regression #3: Properties of OLS Estimator Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #3 1 / 20 Introduction In this lecture, we establish some desirable properties associated with

More information

Quick Review on Linear Multiple Regression

Quick Review on Linear Multiple Regression Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,

More information

Instrumental Variables and the Problem of Endogeneity

Instrumental Variables and the Problem of Endogeneity Instrumental Variables and the Problem of Endogeneity September 15, 2015 1 / 38 Exogeneity: Important Assumption of OLS In a standard OLS framework, y = xβ + ɛ (1) and for unbiasedness we need E[x ɛ] =

More information

Instrumental Variables, Simultaneous and Systems of Equations

Instrumental Variables, Simultaneous and Systems of Equations Chapter 6 Instrumental Variables, Simultaneous and Systems of Equations 61 Instrumental variables In the linear regression model y i = x iβ + ε i (61) we have been assuming that bf x i and ε i are uncorrelated

More information

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Estimation - Theory Department of Economics University of Gothenburg December 4, 2014 1/28 Why IV estimation? So far, in OLS, we assumed independence.

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

CHAPTER 6: SPECIFICATION VARIABLES

CHAPTER 6: SPECIFICATION VARIABLES Recall, we had the following six assumptions required for the Gauss-Markov Theorem: 1. The regression model is linear, correctly specified, and has an additive error term. 2. The error term has a zero

More information

Recitation 1: Regression Review. Christina Patterson

Recitation 1: Regression Review. Christina Patterson Recitation 1: Regression Review Christina Patterson Outline For Recitation 1. Statistics. Bias, sampling variance and hypothesis testing.. Two important statistical theorems: Law of large numbers (LLN)

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

A Significance Test for the Lasso

A Significance Test for the Lasso A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical

More information

Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation

Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation Michele Aquaro University of Warwick This version: July 21, 2016 1 / 31 Reading material Textbook: Introductory

More information

ECON Introductory Econometrics. Lecture 16: Instrumental variables

ECON Introductory Econometrics. Lecture 16: Instrumental variables ECON4150 - Introductory Econometrics Lecture 16: Instrumental variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 12 Lecture outline 2 OLS assumptions and when they are violated Instrumental

More information

Freeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94

Freeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94 Freeing up the Classical Assumptions () Introductory Econometrics: Topic 5 1 / 94 The Multiple Regression Model: Freeing Up the Classical Assumptions Some or all of classical assumptions needed for derivations

More information

Statistics 910, #5 1. Regression Methods

Statistics 910, #5 1. Regression Methods Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known

More information

Lecture 4: Heteroskedasticity

Lecture 4: Heteroskedasticity Lecture 4: Heteroskedasticity Econometric Methods Warsaw School of Economics (4) Heteroskedasticity 1 / 24 Outline 1 What is heteroskedasticity? 2 Testing for heteroskedasticity White Goldfeld-Quandt Breusch-Pagan

More information

Heteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance.

Heteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance. Heteroskedasticity y i = β + β x i + β x i +... + β k x ki + e i where E(e i ) σ, non-constant variance. Common problem with samples over individuals. ê i e ˆi x k x k AREC-ECON 535 Lec F Suppose y i =

More information

A Course in Applied Econometrics Lecture 7: Cluster Sampling. Jeff Wooldridge IRP Lectures, UW Madison, August 2008

A Course in Applied Econometrics Lecture 7: Cluster Sampling. Jeff Wooldridge IRP Lectures, UW Madison, August 2008 A Course in Applied Econometrics Lecture 7: Cluster Sampling Jeff Wooldridge IRP Lectures, UW Madison, August 2008 1. The Linear Model with Cluster Effects 2. Estimation with a Small Number of roups and

More information

Econ 582 Fixed Effects Estimation of Panel Data

Econ 582 Fixed Effects Estimation of Panel Data Econ 582 Fixed Effects Estimation of Panel Data Eric Zivot May 28, 2012 Panel Data Framework = x 0 β + = 1 (individuals); =1 (time periods) y 1 = X β ( ) ( 1) + ε Main question: Is x uncorrelated with?

More information

Instrumental Variables

Instrumental Variables Università di Pavia 2010 Instrumental Variables Eduardo Rossi Exogeneity Exogeneity Assumption: the explanatory variables which form the columns of X are exogenous. It implies that any randomness in the

More information

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9. Section 7 Model Assessment This section is based on Stock and Watson s Chapter 9. Internal vs. external validity Internal validity refers to whether the analysis is valid for the population and sample

More information

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE Chapter 6. Panel Data Joan Llull Quantitative Statistical Methods II Barcelona GSE Introduction Chapter 6. Panel Data 2 Panel data The term panel data refers to data sets with repeated observations over

More information

Comprehensive Examination Quantitative Methods Spring, 2018

Comprehensive Examination Quantitative Methods Spring, 2018 Comprehensive Examination Quantitative Methods Spring, 2018 Instruction: This exam consists of three parts. You are required to answer all the questions in all the parts. 1 Grading policy: 1. Each part

More information

Applied Quantitative Methods II

Applied Quantitative Methods II Applied Quantitative Methods II Lecture 4: OLS and Statistics revision Klára Kaĺıšková Klára Kaĺıšková AQM II - Lecture 4 VŠE, SS 2016/17 1 / 68 Outline 1 Econometric analysis Properties of an estimator

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, 2016-17 Academic Year Exam Version: A INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This

More information

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han Econometrics Honor s Exam Review Session Spring 2012 Eunice Han Topics 1. OLS The Assumptions Omitted Variable Bias Conditional Mean Independence Hypothesis Testing and Confidence Intervals Homoskedasticity

More information

Dealing With Endogeneity

Dealing With Endogeneity Dealing With Endogeneity Junhui Qian December 22, 2014 Outline Introduction Instrumental Variable Instrumental Variable Estimation Two-Stage Least Square Estimation Panel Data Endogeneity in Econometrics

More information

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N 1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the

More information

Properties of the least squares estimates

Properties of the least squares estimates Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares

More information

Lecture 4: Testing Stuff

Lecture 4: Testing Stuff Lecture 4: esting Stuff. esting Hypotheses usually has three steps a. First specify a Null Hypothesis, usually denoted, which describes a model of H 0 interest. Usually, we express H 0 as a restricted

More information

Homoskedasticity. Var (u X) = σ 2. (23)

Homoskedasticity. Var (u X) = σ 2. (23) Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This

More information

LECTURE 18: NONLINEAR MODELS

LECTURE 18: NONLINEAR MODELS LECTURE 18: NONLINEAR MODELS The basic point is that smooth nonlinear models look like linear models locally. Models linear in parameters are no problem even if they are nonlinear in variables. For example:

More information

Asymptotic Theory. L. Magee revised January 21, 2013

Asymptotic Theory. L. Magee revised January 21, 2013 Asymptotic Theory L. Magee revised January 21, 2013 1 Convergence 1.1 Definitions Let a n to refer to a random variable that is a function of n random variables. Convergence in Probability The scalar a

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

Section 3: Simple Linear Regression

Section 3: Simple Linear Regression Section 3: Simple Linear Regression Carlos M. Carvalho The University of Texas at Austin McCombs School of Business http://faculty.mccombs.utexas.edu/carlos.carvalho/teaching/ 1 Regression: General Introduction

More information

ECO Class 6 Nonparametric Econometrics

ECO Class 6 Nonparametric Econometrics ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................

More information

BIOS 6649: Handout Exercise Solution

BIOS 6649: Handout Exercise Solution BIOS 6649: Handout Exercise Solution NOTE: I encourage you to work together, but the work you submit must be your own. Any plagiarism will result in loss of all marks. This assignment is based on weight-loss

More information

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that Linear Regression For (X, Y ) a pair of random variables with values in R p R we assume that E(Y X) = β 0 + with β R p+1. p X j β j = (1, X T )β j=1 This model of the conditional expectation is linear

More information

Problem Set - Instrumental Variables

Problem Set - Instrumental Variables Problem Set - Instrumental Variables 1. Consider a simple model to estimate the effect of personal computer (PC) ownership on college grade point average for graduating seniors at a large public university:

More information

Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data

Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data Panel data Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data - possible to control for some unobserved heterogeneity - possible

More information

Probability. Hosung Sohn

Probability. Hosung Sohn Probability Hosung Sohn Department of Public Administration and International Affairs Maxwell School of Citizenship and Public Affairs Syracuse University Lecture Slide 4-3 (October 8, 2015) 1/ 43 Table

More information

We begin by thinking about population relationships.

We begin by thinking about population relationships. Conditional Expectation Function (CEF) We begin by thinking about population relationships. CEF Decomposition Theorem: Given some outcome Y i and some covariates X i there is always a decomposition where

More information

Lecture 4: Multivariate Regression, Part 2

Lecture 4: Multivariate Regression, Part 2 Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables. Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Chapter 12 - Lecture 2 Inferences about regression coefficient

Chapter 12 - Lecture 2 Inferences about regression coefficient Chapter 12 - Lecture 2 Inferences about regression coefficient April 19th, 2010 Facts about slope Test Statistic Confidence interval Hypothesis testing Test using ANOVA Table Facts about slope In previous

More information

IV estimators and forbidden regressions

IV estimators and forbidden regressions Economics 8379 Spring 2016 Ben Williams IV estimators and forbidden regressions Preliminary results Consider the triangular model with first stage given by x i2 = γ 1X i1 + γ 2 Z i + ν i and second stage

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science

UNIVERSITY OF TORONTO Faculty of Arts and Science UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information