Problem Set 2
|
|
- Charles Jonas Summers
- 6 years ago
- Views:
Transcription
1 Problem Set 2 Paul Schrimpf Due: Friday, March 14, 2008 Problem 1 IV Basics Consider the model: y = Xβ + ɛ where X is T K. You are concerned that E[ɛ X] 0. Luckily, you have M > K variables, Z, which you think might serve as instruments. Solution to this question was written by Konrad last year. (a) What properties are required for Z to be a valid instrument? What is the exclusion restric- Question: tion? Answer: Z must satisfy two conditions: it must (1) have sufficient covariation with X in order to satisfy the rank condition ( rank plim 1 ) T Z X = rank(q ZX ) = k and (2) be uncorrelated with the disturbance ε, i.e. (and this is the exclusion restriction) (b) E[Z ε] = 0 Question: What is the IV estimator using Z as an instrument? What is its asymptotic distribution? Assume E[ɛɛ Z] = Σ. Answer In order to estimate β we construct instruments W = ZA T, where A T is a full-rank m k matrix that converges in probability to a limit A (I m allowing A T to be stochastic in order to nest the 2SLS case in the next part of the problem). Then we can form the usual IV estimator using W as instruments, ˆβ AT = (W X) 1 W y = (A TZ X) 1 A TZ y = β + (A TZ X) 1 A TZ ε From the exclusion restriction, plim ˆβ A = β for any A, and by the usual arguments, plim 1 T A TZ X = A Q ZX 1
2 e.g. from the Chebychev LLN. From an appropriate CLT (e.g. Lindeberg-Levy if we only have heteroskedasticity but individuals are i.i.d. draws from one stable distribution, or a CLT for mixing processes if we have (not too much) autocorrelation) and the Cramér-Wold device where by the law of iterated expectations T V 1 2 T Z ε = N (0, I k ) V T = 1 T E[Z εε Z] = 1 T E[Z E[εε Z]Z] = 1 T E[Z ΣZ] Since A T A, we can apply the continuous mapping theorem to the asymptotic variance of the IV estimator in order to obtain ( plimt Var( ˆβ A AT ) = plim T Z ) 1 X A ( T Z ΣZA T X ) 1 ZA T = (A Q ZX) 1 A Q ZΣZA(Q T T T ZXA) 1 where Q ZΣZ := plim 1 T E[Z ΣZ]. (c) Question: Suppose we form an instrument, Ŵ = ZÂ, where  = (Z Z) 1 Z X, so Ŵ is T K. Show that ˆβ 2SLS = (Ŵ X) 1 Ŵ y is consistent. Calculate the asymptotic variance of ˆβ 2SLS. Answer: 2SLS is a special case of the IV estimator discussed above, and therefore it is consistent whenever the rank condition is satisfied, and by plugging in, the variance becomes (d) ( T Var( ˆβ X Z(Z Z) 1 Z ) 1 X Z ( ΣZ X Z(Z Z) 1 Z X 2SLS ) = T T T Question: Suppose that ˆΩ is a consistent estimator of the limit of Z ΣZ/T. Consider the following minimum chi-squared estimator: ˆβ χ = arg min(y Xβ) Z ˆΩ 1 Z (y Xβ) β Show that ˆβ χ = (à Z X) 1 (à Z y) (so ˆβ χ is an IV estimator) for some Ã. ) 1 Answer: The first-order condition for this minimization problem is X Z ˆΩ 1 Z (y Xβ) = 0 ˆβ IV = (X Z ˆΩ 1 ZX) 1 X Z ˆΩ 1 Z y which fits into the framework in (b) by choosing à = ˆΩ 1 ZX, which corresponds to 2SLS after a GLS transformation of the data. (e) Question: Show that the à is optimal in minimizing the asymptotic variance of IV. What does this imply about the asymptotic efficiency of ˆβ IV relative to 2SLS? (Hint: you can think of this as a linear model of the form Z y = Z Xβ + Z u and apply the Gauss-Markov theorem asymptotically.) 2
3 Answer: After the GLS transformation of the data by premultiplying all variables (including the constant) with (ˆΩ 1 ) 1 2, the linear model Z y = Z Xβ 0 + Z ε satisfies the Gauss-Markov assumptions, and therefore in the limit, 2SLS is BLUE (only asymptotically, there s still bias in finite samples!). Problem 2 Measurement Error This question comes from last year s midterm. Consider the simple linear model y i = β 1 + β 2 x i + ɛ i i = 1,..., n, where x i is unobserved but a measurement of x i of x i is observed with x i = x i + v i, where E(x i v i) = 0 and E(v i ɛ i ) = 0, E(x i ɛ i) = 0. (a) Question: What is the probability limit of the least squares estimator of β? How big would V ar(v i ) have to be relative to V ar(x i ) have to be for the plim of the least square estimator to be half the size of the true parameter. Answer: Let X i = x i x. The limit of ˆβ 2 is plim ˆβ 2 =β 2 + (plim 1 n X X) 1 (plim 1 n X (ɛ β 2 v)) ( ) =β 2 1 σ2 v σx 2 + σv 2 where σ 2 x = Var(x i ) and σ2 v = Var(v i ). Generally, we don t care very much about the constant in a regression, so it would be probably fine to answer this question without saying anything about ˆβ 1, but for completeness: ˆβ 1 =(1 nm x 1 n ) 1 (1 nm x y) where 1 n is an n 1 vector of ones, and M x is the orthogonal projection to x. Expanding gives: ˆβ 1 =β nm x (β 2 x + ɛ) n 1 nx(x x) 1 x 1 n 2 x +σ2 x ) p β 1 + β 2µ x µxβ 2(µ σx 2+µ2 x p β 1 1 µ2 x µ 2 x +σ2 x so ˆβ 1 is consistent. The reason is that even though x is measured with error, the constant is uncorrelated with x, so it s coefficient remains consistent. (b) Question: Suppose you have an instrumental variable z i for x i. What properties does z i need to have for consistent estimation in part (i)? Could your instrument be measured with error and still have a consistent estimator? 3
4 Answer: Like in problem 1, z must be correlated with x i, but uncorrelated with the error, ɛ i β 2 v i. This is possible even if z is measured with error, as long as the measurement error in z is uncorrelated with ɛ and the measurement error in x. (c) Question: You are worried that your instrumental variable may have a little bit of correlation with v. However, you are quite confident that Corr(x i, v i ) > Corr(z i, v i ). Can you make any conclusion that the IV estimator will be less inconsistent than the OLS estimator? Answer: From part (a), we know that ( ) plim ˆβ 2 O LS = β 2 1 σ2 v σx 2 + σv 2 = β(1 ρ xv ) where ρ xv is the correlation between x and v. Let s look at the plim of IV. Remove the mean from z and x. Since z and x are scalars, the choice of A in our IV estimator does not matter, so let s just take A = 1. Then, ˆβ IV =(Z X) 1 (Z y) =β 2 + (Z X) 1 (Z (ɛ β 2 v)) p β 2 (1 σ zv σ zx + σ zv ) p ρ zv β 2 (1 ) ρ zx + ρ zv So, the bias of IV depends on both the correlation of z and v, and the correlation of z and x. Regardless of ρ zv, if ρ zx is small, then ˆβ IV can be more biased than OLS. (d) Question: Suppose you have a nonclassical measurement error (E[x i v i] 0) but you know that Cov(x i, ɛ i ) = 0. What will be the plim of the least squares estimator of β? Show that you will have attenuation bias if Cov(x i, x i ) > 0 and V ar(x i) > V ar(x i ). Answer: The plim of ˆβ 2 is: plim ˆβ 2 =(plim 1 n X X) 1 (plim 1 n X X β 2 ) =β 2 cov(x, x ) Var(x) In general, the Cauchy-Schwarz inequality tells us that cov(x, y) 2 Var(x)Var(y). Then, since Var(x ) Var(x), we have cov(x,x ) Var(x) < 1, so there is attenuation bias. Problem 3 Misspecified First Stage Suppose you re interested in the effect of an educational policy on child outcomes. covariates, you have a model that looks like: After partialing out outcome i = β 1 policylevel i + ɛ i 4
5 where outcome is something like a standardized test score, and policylevel is something like the studentteacher ratio. For this exercise, think of the policy as a continuous choice variable. You re worried that policy level is endogenous. A school that has a lower student-teacher ratio is also likely to be higher quality in other, unobserved dimensions. You think that prior test scores are a valid instrument for the policy level. You think that prior test scores are related to the policy level through: policylevel i = α 1 priortest i + v i Sadly, you re wrong about the functional form of this equation. The truth is that policylevel i = α 0 priortest i p + v i that is, the policy level is actually a function of how far prior test scores are from some cutoff level. This question explores how misspecifying the first stage affects IV. Throughout, you may assume that E[ɛ priortest] = 0 and that ɛ i is i.i.d. The very last part of this question is especially difficult. You probably have not yet seen all the tools required to answer it, so don t worry if you can t. However, if you can answer the last question and prove your answer, you can skip all the other parts of this question. (a) Question: Suppose you had correctly specified the first stage and know p. You estimate β 1 using Z = priortest i p as an instrument. ˆβ IV = (A Z X) 1 (A Z y) where X = policylevel and y = outcome. Is this estimator consistent? What is its asymptotic variance? (You can just state your answer, no proof necessary.) Answer: Yes, this estimator is consistent because we know that E[ɛZ] = 0. As in problem 1, its asymptotic variance is: V ( ˆβ IV ) = (A Q ZX ) 1 A Q ZΣZ A(Q ZXA) 1 In this case, assuming that we have partialed out a constant so all variables are mean zero Q ZX = plim 1 N Z X =α 0 Var( priortest i p ) Q ZΣZ = plim 1 N Z ɛɛ Z =σ 2 Var( priortest i p ) so, V ( ˆβ IV ) = σ 2 α 2 0 Var( priortest i p ) (b) Question: estimate Now, suppose you choose the wrong first stage and use Z = priortest i as an instrument to β IV = (Ã Z X) 1 (Ã Z y) Is this estimator consistent? What is its asymptotic variance? 5
6 Answer: Yes, this estimator is consistent, provided that cov(x, Z) 0. As above, its asymptotic variance is given by the formula in question 1. As in part (a), if we substitute for Q ZX and Q ZΣZ we get V ( β IV ) = σ 2 α 2 1 Var(priortest i) (c) (i) Finite Sample Properties Question: How would you expect β IV to perform relative to ˆβ IV in small samples? Is either estimator unbiased? Would you prefer one over the other? Give some intuition for your answer. Answer: Neither IV estimator is unbiased. The reason is that E[ ˆβ IV β] = E [ (A Z X) 1 (A Z ɛ) ] Since X is correlated with ɛ, this expectation is generally non-zero. After spring break, we ll look at the finite sample bias of IV in greater detail. If we get the functional form wrong, the first stage may be substantially weaker (the denominator of the bias smaller), and as you will see this typically increases the finite-sample bias. (ii) Let s simulate the model to compare the two estimators. Let: x i = α z i z + u i + v i y i =βx i + u i + e i where α = β = 1. Let z, u, v, and e be i.i.d, N(0, 1) variables. Set z = 0.5. Simulate the model many times and compute ˆβ OLS, ˆβ2SLS = (X P Z X) 1 (X P Z y), and β 2SLS = (X P ZX) 1 (X P Zy). Report the MSE, median, and probability of reject H 0 : β = 1 for each of these estimators. Examine how the results depend on the sample size. Answer: From Konrad last year: here are my simulation results: OLS coefficient (median) IV coefficient, (median) OLS coefficient (iqr) IV coefficient, (iqr) OLS Rejection Rates IV Rejection Rates Here you can again see that the bias of OLS is unaffected by sample size, whereas the substantial median bias in the IV coefficient for N = 20 goes away as the sample size increases. Looking at the inter-quartile range, the IV estimator seems to have a standard error about 5 times the size of that of OLS, which is much more than in problem 5 even though most other magnitudes were comparable, and this is in fact a consequence of the bad approximation of the functional form of the first stage. In fact, if z were zero, this strategy breaks down completely: beta = OLS coefficient (median) IV coefficient, (median)
7 OLS coefficient (iqr) IV coefficient, (iqr) OLS Rejection Rates IV Rejection Rates You can see that IV has about the same bias as OLS, and furthermore, the interquartile range doesn t decrase at all even though from the first to the third column, we increased the sample size by a factor of 100! This is a consequence of having no first stage anymore, and again this should give you a taste for what the second half of the course is going to be about. (d) (i) Asymptotically Optimal Instruments Question: What choices of A and à minimize the asymptotic variances of ˆβ IV and β IV? Which estimator has a smaller asymptotic variance with this A and Ã? Answer: The easiest answer is that A does not matter, since in both cases, Z and X are scalars. If this were not the case, A would matter, but we already know from problem 1 that the optimal A is A = Ω 1 Z X. Here, since ɛ is iid, Ω = Z Zσ 2, so the optimal A is just (Z Z) 1 Z X, which gives us 2SLS. Below, we ll show that Z E[x z] yields the minimum asymptotic variance among instruments of the form Z = g(z). Applying this result, it must be that ˆβ IV has smaller variance. (ii) Question: Consider a more general setup where instead of the first stage given above you have E[X Z] = f(z) for some function. Among IV estimators of the form: what choice of g() and A is asymptotically efficient? ˆβ IV (g(z), A) =(A g(z) X) 1 (A g(z) y) Answer: For a given g(), we know that A = (g(z) g(z)) 1 g(z) X is optimal for the same reason as in part (d)(i). Alternatively, we know that A doesn t matter when x is a scalar and g(z) is scalar valued. Sticking to the scalar case and assuming all variables have been de-meaned, the asymptotic variance for ˆβ using g(z) as an instrument is: V (β g ) = σ 2 Var(g(z)) cov(x, g(z)) 2 We can rewrite cov(x, g(z)) as: cov(x, g(z)) =E[xg(z)] =E[E[x z]g(z)] =E[f(z), g(z)] = cov(f(z), g(z)) Now comes the tricky part. Observe that the set of functions that map z to R, equipped with innner product f, g = cov(f(z), g(z)) forms an inner product space. The Cauchy-Schwarz inequality tells us that: f, g 2 f, f g, g Therefor, cov(f(z), g(z)) 2 Var(f(z))Var(g(z)) and in particular, V (β g ) = σ 2 Var(g(z)) cov(f(z), g(z)) 2 V (β f ) = σ 2 1 Var(f(z)) 7
8 (iii) Question: Suppose you don t know f(z) = E[X Z] exactly, but know it up to a finite dimensional parameter, θ. Can you construct a feasible estimator that has the same asymptotic variance as the efficient estimator when f(z) is known? Suppose you don t know anything about f(z), except that it s a wellbehaved function. Is there a feasible estimator with the same asymptotic variance as the efficient estimator when f(z) is known? Answer: The answer is yes, you can construct a feasible efficient estimator in both cases. Newey (1990) showed this result in a slightly more general setting when f(z) is completely unknown. To begin with, let s suppose we have some estimate of f(z), ˆf(z), and look at the asymptotic distribution of ˆβ ˆf. For simplicity, we will stick with the case where x, z, and f(z) are all scalars. We have: ˆβ ˆf =( ˆf X) 1 ˆf y =β + ( ˆf X) 1 ˆf ɛ where ˆf is short for ˆf(z). Notice that we can always write x = E[x z] + v = f(z) + v where E[v z] = 0. Thus, we have ˆβ ˆf β = ˆf ɛ ˆf (f + v) f ˆβ ɛ + ( ˆf β = ˆf f) ɛ f (f + v) + ( ˆf f) (f + v) n( ˆβ ˆf β) = 1 n f ɛ + 1 n ( ˆf f) ɛ 1 n f (f + v) + 1 n ( ˆf f) (f + v) The first terms in the numerator and denominator are exactly what we would get if f were known. Therefore, ˆβ ˆf has the same distribution as ˆβ f if and only if and plim 1 n ( ˆf f) ɛ = 0 (1) plim 1 n ( ˆf f) (f + v) = 0 (2) Equation (2) will be satisfied, if ˆf converges to f in probability uniformly in z. That is, if for any ɛ > 0, there s an N such that P ( ˆf(z) f(z) > ɛ) < ɛ z Z where Z is the domain of z. When f is known up to a finite dimensional parameter, then we know that we can estimate θ consistently by doing nonlinear-least squares. In this case, ˆf(z) f(z) =f(z, ˆθ) f(z, θ) = f(z, θ) (ˆθ θ) θ so ˆf p f uniformly if f θ is bounded. More generally, in this case plim 1 n ( ˆf f) (f + v) = plim(ˆθ θ) 1 n f(z i, θ) (f(z i ) + v i ) n θ i=1 [ ] f(zi, θ) =0 E (f(z i ) + v i ) θ 8
9 [ ] so we just need E f(zi,θ) 1 θ (f(z i ) + v i ) to be finite. A similar argument shows that if n f θ ɛ converges in distribution, then equation (1) is satisfied. Therefore, in the parametric case (when f is known up a finite number of parameters) a sufficient condition (in addition to the usual conditions needed for IV) for ˆβ ˆf to be [ ( ) ] 2 asymptotically efficient is that E f θ is finite. When f is completely unknown, we must estimate it nonparametrically. The details of showing that (1) and (2) are satisfied depend on the particular nonparametric estimator we use. Problem 4 Inference with Difference in Differences In this question, you ll reproduce some of the results from Bertrand, Duflo, and Mullainathan (2004). You ll then explore how the bias and size corrected test of Hausman and Kuersteiner (2007) performs in this setup. On the course website, I posted data from the CPS merged outgoing rotation groups. I originally obtained the data from I merged the CPS-MORG data for all the years and kept a few of the variables. If you re curious, I posted the do-file I used. This is the same data as used by Bertrand, Duflo, and Mullainathan (2004). The variables in the data are: ˆ state state census code ˆ year ˆ age ˆ female 1 = female, 0 = male ˆ yrsedu years of education ˆ earnwke weekly earnings ˆ employed whether employed ˆ white 1=white, 0=non-white ˆ black 1=black, 0=non-black The data set is rather large. If it is annoyingly slow to work with, you might want to limit the sample in some way. Bertrand, Duflo, and Mullainathan just use prime age white women. They also aggregate the data to state-year means for most of their simulations (the ones where the first column of the table is labeled CPS agg). You will probably also want to aggregate the data to save time. You will need to aggregate the data to impliment Hausman and Kuersteiner s estimator. (a) Reproducing Bertrand, Duflo, and Mullainathan Question: Perform similar simulations as Bertrand, Duflo, and Mullainathan (2004), and reproduce their key results. More specifically, randomly simulate policy changes and use difference in differences to estimate the effect of these changes. For each simulation, first sample 50 states with replacement from the data. Then, simulate the policy changes by randomly picking a year between 1985 and Randomly choose half of the states to have been affected. Then, estimate y ist = α s + λ t + γt st + x ist β + ɛ ist Estimate γ using OLS with various types of standard errors. Specifically, consider the usual, homoskedastic, iid OLS standard errors, standard errors clustered at state and year, and standard errors clustered at state. (If you want, you can also try some of the other things Bertrand, Duflo, and Mullainathan did). Report the fraction of simulations where you reject H 0 : γ = 0. Do this under both γ = 0 and γ = Explore how the results vary with N (the number of states) and T (the number of years). 9
10 Answer: See the code I posted. Table 1 shows the results of the simulations. The results are similar to Bertrand, Duflo, and Mullainathan. OLS and OLS state-year clustered over-reject for large N and T. OLS state clustered works well for N=50 and 20, but less well for smaller N. (b) FGLS Question: Perform the same simulations, but estimate γ using FGLS. Compare the results to OLS. As in Hausman and Kuersteiner (or section IV.D of Bertrand, Duflo, and Mullainathan), assume that the autocorrelation process is the same in all states and there is no heteroskedasticity. Again, explore how the results depend on N and T. Answer: The presence of fixed effects makes doing FGLS somewhat tricky. In particular, the natural estimate of the variance matrix, S ˆΣ = ê s ê s will be rank T 1 instead of rank T. You can proceed as in Kiefer (1980). Specifically: s=1 1. Aggregate the data to state-year averages. As in Bertrand, Duflo, and Mullainathan, first regress the outcome on any demographics you want to control for. Then, take the mean of the residuals within each state and year. Call this y s,t. Using OLS, etimate: y s,t = α s + λ t + γt s,t + e s,t 2. Form ˆΣ = S s=1 êsê s where ê s is the T 1 vector of residuals from state s. ˆΣ is singular, so we need to take a generalized inverse instead of its inverse. Following Kiefer (1980), let S be the T 1 T 1 matrix consisting of the first T 1 rows and columns of ˆΣ. Let S = S Estimate ˆβ F GLS = (X (I s M 1t S 1 M 1t )X) 1 (X (I s M 1t S 1 M 1t )y), where M 1t = I t 1 T 1 T 1 T is the matrix that projects out fixed effects, and X consists of time fixed effects and the policy indicator, and y = [y 11, y 12,..., y 1T,..., y ST ] and X is similarly arranged. See the code for more details. The results are in Table 1. FGLS does not perform very well. It only has nearly correct size for N=50 and T =10. However, in these situations FGLS is dominated in terms of power by OLS state clustered. (c) Bias-corrected FGLS Since T is small, and the regression includes state fixed effects, estimates of the variance matrix will be biased. Hausman and Kuersteiner show how to correct this bias (page 7). Implement this correction and repeat the simulations in part (b). How much does this bias correction help? How large does T need to be for the bias to not matter? Hint: there were two things that tripped me up when I was trying to do this. First, Hausman and Kuersteiner sort the variables as time 1 state 1, time 1 state 2,... time 1 state S, time 2 state 1, etc. This is the opposite of what I had been doing. You ll need to either reorder your variables, or reverse the order of the kronecker products in Hausman and Kuersteiner s formulas. The other thing that was difficult was figuring out what was meant by V. V s is supposed to be a vector of all right hand side variables at all times, except for state fixed effects. Since, after aggregating, the only variables we have are a policy indicator and time fixed effects, if you naively include all of these, V will not be full. 10
11 rank. In our setup, V s should just be two variables, a constant, and an indicator for whether the policy was ever impleented. Once you figure out the two things mentioned in the hint, implementing the bias correction is just a matter of following Hausman and Kuersteiner s formulas. See the code for details. In the table, we that bias-corrected FGLS almost always does better than plain FGLS. (d) Size-corrected Tests Question: Hausman and Kuersteiner (2007) also propose a size correction for testing H 0 : γ = γ 0 for when N is small. 1 Repeat the simulations in the previous sub-question using size corrected tests. Discuss the results. Answer: Following the theorem 3.1 of Hausman and Kuersteiner, the size-corrected tests, replaces the usual t = 1.96 critical value with ( t sc = t 1 + (1 + ) t2 ) + 2(T 2) 2S Since t sc > t, we know that size and bias corrected FGLS will always have a smaller size and power than bias corrected FGLS. However, this is a good thing, since bias corrected FGLS tends to over-rejct. From the table, we see that the size correction works fairly well, except for N=6, T=5. I am fairly impressed by how well the size correction works. I expected the results to be much worse. One method of inference in difference in differences setups, which we have not talked about, is the synthetic control method of Abadie, Diamond, and Hainmueller (2007). Abadie presented this paper at the econometrics seminar recently, and it appears quite promising. If you ever estimate a difference in differences in your research, the Abadie paper is definitely worth reading. Table 1: Simulation Results Estimator N T Rejection Rate No Effect Rejection Rate 2% Effect OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected Notice that this setup falls under the special case in section 3 of the paper, so you can use the simpler formulas for the test given in theorem
12 FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected
13 OLS OLS state-year clustered OLS state clustered OLS aggregate FGLS FGLS bias-corrected FGLS bias and size corrected Based on 100 simulations 13
1 Motivation for Instrumental Variable (IV) Regression
ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data
More information1 Introduction to Generalized Least Squares
ECONOMICS 7344, Spring 2017 Bent E. Sørensen April 12, 2017 1 Introduction to Generalized Least Squares Consider the model Y = Xβ + ɛ, where the N K matrix of regressors X is fixed, independent of the
More informationMultiple Regression Analysis
Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,
More informationFinal Exam. Economics 835: Econometrics. Fall 2010
Final Exam Economics 835: Econometrics Fall 2010 Please answer the question I ask - no more and no less - and remember that the correct answer is often short and simple. 1 Some short questions a) For each
More informationCasuality and Programme Evaluation
Casuality and Programme Evaluation Lecture V: Difference-in-Differences II Dr Martin Karlsson University of Duisburg-Essen Summer Semester 2017 M Karlsson (University of Duisburg-Essen) Casuality and Programme
More informationGLS and FGLS. Econ 671. Purdue University. Justin L. Tobias (Purdue) GLS and FGLS 1 / 22
GLS and FGLS Econ 671 Purdue University Justin L. Tobias (Purdue) GLS and FGLS 1 / 22 In this lecture we continue to discuss properties associated with the GLS estimator. In addition we discuss the practical
More informationThe Simple Linear Regression Model
The Simple Linear Regression Model Lesson 3 Ryan Safner 1 1 Department of Economics Hood College ECON 480 - Econometrics Fall 2017 Ryan Safner (Hood College) ECON 480 - Lesson 3 Fall 2017 1 / 77 Bivariate
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationMultiple Equation GMM with Common Coefficients: Panel Data
Multiple Equation GMM with Common Coefficients: Panel Data Eric Zivot Winter 2013 Multi-equation GMM with common coefficients Example (panel wage equation) 69 = + 69 + + 69 + 1 80 = + 80 + + 80 + 2 Note:
More informationNon-Spherical Errors
Non-Spherical Errors Krishna Pendakur February 15, 2016 1 Efficient OLS 1. Consider the model Y = Xβ + ε E [X ε = 0 K E [εε = Ω = σ 2 I N. 2. Consider the estimated OLS parameter vector ˆβ OLS = (X X)
More informationEconometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague
Econometrics Week 8 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 25 Recommended Reading For the today Instrumental Variables Estimation and Two Stage
More informationIV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors
IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors Laura Mayoral IAE, Barcelona GSE and University of Gothenburg Gothenburg, May 2015 Roadmap Deviations from the standard
More informationThe BLP Method of Demand Curve Estimation in Industrial Organization
The BLP Method of Demand Curve Estimation in Industrial Organization 9 March 2006 Eric Rasmusen 1 IDEAS USED 1. Instrumental variables. We use instruments to correct for the endogeneity of prices, the
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More informationEconomics 582 Random Effects Estimation
Economics 582 Random Effects Estimation Eric Zivot May 29, 2013 Random Effects Model Hence, the model can be re-written as = x 0 β + + [x ] = 0 (no endogeneity) [ x ] = = + x 0 β + + [x ] = 0 [ x ] = 0
More informationProblem Set #6: OLS. Economics 835: Econometrics. Fall 2012
Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationIntroduction to Econometrics. Heteroskedasticity
Introduction to Econometrics Introduction Heteroskedasticity When the variance of the errors changes across segments of the population, where the segments are determined by different values for the explanatory
More informationEconometrics - 30C00200
Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business
More informationInstrumental Variables
Instrumental Variables Department of Economics University of Wisconsin-Madison September 27, 2016 Treatment Effects Throughout the course we will focus on the Treatment Effect Model For now take that to
More informationPanel Data Models. James L. Powell Department of Economics University of California, Berkeley
Panel Data Models James L. Powell Department of Economics University of California, Berkeley Overview Like Zellner s seemingly unrelated regression models, the dependent and explanatory variables for panel
More informationOrdinary Least Squares Regression
Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section
More informationSpatial Regression. 3. Review - OLS and 2SLS. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved
Spatial Regression 3. Review - OLS and 2SLS Luc Anselin http://spatial.uchicago.edu OLS estimation (recap) non-spatial regression diagnostics endogeneity - IV and 2SLS OLS Estimation (recap) Linear Regression
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationRegression with a Single Regressor: Hypothesis Tests and Confidence Intervals
Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More informationReview of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley
Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 6: Bias and variance (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 49 Our plan today We saw in last lecture that model scoring methods seem to be trading off two different
More informationIntermediate Econometrics
Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the
More informationLecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)
Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) 1 2 Model Consider a system of two regressions y 1 = β 1 y 2 + u 1 (1) y 2 = β 2 y 1 + u 2 (2) This is a simultaneous equation model
More informationF9 F10: Autocorrelation
F9 F10: Autocorrelation Feng Li Department of Statistics, Stockholm University Introduction In the classic regression model we assume cov(u i, u j x i, x k ) = E(u i, u j ) = 0 What if we break the assumption?
More informationEconometrics Master in Business and Quantitative Methods
Econometrics Master in Business and Quantitative Methods Helena Veiga Universidad Carlos III de Madrid Models with discrete dependent variables and applications of panel data methods in all fields of economics
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More informationECONOMICS 210C / ECONOMICS 236A MONETARY HISTORY
Fall 018 University of California, Berkeley Christina Romer David Romer ECONOMICS 10C / ECONOMICS 36A MONETARY HISTORY A Little about GLS, Heteroskedasticity-Consistent Standard Errors, Clustering, and
More informationECO375 Tutorial 8 Instrumental Variables
ECO375 Tutorial 8 Instrumental Variables Matt Tudball University of Toronto Mississauga November 16, 2017 Matt Tudball (University of Toronto) ECO375H5 November 16, 2017 1 / 22 Review: Endogeneity Instrumental
More informationTopic 6: Non-Spherical Disturbances
Topic 6: Non-Spherical Disturbances Our basic linear regression model is y = Xβ + ε ; ε ~ N[0, σ 2 I n ] Now we ll generalize the specification of the error term in the model: E[ε] = 0 ; E[εε ] = Σ = σ
More informationEC402 - Problem Set 3
EC402 - Problem Set 3 Konrad Burchardi 11th of February 2009 Introduction Today we will - briefly talk about the Conditional Expectation Function and - lengthily talk about Fixed Effects: How do we calculate
More informationRecent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data
Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)
More informationIntroduction to Econometrics Midterm Examination Fall 2005 Answer Key
Introduction to Econometrics Midterm Examination Fall 2005 Answer Key Please answer all of the questions and show your work Clearly indicate your final answer to each question If you think a question is
More informationIntroduction to Econometrics Final Examination Fall 2006 Answer Sheet
Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Please answer all of the questions and show your work. If you think a question is ambiguous, clearly state how you interpret it before
More informationStat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS
Stat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS 1a) The model is cw i = β 0 + β 1 el i + ɛ i, where cw i is the weight of the ith chick, el i the length of the egg from which it hatched, and ɛ i
More informationLinear Regression. Junhui Qian. October 27, 2014
Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency
More informationEconometrics II - EXAM Answer each question in separate sheets in three hours
Econometrics II - EXAM Answer each question in separate sheets in three hours. Let u and u be jointly Gaussian and independent of z in all the equations. a Investigate the identification of the following
More informationThe returns to schooling, ability bias, and regression
The returns to schooling, ability bias, and regression Jörn-Steffen Pischke LSE October 4, 2016 Pischke (LSE) Griliches 1977 October 4, 2016 1 / 44 Counterfactual outcomes Scholing for individual i is
More informationA Primer on Asymptotics
A Primer on Asymptotics Eric Zivot Department of Economics University of Washington September 30, 2003 Revised: October 7, 2009 Introduction The two main concepts in asymptotic theory covered in these
More informationNew Developments in Econometrics Lecture 11: Difference-in-Differences Estimation
New Developments in Econometrics Lecture 11: Difference-in-Differences Estimation Jeff Wooldridge Cemmap Lectures, UCL, June 2009 1. The Basic Methodology 2. How Should We View Uncertainty in DD Settings?
More informationReview of Econometrics
Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,
More informationQuestions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares
Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares L Magee Fall, 2008 1 Consider a regression model y = Xβ +ɛ, where it is assumed that E(ɛ X) = 0 and E(ɛɛ X) =
More informationShort Questions (Do two out of three) 15 points each
Econometrics Short Questions Do two out of three) 5 points each ) Let y = Xβ + u and Z be a set of instruments for X When we estimate β with OLS we project y onto the space spanned by X along a path orthogonal
More informationRegression #3: Properties of OLS Estimator
Regression #3: Properties of OLS Estimator Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #3 1 / 20 Introduction In this lecture, we establish some desirable properties associated with
More informationQuick Review on Linear Multiple Regression
Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,
More informationInstrumental Variables and the Problem of Endogeneity
Instrumental Variables and the Problem of Endogeneity September 15, 2015 1 / 38 Exogeneity: Important Assumption of OLS In a standard OLS framework, y = xβ + ɛ (1) and for unbiasedness we need E[x ɛ] =
More informationInstrumental Variables, Simultaneous and Systems of Equations
Chapter 6 Instrumental Variables, Simultaneous and Systems of Equations 61 Instrumental variables In the linear regression model y i = x iβ + ε i (61) we have been assuming that bf x i and ε i are uncorrelated
More informationApplied Econometrics (MSc.) Lecture 3 Instrumental Variables
Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Estimation - Theory Department of Economics University of Gothenburg December 4, 2014 1/28 Why IV estimation? So far, in OLS, we assumed independence.
More informationPeter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8
Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall
More informationCHAPTER 6: SPECIFICATION VARIABLES
Recall, we had the following six assumptions required for the Gauss-Markov Theorem: 1. The regression model is linear, correctly specified, and has an additive error term. 2. The error term has a zero
More informationRecitation 1: Regression Review. Christina Patterson
Recitation 1: Regression Review Christina Patterson Outline For Recitation 1. Statistics. Bias, sampling variance and hypothesis testing.. Two important statistical theorems: Law of large numbers (LLN)
More informationRegression and Statistical Inference
Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF
More informationA Significance Test for the Lasso
A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical
More informationWarwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation
Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation Michele Aquaro University of Warwick This version: July 21, 2016 1 / 31 Reading material Textbook: Introductory
More informationECON Introductory Econometrics. Lecture 16: Instrumental variables
ECON4150 - Introductory Econometrics Lecture 16: Instrumental variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 12 Lecture outline 2 OLS assumptions and when they are violated Instrumental
More informationFreeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94
Freeing up the Classical Assumptions () Introductory Econometrics: Topic 5 1 / 94 The Multiple Regression Model: Freeing Up the Classical Assumptions Some or all of classical assumptions needed for derivations
More informationStatistics 910, #5 1. Regression Methods
Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known
More informationLecture 4: Heteroskedasticity
Lecture 4: Heteroskedasticity Econometric Methods Warsaw School of Economics (4) Heteroskedasticity 1 / 24 Outline 1 What is heteroskedasticity? 2 Testing for heteroskedasticity White Goldfeld-Quandt Breusch-Pagan
More informationHeteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance.
Heteroskedasticity y i = β + β x i + β x i +... + β k x ki + e i where E(e i ) σ, non-constant variance. Common problem with samples over individuals. ê i e ˆi x k x k AREC-ECON 535 Lec F Suppose y i =
More informationA Course in Applied Econometrics Lecture 7: Cluster Sampling. Jeff Wooldridge IRP Lectures, UW Madison, August 2008
A Course in Applied Econometrics Lecture 7: Cluster Sampling Jeff Wooldridge IRP Lectures, UW Madison, August 2008 1. The Linear Model with Cluster Effects 2. Estimation with a Small Number of roups and
More informationEcon 582 Fixed Effects Estimation of Panel Data
Econ 582 Fixed Effects Estimation of Panel Data Eric Zivot May 28, 2012 Panel Data Framework = x 0 β + = 1 (individuals); =1 (time periods) y 1 = X β ( ) ( 1) + ε Main question: Is x uncorrelated with?
More informationInstrumental Variables
Università di Pavia 2010 Instrumental Variables Eduardo Rossi Exogeneity Exogeneity Assumption: the explanatory variables which form the columns of X are exogenous. It implies that any randomness in the
More informationInternal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.
Section 7 Model Assessment This section is based on Stock and Watson s Chapter 9. Internal vs. external validity Internal validity refers to whether the analysis is valid for the population and sample
More informationChapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE
Chapter 6. Panel Data Joan Llull Quantitative Statistical Methods II Barcelona GSE Introduction Chapter 6. Panel Data 2 Panel data The term panel data refers to data sets with repeated observations over
More informationComprehensive Examination Quantitative Methods Spring, 2018
Comprehensive Examination Quantitative Methods Spring, 2018 Instruction: This exam consists of three parts. You are required to answer all the questions in all the parts. 1 Grading policy: 1. Each part
More informationApplied Quantitative Methods II
Applied Quantitative Methods II Lecture 4: OLS and Statistics revision Klára Kaĺıšková Klára Kaĺıšková AQM II - Lecture 4 VŠE, SS 2016/17 1 / 68 Outline 1 Econometric analysis Properties of an estimator
More informationThis does not cover everything on the final. Look at the posted practice problems for other topics.
Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry
More informationWISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A
WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, 2016-17 Academic Year Exam Version: A INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This
More informationEconometrics Honor s Exam Review Session. Spring 2012 Eunice Han
Econometrics Honor s Exam Review Session Spring 2012 Eunice Han Topics 1. OLS The Assumptions Omitted Variable Bias Conditional Mean Independence Hypothesis Testing and Confidence Intervals Homoskedasticity
More informationDealing With Endogeneity
Dealing With Endogeneity Junhui Qian December 22, 2014 Outline Introduction Instrumental Variable Instrumental Variable Estimation Two-Stage Least Square Estimation Panel Data Endogeneity in Econometrics
More informationEcon 836 Final Exam. 2 w N 2 u N 2. 2 v N
1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the
More informationProperties of the least squares estimates
Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares
More informationLecture 4: Testing Stuff
Lecture 4: esting Stuff. esting Hypotheses usually has three steps a. First specify a Null Hypothesis, usually denoted, which describes a model of H 0 interest. Usually, we express H 0 as a restricted
More informationHomoskedasticity. Var (u X) = σ 2. (23)
Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This
More informationLECTURE 18: NONLINEAR MODELS
LECTURE 18: NONLINEAR MODELS The basic point is that smooth nonlinear models look like linear models locally. Models linear in parameters are no problem even if they are nonlinear in variables. For example:
More informationAsymptotic Theory. L. Magee revised January 21, 2013
Asymptotic Theory L. Magee revised January 21, 2013 1 Convergence 1.1 Definitions Let a n to refer to a random variable that is a function of n random variables. Convergence in Probability The scalar a
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationSection 3: Simple Linear Regression
Section 3: Simple Linear Regression Carlos M. Carvalho The University of Texas at Austin McCombs School of Business http://faculty.mccombs.utexas.edu/carlos.carvalho/teaching/ 1 Regression: General Introduction
More informationECO Class 6 Nonparametric Econometrics
ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................
More informationBIOS 6649: Handout Exercise Solution
BIOS 6649: Handout Exercise Solution NOTE: I encourage you to work together, but the work you submit must be your own. Any plagiarism will result in loss of all marks. This assignment is based on weight-loss
More informationThis model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that
Linear Regression For (X, Y ) a pair of random variables with values in R p R we assume that E(Y X) = β 0 + with β R p+1. p X j β j = (1, X T )β j=1 This model of the conditional expectation is linear
More informationProblem Set - Instrumental Variables
Problem Set - Instrumental Variables 1. Consider a simple model to estimate the effect of personal computer (PC) ownership on college grade point average for graduating seniors at a large public university:
More informationRepeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data
Panel data Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data - possible to control for some unobserved heterogeneity - possible
More informationProbability. Hosung Sohn
Probability Hosung Sohn Department of Public Administration and International Affairs Maxwell School of Citizenship and Public Affairs Syracuse University Lecture Slide 4-3 (October 8, 2015) 1/ 43 Table
More informationWe begin by thinking about population relationships.
Conditional Expectation Function (CEF) We begin by thinking about population relationships. CEF Decomposition Theorem: Given some outcome Y i and some covariates X i there is always a decomposition where
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationFor more information about how to cite these materials visit
Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/
More informationRandom vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.
Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just
More information401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.
401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis
More informationChapter 12 - Lecture 2 Inferences about regression coefficient
Chapter 12 - Lecture 2 Inferences about regression coefficient April 19th, 2010 Facts about slope Test Statistic Confidence interval Hypothesis testing Test using ANOVA Table Facts about slope In previous
More informationIV estimators and forbidden regressions
Economics 8379 Spring 2016 Ben Williams IV estimators and forbidden regressions Preliminary results Consider the triangular model with first stage given by x i2 = γ 1X i1 + γ 2 Z i + ν i and second stage
More informationUNIVERSITY OF TORONTO Faculty of Arts and Science
UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More information