Identi cation of Average E ects under Magnitude and Sign Restrictions on Confounding

Size: px
Start display at page:

Download "Identi cation of Average E ects under Magnitude and Sign Restrictions on Confounding"

Transcription

1 Identi cation of Average E ects under Magnitude and Sign Restrictions on Confounding Karim Chalak y Boston College August 1, 214 Abstract This paper studies measuring the average e ects of X on Y in structural systems by imposing magnitude and sign restrictions on confounding without requiring (conditional) exogeneity of causes, treatment, or instruments. e study the identi cation of covariateconditioned average random coe cients, average nonparametric discrete and marginal e ects, local and marginal treatment e ects as well as average treatment e ects for the population, treated, and untreated. e characterize the omitted variables bias, due to confounders U, of regression and instrumental variables methods for the identi cation of these various average e ects, thereby generalizing the classic linear regression omitted variable bias representation. Then, using proxies for U, we ask how do the average direct e ects of U on Y compare in magnitude and sign to those of U on. Exogeneity and proportional confounding are limit cases yielding full identi cation. Alternatively, the e ects of X on Y are partially identi ed in sharp bounded intervals if is su ciently sensitive to U, and sharp upper or lower bounds may obtain otherwise. After studying estimation and inference, we apply this method to study the nancial return to education and the black-white wage gap. Keywords: causality, confounding, endogeneity, omitted variable, partial identi cation, proxy. Department of Economics, Boston College (University of Virginia starting Fall 214). chalak@bc.edu. y Acknowledgments: I thank the participants in the Northwestern Junior Festival of Recent Developments in Microeconometrics, Harvard Causal Inference Seminar, 212 California Econometrics Conference, 213 BU-BC Green Line Econometrics Conference, 213 North American inter Meeting of the Econometric Society, New York Camp Econometrics VIII, 23rd annual meeting of the Midwest Econometrics Group, 9th Greater New York Metropolitan Area Econometrics Colloquium, Cowles Foundation Conference on Econometrics, and the seminars at BC, the Federal Reserve Bank of Cleveland, UCSD, UCLA, USC, University of Pittsburgh, IUPUI, University of isconsin-milwaukee, Oxford, Royal Holloway University of London, University of Leicester, UVA, and University of Montreal as well as Kate Antonovics, Andrew Beauchamp, Stéphane Bonhomme, Donald Cox, Julie Cullen, Stefan Hoderlein, Arthur Lewbel, Matthew Masten, and Elie Tamer for helpful comments. I thank Rossella Calvi, Daniel Kim, and Tao Yang for excellent research assistance. Any errors are the author s responsibility. 1

2 1 Introduction This paper studies identifying and estimating average causal e ects in structural systems by imposing restrictions on the magnitude and sign of confounding without requiring conditional exogeneity of causes, treatment, or instruments given covariates. In particular, we study the full (point) and partial identi cation of covariate-conditioned average random coe cients, average nonparametric discrete and marginal e ects, local and marginal treatment e ects as well as average treatment e ects for the population, treated, and untreated. To illustrate the paper s main ideas, consider a Mincer (1974) earning structural equation, frequently employed in empirical work (see e.g. discussion in Card, 1999), given by Y = Y + X + U Y ; (1a) where Y denotes the logarithm of hourly wage, X denotes observed determinants of wage including years of education, and the scalar U, commonly referred to as ability in the literature, denotes unobserved skill. Thus, both X and U are potential structural determinants (causes) of Y, albeit realizations of Y and X are observed whereas those of U are not. As discussed below, we emphasize that this paper s approach does not require a linear or parametric speci cation. Nevertheless, in order to introduce the main ideas in their simplest form, we let U be scalar and consider constant slope coe cients and Y for now but allow for a random intercept which may be correlated with X. e also leave implicit conditioning on covariates. Our Y object of interest here is, the vector of (average) direct e ects of the elements of X on Y (e.g. average nancial return to education). Because U is freely associated with X and may cause Y (e.g. education choices and wage may depend on ability), we say that U is an unobserved confounder and X is endogenous. The researcher observes realizations of a vector Z of potential instruments that are uncorrelated with Y but possibly freely correlated with ability U and therefore invalid. This allows for the possibility that a potential instrument for education, e.g. proximity to a college, may be correlated with ability U, e.g. due to unobserved parental characteristics or choices. e let Z and X have the same dimension; in particular, Z may equal X. Suppose that the researcher observes realizations of a proxy for U that is possibly error-laden and given by = + U ; (1b) where, for now, we consider a constant slope coe cient and random intercept which may be correlated with X. For example, may denote the logarithm of a test score commonly used as a proxy for ability, such as IQ (Intelligence Quotient) or K (Knowledge of the orld of ork). This parsimonious speci cation facilitates comparing the slope coe cients on U in the Y and equations while maintaining the commonly used log-level speci cation for 2

3 the wage equation. In particular, Y and denote respectively the semi-elasticities of wage and test score with respect to unobserved ability (i.e. 1 Y % and 1 % are the average approximate percentage changes in wage and test score respectively due directly to a unit or percentile increase in U). Alternatively, one could consider standardizing the variables in these two equations, in which case the slope coe cients on standardized ability denote standard deviation shifts in wage and test score respectively due to a standard deviation shift in ability. Let Z be uncorrelated with ; thus, the proxy is informative, in the sense that any correlation between Z and arises solely due to 1 U. De ne Z ~ Z E(Z). e then have E( ZY ~ ) = E( ZX ~ ) + E( ZU) ~ Y and E( Z ~ ) = E( ZU) ~ : Provided E( ZX ~ ) is nonsingular and 6=, we obtain = E( ~ ZX ) 1 E( ~ ZY ) E( ~ ZX ) 1 E( ~ Z ) Y : This expression for involves two linear instrumental variables (IV) regression estimands E( ZX ~ ) 1 E( ZY ~ ) and E( ZX ~ ) 1 E( Z ~ ). It also involves the unknown Y denoting the ratio of the (average) direct e ect of U on Y to that of U on. Importantly, the IV regression omitted variable bias (or inconsistency) E( ZX ~ ) 1 E( Z ~ ) Y in measuring is known up to this ratio. As we show, similar expressions for the average e ects of X on Y obtain, under suitable assumptions, in the cases of random slope coe cients and nonparametric e ects. e ask the following questions: 1. How does the average direct e ect of U on Y compare in magnitude to that of U on? 2. How does the average direct e ect of U on Y compare in sign to that of U on? The answers to these questions impose restrictions on the magnitude and sign of confounding which fully or partially identify the average e ects of X on Y. The paper does not require particular answers to these questions. Instead, it characterizes the mapping 2 from every possible answer to the corresponding identi cation region for the average e ects of X on Y. In particular, exogeneity is a limiting special case, which can obtain if the average direct e ect Y of U on Y is zero, yielding full (point) identi cation. Proportional confounding is another limiting case, in which the average direct e ect of U on Y equals a known proportion of that of U on, also yielding full identi cation. Alternatively, weaker restrictions on how the average direct e ect of U on Y compares in magnitude and/or sign to that of U on partially identify elements of, 1 More generally, we let U denote the vector of unobservables that drive Y or the proxies and are thought to be freely correlated with Z (conditional on covariates S), and we absorb into Y and respectively the unobserved drivers of Y and that are (conditionally) uncorrelated with Z. It then su ces to have as many proxies as confounders U. Recall that Z may equal X. 2 Leamer (1983) suggests the slogan the mapping is the message. 3

4 yielding sharp bounded intervals when the proxy is su ciently sensitive to the confounder U, and sharp lower or upper bounds otherwise. Sometimes, economic theory and evidence can provide guidance to answering these questions in particular contexts. For example, in the earning equation illustrative example, it may be reasonable to assume that, given the observables, wage is on average less elastic or sensitive to unobserved ability than the test score is, i.e. Y. Speci cally, given the observed characteristics, a change in U may, on average, directly cause a higher percentage change in the test score than in wage. Moreover, we sometimes further assume that ability, on average, directly a ects wage and the test score in the same direction, i.e. Y. These assumptions are in accord with several theoretical and empirical ndings. For instance, Cawley, Heckman, and Vytlacil (21) nd that the fraction of wage variance explained by measures of cognitive ability is modest and that personality traits are correlated with earnings primarily through schooling attainment. Provided that ability measures, such as IQ or K, are su ciently associated with unobserved ability U, this suggests that the average direct e ects of U on Y may be modest. Second, when ability is not revealed to employers, they may statistically discriminate based on observables such as education (see e.g. Altonji and Pierret, 21; Arcidiacono, Bayer, and Hizmo, 21). This also suggests a modest average direct e ect of U on Y. Third, the empirical ndings in this paper corroborate the assumption Y since allowing Y to be larger than often extends the estimated identi cation regions to include a negative average return to education and a black-white wage gap in favor of blacks, which is inconsistent with the general ndings in the literature. Last, the assumptions Y, Y, or Y 1, underlying partial identi cation, are a weakening of the commonly employed assumption of exogeneity. Speci cally, with Z and U freely (conditionally) correlated, exogeneity requires the coe cient Y on U to be zero; we allow but don t require this. More generally, this paper s approach doesn t require a linear or parametric speci cation; Sections 4 and 5 give key general results. In particular, Section 4 studies the structural system: Y = r(x; S; U; U Y ) and = q(s; U; U ), (2) where the vectors of unobservables U Y and of confounders U interact nonseparably with X and a vector of observed covariates S to drive Y according to the unknown nonparametric structural function r. Further, U interacts with the unobserved vector U and the covariates S to drive the vector of proxies according to the unknown function q. Unlike U Y and U, U may statistically depend on X given the covariates S. Here, we study the identi cation of conditional average discrete and marginal causal e ects of X on Y given covariates S under magnitude and sign restrictions on confounding. For example, in the nonseparable nonparametric case, one can apply a probability transformation, e.g. for continuous response, proxy, and confounder, to 4

5 rewrite the response and proxy equations such that U, Y, and have uniform distributions. The researcher can then contrast the average sign and magnitude of the percentile changes in the response and proxy due to a percentile change in the confounder. To illustrate, consider rst the additively separable case in which Y = r(x; S; U Y ) + U Y and = + U, (3) where Y is a vector of random coe cients that can depend on S and U Y and where is a vector, and a matrix, of random coe cients that can depend on S and U. hen = and 3 U Y?XjS, we obtain the speci cation for the Y equation studied in e.g. Altonji Y and Matzkin (25), Hoderlein and Mammen (27), and Imbens and Newey (29), yielding full identi cation of various average e ects of X on Y. Section 4 studies the full and partial identi cation of conditional average e ects of X on Y in systems with U Y?XjS = s but where U may depend on X given S, rst in the additively separable case in equations (3) with the random vector Y possibly nonzero, and second when the e ect of U on Y in the nonseparable equations (2) is possibly nonzero. In section 3, we focus on the special case in which r(x; S; U Y ) = r (S; U Y ) + P k j=1 X jr j (S; U Y ) Y + X, with Y and denoting random intercept and slope coe cients. In this case, we study the identi cation of conditional averages of under magnitude and sign restrictions on confounding, while allowing for X and the potential instruments Z to be freely (conditionally) correlated with U and for Y to be nonzero. Appendix A contains additional extensions in the random coe cients case to allow for a panel structure and proxies included in the Y equation. Last, Section 5 studies treatment e ects and augments equations (2) and (3) from Section 4 with a threshold crossing equation generating the binary treatment X: X = 1fU X (Z; S)g; where U X is an unobserved variable and the function is unknown. For example, when Y = in the separable equations (3) and (U X ; U Y )?ZjS, we obtain the speci cation for the X and Y equations studied e.g. in Imbens and Angrist (1994) and Heckman and Vytlacil (25). Section 5 studies the full and partial identi cation of conditional local and marginal treatment e ects as well as conditional average treatment e ects for the population, treated, and untreated, when (U X ; U Y )?ZjS = s but U may depend on Z given S in both the separable case in equations (3) with Y possibly nonzero and the nonseparable case in equations (2) when the e ect of U on Y may be nonzero. 3 Throughout, we use A?BjS to denote conditional independence as in Dawid (1979). Further, we write A?BjS = s to denote conditional independence at S = s. 5

6 This paper s method provides a simple alternative to the common practice which informally assumes that conditioning on proxies for confounders ensures conditional exogeneity. For example, consider the illustrative linear example in equations (1a, 1b) with Cov(X; ( Y ; ) ) =. Then conditioning on doesn t ensure that the coe cient on X from a regression of Y on (1; X ; ) identi es. In particular, substituting for scalar U gives Y = Y Y + X + Y ; (4) and the possible conditional correlation 4 between X and given leads to regression bias. Indeed, more generally, conditioning on may, but need not, attenuate the regression bias (see e.g. ickens, 1972; Battistin and Chesher, 29; and Ogburna and Vandereele, 212). Note, however, that this regression consistently estimates ( ; Y ) if one assumes that Cov(U; Y ) = and = so that the proxy = U is a perfect rescaling of U. e don t require these assumptions. Rather than conditioning on mismeasured proxies, this paper employs proxies to bound the nonparametric regression bias (see Section 4). Furthermore, this paper s method provides a practical alternative to IV methods when potential instruments may be weak or (conditionally) endogenous. In particular, we don t require Cov(Z; ( Y ; U) ) = in equations (1a, 1b). Instead we need only impose Cov(Z; ( Y ; ) ) =. Moreover, is under-identi ed in equation (4) from the linear illustrative example since Z and X have the same dimension and thus there are fewer exogenous instruments for (X ; ) than needed for full identi cation. Similar di culties arise in more general nonlinear cases. As discussed above, imposing sign and magnitude restrictions on confounding is a weakening of exogeneity. Manski and Pepper (2) employ alternative assumptions to bound nonparametric average e ects. In particular, they assume known bounds on the range of Y and that E[r(x; s; U; U Y )jz = z; S = s] is monotonic in z. They also consider the assumption that r is monotonic in x. Okumura and Usui (214) combine the last two assumptions for Z = X along with the assumption that r is concave in x. Also, several recent papers employ alternative assumptions to partially identify linear or parametric e ects of endogenous variables 5. For example, Altonji, Conley, Elder, and Taber (211) assume that the selection on unobservables occurs similarly to that on observables. Also, Reinhold and outersen (29) and Nevo and Rosen (212) assume that the correlation between the potential instrument and U and that between the endogenous variable and U have the same sign and then further assume that the potential instrument is less correlated with U than the endogenous variable is. Lewbel 4 From = + U, we have that is generally correlated with U given. Since X (or Z) and U are freely correlated, it follows that is generally correlated with X (or Z) given. 5 See also e.g. Klepper and Leamer (1984) and Leamer (1987) who study linear systems where the measurement errors in the explanatory variables are uncorrelated with the explanatory variables and with the disturbances in the explained equations (as well as sometimes with each other). 6

7 (212) restricts the covariance of Z and the product of heteroskedastic error terms. Bontemps, Magnac, and Maurin (212) provide additional examples and a general treatment of set identi ed linear models. e don t require the assumptions in these papers. Instead, we employ proxies to identify average e ects under magnitude and sign restrictions on confounding. Of course, which identifying assumption is more appropriate depends on the context. After deriving sharp identi cation regions for the average direct e ects of X on Y in Sections 3, 4, and 5 under magnitude and sign restrictions on confounding, Section 6 studies estimation and inference. Last, Section 7 applies these results to study the return to education and the black-white wage gap. Using the data in Card (1995), we employ restrictions on confounding ( Y 1 and Y 1) to partially identify in sharp bounded intervals the covariate-conditioned average nancial incremental return to each year of education as well as the average black-white wage gap. Importantly, we don t require that instruments or regressors are conditionally exogenous. Generally, we nd that regression estimates, which would be con- sistent under exogeneity ( Y = ), provide an upper bound on the average return to education (e.g. 19:5% for the return to the 16 th year) and black-white wage gap ( 17:8%) and that the regression-based bound estimates are generally narrower than the IV-based ones, with especially narrower con dence intervals. Note that if one assumes that the proxy (log(k )) is a perfect rescaling of U then Y is estimated to be :23 with 95% con dence interval (CI) [:141; :264]. Allowing for error-laden proxies, the regression-based estimated sharp identi cation region for the black-white wage gap under sign and magnitude restrictions on confounding ( Y 1) is relatively wide, [ 17:8%; 1:9%] with a 95% CI [ 21%; 5:4%]. Thus, under these weaker than exogeneity assumptions on confounding, this data set is inconclusive about the extent of discrimination in the labor market. In contrast, the average return to education for the black subpopulation may di er slightly from the nonblack subpopulation, if at all. Further, we nd nonlinearity in the return to education, with the 12 th, 16 th, and 18 th years, corresponding to obtaining a high school, college, and possibly a graduate degree, yielding a high average return. For example, under sign and magnitude restrictions on confounding, the estimated identi cation region for the average return to the 16 th year is [13:33%; 19:5%] with 95% CI [7:5%; 25:1%] whereas that for the 13 th year is [:7%; 7:8%] with 95% CI [ 3:4%; 11:6%]. This nonlinearity may partly explain why, contrary to the expected direction of ability bias, linear IV estimates of the average return to education often exceed linear regression estimates. In particular, both types of estimates are weighted averages of yearly incremental returns for di erent subpopulations and the large IV estimates may re ect the relatively high return to graduation years for the subpopulation whose graduation outcomes depends on instruments such as proximity to college (see e.g. Card 1995, 1999). Section 8 concludes and mathematical proofs are gathered in Appendix B. 7

8 2 Data Generation The next assumption de nes the data generating process. Assumption 1 (S.1) (i) Let M ( S ; Z ; X ; ; Y ) be a random vector with unknown 1 `1 k1 m1 11 distribution P 2 P. (ii) Let a structural system generate the unobserved vectors U and U Y of countable dimension and confounders U collected in L (U ; U Y ; U ), covariates S, potential l1 instruments Z, causes X, proxies, and response Y such that Y = r(x; S; U; U Y ) and = q(s; U; U ); where r and q are unknown real- and vector-valued measurable functions respectively and E(Y; ) < 1. Realizations of M are observed whereas those of L are not. S.1(i) de nes the notation for observables. S.1(ii) imposes structure on the data generating process. e distinguish between the observed (or measured) variables M and unobserved (or latent) variables L. The vectors of unobservables U Y and of confounders U may interact nonseparably with the causes of interest X and covariates S to impact the response Y according to the nonparametric structural function r. e allow but do not require the availability of covariates S; if these are absent, set S = 1. Further, we allow but don t require elements of S to directly a ect Y ; if r doesn t directly depend on S, these may nevertheless serve as conditioning variables. e observe realizations of a vector of proxies for U. e sometimes allow for and X to have common elements. Analogously to U Y, the vector U interacts with S and U to generate the proxies according to q. Last, we also observe realizations of a vector of potential instruments Z possibly equal to, or containing elements of, X. Importantly, unlike U and U Y, U may statistically depend on Z or X given S, thereby creating di culties for the identi cation of the e ects of X on Y. Thus, elements of Z may but need not be valid instruments since these may be included in the Y equation and are freely (conditionally) correlated with U. 2.1 Causal E ects For 6 (x; s; u; u y ) and (x ; s; u; u y ) in X S U U Y, the direct e ect of X on Y at (x; x ) given (s; u; u y ) is (x; x ; s; u; u y ) r(x ; s; u; u y ) r(x; s; u; u y ). Further, when r is di erentiable in a particular cause of interest, we set k = 1 to denote this cause by X and we subsume, without loss of generality, the remaining elements of X into S. Then (x; s; u; u r(x; s; u; u y) is the direct marginal e ect of X on Y at x given (s; u; u y ). If U enters r separably from X A b. 6 Throughout, for random vectors A and B, we denote the support of A by A and that of A given B = b by 8

9 then (x; s; u; u y ) = (x; s; u y ) does not depend on U. Further, if the e ect of X on Y is linear, (x; s; u y ) does not depend on X, and we obtain the random coe cient speci cation (S; U Y ). Similarly, for (x; s; u; u y ) and (x; s; u ; u y ) in X S U U Y, the direct e ect of U on Y at (u; u ) given (x; s; u y ) is Y (u; u ; x; s; u y ) r(x; s; u ; u y ) r(x; s; u; u y ). Further, for l = 1 and r di erentiable in u, the direct marginal e ect of U on Y at u given (x; s; u y ) is Y (u; x; s; u y r(x; s; u; y). If U enters r separably from X and its e ect on Y is linear, we obtain the random coe cient speci cation Y Y (S; U Y ). Analogously, for (s; u; u w ) and (s; u ; u w ) in S U U, the direct e ect of U on h, for h = 1; :::; m; are h (u; u ; s; u w ) q h (s; u ; u w ) q h (s; u; u w ) and the direct marginal e ect of U on h is h (u; s; u q h(s; u; u w ). Thus, causal e ects, such as, Y, and, are features of the structural system and can derive from economic theory whereas the observability of X and U is an empirical matter. In this paper, we re interested in measuring certain (conditional) averages of (x; x ; s; u; u y ) and (x; s; u; u y ) such as the conditional average direct e ect of X on Y at (x; x ) given X = x and S = s: (x; x jx ; s) E[(x; x ; s; U; U Y )jx = x ; S = s] E[r(x ; S; U; U Y ) r(x; S; U; U Y )jx = x ; S = s]: For instance, for binary treatment X, averaging (; 1j1; s) over the distribution of S given X = 1 gives the average treatment e ect on the treated (; 1j1). As another example, we re interested in measuring the conditional average direct marginal e ect of X on Y at x given S = s: (xjs) E[(x; s; U; U Y )js = r(x; s; U; U Y )js = s]: It s also useful to give a succinct notation for the average direct e ects of U on Y and. For example, Y (u; xjs) E[ Y (u; x; s; U Y )js = r(x; s; u; U Y )js = s]. Similarly, for scalar, (u; u js) E[ (u; u ; s; U )js = s] E[q(s; u ; U ) q(s; u; U )js = s]: 3 Identi cation of Average Random Coe cients hile the paper s method doesn t require a linear or parametric e ect of X on Y, we nd it instructive to begin our analysis of the identi cation of average e ects under magnitude and sign restrictions on confounding by studying linear structures with random coe cients. e 9

10 relax the linearity assumption when studying the identi cation of conditional average nonparametric e ects in Section 4 and of local, marginal, and average treatment e ects in Section 5. Speci cally, we impose the following assumption in Section 3. Assumption 2 (S.2) Linearity: Assume S.1 with Cov[Z; (Y; ) ] < 1 and Y = r(x; S; U; U Y ) = r (S; U Y ) + kx X j r j (S; U Y ) + j=1 and let the h th component q h of q be given by h = q h (U; U ) = q h; (S; U ) + so that stacking h, h = 1; :::; m; into gives lx U g r g (S; U Y ) Y + X + U Y ; g=1 lx U g q h;g (S; U ) h + U h ; g=1 = + U : e collect the random coe cients 7 in ( ; vec( ) ; Y ; ; Y ). m1 lm 11 k1 l1 Thus, under S.2, for each observation i in a sample, we have Y i = Y;i + X i i + U i Y;i and i = ;i + U i ;i. This allows for random e ects for each individual or unit and encompasses constant slope coe cients as a special case. e suppress the index i when referring to the population. In particular, is the vector of random direct e ects of the elements of X on Y. To simplify the exposition, we set S = 1 in Sections 3.2 and 3.3 and study identifying the average e ect E() of X on Y. Section 3.4 accommodates covariates S and studies the identi cation of the conditional average e ect (S) E(jS) of X on Y given S. 3.1 IV Regression Notation Throughout this paper, for a generic random vector A with nite mean, we write: A E(A) and A ~ A A: For example, we write E() and Y E( Y ). Further, we employ the following succinct notation for IV regression coe cients and residuals. For generic random vector A and vectors B and C of equal dimension with E( C ~ A ~ ) nite and E( C ~ B ~ ) nite and nonsingular, we let R A:BjC E( ~ C ~ B ) 1 E( ~ C ~ A ) and A:BjC ~ A ~ B R A:BjC 7 e slightly abuse notation in the linear case and collect in and Y the random direct e ects of the elements of X and U on Y. Similarly, we collect in the random direct e ects of the elements of U on those of. 1

11 denote the linear IV regression estimand and residual respectively, so that by construction E( C ~ A:BjC ) =. For example, for k = `, R Y:XjZ E( Z ~ X ~ ) 1 E( Z ~ Y ~ ) is the vector of slope coe cients associated with X in a linear IV regression of Y on (1; X ) using instruments (1; Z ). In the special case where B = C, we obtain the linear regression coe cients and residuals: R A:B R A:BjB E( B ~ B ~ ) 1 E( B ~ A ~ ) and A:B A:BjB A ~ B ~ R A:B : 3.2 Characterization and Full Identi cation e begin by characterizing and studying conditions for full identi cation. Section 3.3 studies partial identi cation of elements of under sign and magnitude restrictions on confounding. For illustration, consider the example from the Introduction with scalar proxy and confounder U and constant slope coe cients: Y = Y + X + UY and = + U : Using the IV regression succinct notation, recall that, provided E( Z ~ X ~ ) is nonsingular, 6=, and Cov(Z; ( Y ; ) ) =, we have: = R Y:XjZ R :XjZ Y : In particular, the IV regression (omitted variable) bias R Y :XjZ depends on the ratio Y of the (average) direct e ect of U on the response Y to that of U on the proxy. Theorem 3.1 extends this result to allow for vectors U and and for random slope coe cients. e set S = 1 here; Section 3.4 explicitly conditions on covariates. Theorem 3.1 Assume S.2 with S = 1, ` = k, m = l, and that (i) E( ~ Z ~ X ) and are nonsingular, (ii) Cov( Y ; Z) =, E( ~ jz; X) =, and E( ~ Y ju; Z) =, (iii) Cov( ; Z) = and E( ~ ju; Z) =. Let 1 Y then = R Y:XjZ R :XjZ : e let B R Y:XjZ = R:XjZ denote the IV regression bias (or inconsistency) in measuring. hen Z = X, Theorem 3.1 gives that = R Y:X R :X. Next, we discuss the conditions in Theorem 3.1. Condition (i) requires ` = k and m = l, with E( Z ~ X ~ ) and nonsingular. More generally, ` k and m l su ce for full or partial identi cation and one may use a weighted combination of the resulting moments. In particular, 11

12 having ` k + m may fully identify. For example, if ` = k + m and m = l then ( ; ) = R Y:(X ; ) jz provided E[ ~ Z( ~ X ; ~ ) ] is nonsingular 8. e do not require this many instruments here; as such is under-identi ed. In particular, we let ` = k, with Z possibly equal to X. Conditions (ii) and (iii) are implied by the assumption that the coe cients are mean independent 9 of (U; Z; X) or the stronger assumption that (U ; U Y )? (U; Z; X). Note that when, Y, and are constants, conditions (ii) and (iii) reduce to Cov(Z; ( Y ; ) ) =. In particular, condition (ii) imposes assumptions on the random coe cients in the Y equation. It requires that Z is uncorrelated with Y, is mean independent 1 of (Z; X), and Y is mean independent of (U; Z). Roughly speaking, (ii) isolates U as the source of the di culty in identifying. Had U been observed with ` = k +l and E( ~ Z( ~ X ; ~ U )) nonsingular, (ii) would permit identifying the average slope coe cients via IV regression. Note that linearity and condition (ii) can (indirectly) restrict how the return to education depends on ability U (see e.g. Card 1999). In the linear case, and if valid instruments are available, one can consider IV methods for the correlated random coe cient model, e.g. ooldridge (1997, 23) and Heckman and Vytlacil (1998). Similarly, linearity and condition (ii) can restrict how Y relates to Z (and X), e.g. in learning models where the return to ability can vary with experience and depend on educational attainment (e.g. Altonji and Pierret, 21; Arcidiacono, Bayer, and Hizmo, 21). Sections 4 and 5 weaken these restrictions in Theorem 3.1 and give identi cation results for the case in which X and U can interact to determine Y via the nonseparable speci cation in S.1. Importantly, however, the conditions in Theorem 3.1 do not restrict the joint distribution of (U; Z ; X ) other than requiring that E( ~ Z ~ X ) is nonsingular. In particular, Z and X can be freely correlated with U and thus endogenous. Condition (iii) imposes restrictions on the random coe cients in the equation. It requires that Z is uncorrelated with and that the elements of are mean independent of (U; Z). This linearly relates Cov(Z; ) to Cov(Z; U) via the matrix. Note that we do not directly restrict the dependence between and U. Nevertheless, even when is constant and is independent of U, recall from the Introduction that the coe cient on X from a regression of Y on (1; X ; ) need not identify. e note that one means for full identi cation generates su ciently many valid instruments by imposing restrictions involving the components of,, and U. For example, let U, 1, and 2 be scalars and suppose that Y = Y + X + U Y, 1 = 1 + U 1, and 2 = 2 + U 2 ; 8 For example, if Y and are constant, substituting for U gives Y = Y + X +. 9 In the linear case, having ( Y ; ) be mean independent of Z may fully identify by generating a su cient number of instruments as functions of Z. Indeed, under such stronger (mean) independence assumptions involving Z, one can dispense with linearity as we show in Sections 4 and 5. 1 e employ the unnecessary mean independence assumptions in conditions (ii) and (iii) of Theorem 3.1 because of their simple interpretation. However, E[ ZX ~ ] ~ =, E[ ZU ~ ~ Y ] =, and E[ ZU ~ ~ ] = su ce. 12

13 so that there are two proxies for U, with 1 ; 2 6= and Corr[(U; 2 ) ; ( Y ; 1 ) ] =. Then Y = Y Y X + Y 1 1 and ( ; Y 1 ) may be fully identi ed from an IV regression of Y on (1; X ; 1 ) using instruments (1; Z ; 2 ) (see e.g. Blackburn and Neumark, 1992). e don t require such restrictions here, allowing for example, for components of (e.g. test taking skills) to be correlated. (Appendix A studies the case of multiple proxies for U that are components of X.) To illustrate the consequences of Theorem 3.1, consider the example from the Introduction with scalar U and. Observe that R Y:XjZ fully identi es under exogeneity. In this case, the IV regression bias disappears either because U does not determine Y, and in particular Y =, or because Z and U are uncorrelated, and thus R :XjZ =. Alternatively, shape restrictions on the e ects of U on Y and can fully identify. In particular, this occurs under signed proportional confounding, in which case the sign of the ratio Y of the average direct e ect of U on Y to that of U on is known and its magnitude equals a known constant jdj. For example, under equiconfounding, jdj = 1, and U directly a ects Y and equally on average. In this case, is fully identi ed under positive ( Y = ) or negative ( Y = ) equiconfounding by = R Y :XjZ or = R Y +:XjZ respectively (see Chalak, 212). More generally, U may be a vector of potential confounders. Often, to each confounder U h corresponds a proxy h = h +U h h so that = +U with = diag( 1 ; :::; m ). In this case, = R Y:XjZ R :XjZ = RY:XjZ P m h=1 Y;h h R h :XjZ: As before, under exogeneity = R Y:XjZ whereas under e.g. positive equiconfounding Y;h h = 1 for h = 1; :::; m and P = R m Y:XjZ h=1 R h :XjZ. Corollary 3.2 extends these full identi cation results to allow for a general matrix, with possibly equal to a known vector d of constants. However, it is useful throughout to keep in mind the leading one-to-one case where is a diagonal matrix with straightforward interpretation. e use subscripts to denote vector elements. For example, j and R Y:XjZ;j are the j th elements of and R Y:XjZ, and h and d h are the h th elements of and d, respectively. Corollary 3.2 Assume the conditions of Theorem 3.1 and let j = 1; :::; k. (i) If B j = (exogeneity) then j = R Y:XjZ;j. (ii) If = d (signed proportional confounding) then j = P m h=1 d hr h :XjZ;j. R Y:XjZ;j Thus, it su ces for exogeneity that Y = or R :XjZ =. In particular, if one fails to reject the null hypothesis R :XjZ;j = against the alternative R :XjZ;j 6=, say via a t-test in the scalar proxy case, then one cannot reject, under Theorem 3.1 s assumptions, that R Y:XjZ;j identi es j. Further, signed proportional confounding with known jd h j and sign(d h ); h = 1; :::; m, point identi es. 13

14 3.3 Partial Identi cation In the absence of conditions leading to full identi cation, magnitude and sign restrictions on the average direct e ects of U on Y and partially identify the elements of. To illustrate, consider the return to education example from the Introduction with scalar U and proxy (log(k )). As discussed above, exogeneity ( Y = ) and signed proportional confounding ( Y = d ) are limit cases securing full identi cation. Next, we derive sharp identi cation regions for the elements of under weaker sign and magnitude restrictions. In particular, we ask how does the average direct e ect Y average e ect of U on. of U on Y compares in magnitude and sign to the Suppose that Y 1 so that the magnitude of the average direct e ect of U on Y is not larger than that of U on. Here, is, on average, at least as directly responsive to U than Y is. Assume further that the sign of is known. For example, suppose that so that U a ects Y and on average in the same direction. Thus, in this case 2 D = [; 1]. Then the expression for gives the following identi cation regions for j, j = 1; :::; k, which depend on the sign of R :XjZ;j : j 2 B j ([; 1] jr :XjZ;j ) = [R Y:XjZ;j ; R Y:XjZ;j R :XjZ;j ]; j 2 B j ([; 1] j R :XjZ;j ) = [R Y:XjZ;j R :XjZ;j ; R Y:XjZ;j ]: Symmetrically, if we maintain the magnitude restriction 1 and assume that so that 2 D = [ 1; ], we obtain j 2 B j ([ 1; ] jr :XjZ;j ) = [R Y:XjZ;j + R :XjZ;j ; R Y:XjZ;j ]; j 2 B j ([ 1; ] j R :XjZ;j ) = [R Y:XjZ;j ; R Y:XjZ;j + R :XjZ;j ]: Instead, if 1 Y, so that is on average at most as directly responsive to U than Y is, and, so that U a ects Y and on average in the same direction, then 2 D = [1; +1) and we obtain the following identi cation regions for j, j = 1; :::; k: j 2 B j ([1; +1) jr :XjZ;j ) = [R Y:XjZ;j R :XjZ;j ; +1); j 2 B j ([1; +1) j R :XjZ;j ) = ( 1; R Y:XjZ;j R :XjZ;j ]: Note that these identi cation regions exclude the IV estimand R Y:XjZ;j. Symmetrically, if we assume that 1 and so that 2 D = ( 1; 1], we obtain j 2 B j (( 1; 1] jr :XjZ;j ) = ( 1; R Y:XjZ;j + R :XjZ;j ]; j 2 B j (( 1; 1] j R :XjZ;j ) = [R Y:XjZ;j + R :XjZ;j ; +1]: 14

15 ider intervals obtain under either magnitude or sign (but not both) restrictions on the average direct e ects Y and. In particular, if Y ; is on average at least as directly responsive as Y is to U, 2 D = [ 1; 1], and j is partially identi ed as follows: j 2 B j ([ 1; 1]) = [R Y:XjZ;j R:XjZ;j ; RY:XjZ;j + R:XjZ;j ]: Note that B j ([ 1; 1]) is twice as large as B j ([; 1] jsign(r :XjZ;j )) or B j ([ 1; ] jsign(r :XjZ;j )). Also, the closer Z is to exogeneity, the smaller R :XjZ;j is, and the tighter these three identi cation regions are. Alternatively, if Y, is, on average, less directly responsive to U than Y is, 2 D = ( 1; 1] S [1; +1), and j 2 B j (( 1; 1] S [1; +1)) = ( 1; R Y:XjZ;j R:XjZ;j ] S [RY:XjZ;j + R:XjZ;j ; +1). In this case, the farther Z is from exogeneity, the larger R:XjZ;j is, and the more informative B j (( 1; 1] jsign(r :XjZ;j )), B j ([1; +1) jsign(r :XjZ;j )), and B j (( 1; 1] S [1; +1)) are. Alone, sign restrictions determine the direction of the IV regression omitted variable bias. In particular, we have B j (( 1; ] j R :XjZ;j ) = B j ([; +1) j R :XjZ;j ) = [R Y:XjZ;j ; +1); B j (( 1; ] jr :XjZ;j ) = B j ([; +1) j R :XjZ;j ) = ( 1; R Y:XjZ;j ]: The above identi cation regions for j, under magnitude or sign restrictions on confounding (or both) with scalars U and, are sharp. Thus, any point in these regions is feasible under the maintained assumptions. In particular, given S.2 and the distribution of the observables M, for each element b of B j (D j sign(r :XjZ;j )) or B j (D), one can construct constants d Y and d (which, being constant, satisfy the conditions on Y and in Theorem 3.1) such that d Y d 2 D. For example, for R :XjZ;j 6=, it su ces to let d Y 1 d = R :XjZ;j (R Y:XjZ;j b). These identi cation regions obtain in part by asking how the average direct e ects Y and compare in magnitude. If this comparison is ambiguous, a researcher may be more con dent imposing a lower bound d L and upper bound d H on Y so that 2 D = [d L ; d H ]. In this case, similar sharp identi cation regions, involving d L R :XjZ;j and d H R :XjZ;j, derive, with exogeneity or signed proportional confounding as limit cases. Further, suppose more generally that U is a vector and there is a proxy h = h + U h h for each confounder U h, h = 1; :::; m, so that = diag( 1 ; :::; m ). Then P = R m Y:XjZ h=1 h R h :XjZ;j and magnitude and/or sign restrictions on h = Y;h h regions B j ( m h=1 D h j sign h=1;:::;m Y;h 2 D h, h = 1; :::; m, yield the sharp identi cation (R h :XjZ;j)) for j, j = 1; :::; k, de ned next. Corollary 3.3 derives sharp identi cation regions for a general matrix and restrictions 11 h 2 D h = [d L;h; ; d H;h ], h = 1; :::; m (we allow for d L;h = 1 and/or d H;h = +1 yielding 11 Here, we focus on interval restrictions. One can consider other types of restrictions on h. 15

16 identi cation regions that are either half open intervals or the real line). To facilitate the exposition in subsequent sections, we let A R :XjZ so that A jh R h :XjZ;j. Also, when some elements of R :XjZ;j have di erent signs, we partition these, without loss of generality, such that R h :XjZ;j for h = 1; :::; g and R h :XjZ;j for h = g + 1; :::; m. Otherwise, if all the elements of R :XjZ;j have common sign, we omit the irrelevant inequalities and sums. Corollary 3.3 Assume the conditions of Theorem 3.1 and that h 2 D h = [d L;h; ; d H;h ], h = 1; :::; m. Then, for j = 1; :::; k, j 2 B j ( m h=1 D h j sign (A jh )) de ned as follows, and these h=1;:::;m bounds are sharp: B j ([d L;1; ; d H;1 ] ::: [d L;m; ; d H;m ] ja j1 ; :::; A jg ; A j g+1 ; :::; A jm ) = [R Y:XjZ;j P g h=1 A jhd L;h P m h=g+1 A jhd H;h ; R Y:XjZ;j P g h=1 A jhd H;h P m h=g+1 A jhd L;h ]: Note that these sharp identi cation regions may but need not contain R Y:XjZ;j. Sharp identi cation regions under either magnitude or sign restrictions (but not both) on confounding derive by setting the vectors d L and d H suitably. Further, note that di erent potential instruments or proxies may lead to di erent identi cation regions for j, in which case j is identi ed in the intersection of these regions, provided it is nonempty. 3.4 Conditioning on Covariates e extend the results in Sections 3.2 and 3.3 to accommodate conditioning on covariates S. For this, for a generic random vector A with E(A) nite, let A(S) E(AjS) and ~ A(S) A A(S): For example, for s 2 S, (s) E(jS = s) denotes the average direct e ects of X on Y given covariates S = s. Further, for generic random vectors B and C of equal dimension with E[ C(S)( ~ A ~ (S); B ~ (S))jS = s] nite and E( C(S) ~ B ~ (S)jS = s) nonsingular, we de ne the conditional linear IV regression estimand and residual R A:BjC (s) E( ~ C(S) ~ B (S)jS = s) 1 E( ~ C(S) ~ A (S)jS = s) and A:BjC(s) ~ A (s) ~ B (s)r A:BjC (s); so that by construction E( C(S) ~ A:BjC (S)jS = s) =. For example, for k = ` we write R Y:XjZ (s) E( Z(S) ~ X ~ (S)jS = s) 1 E( Z(S) ~ Y ~ (S)jS = s). hen B = C, we obtain R A:B (s) R A:BjB (s) and A:B (s) A:BjB (s): This notation reduces to that de ned in Section 3.1 when S is degenerate, S = 1, in which case we leave S implicit. Theorem 3.4 Assume S.2 with ` = k and m = l, and that, for s 2 S, 16

17 (i) E( ~ Z(S) ~ X (S)jS = s) and (s) are nonsingular, (ii) Cov( Y ; ZjS = s) =, E( ~ (S)jZ; X; S = s) =, and E( ~ Y (S)jU; Z; S = s) =, (iii) Cov( ; ZjS = s) = and E( ~ (S)jU; Z; S = s) =. Let (s) 1 (s) Y (s) then (s) = R Y:XjZ (s) R :XjZ (s) (s): Here, B(s) R Y:XjZ (s) (s) = R:XjZ (s) (s) denotes the conditional IV regression bias in measuring (s). hen Z = X, Theorem 3.4 gives (s) = R Y:X (s) R :X (s) (s): Conditions (ii) and (iii) in Theorem 3.4 are implied by the stronger (unnecessary under linearity) condition (U ; U Y )? (U; Z; X)jS. In particular, the covariate-conditioned conditions (ii) and (iii) in Theorem 3.4 weaken their unconditional uncorrelation and mean independence analogs in Theorem 3.1. Further, conditioning on covariates S may render Z closer to exogeneity and the conditional IV regression bias smaller. Theorem 3.4 reduces to Theorem 3.1 when S is degenerate, S = 1: If the conditions in Theorem 3.4 hold for almost every s 2 S then an expression for derives by averaging the expression for (s) over the distribution of S. If, in addition, the conditional average e ects (S), (S), and Y (S) are constant, so that the mean independence assumptions involving,, and Y in Theorem 3.4 hold unconditionally, the law of iterated expectations gives = E( Z(S) ~ X ~ (S)) 1 E( Z(S) ~ Y ~ (S)) E( Z(S) ~ X ~ (S)) 1 E( Z(S) ~ ~ (S)) : This expression involves two estimands from IV regressions of ~ Y (S) and ~ (S) respectively on ~X(S) using instrument ~ Z(S). Further, if the conditional expectations Z(S), X(S), (S), and Y (S) are a ne functions of S, we obtain = E( Z:S X:S) 1 E( Z:S Y:S) E( Z:S X:S) 1 E( Z:S :S) : Using partitioned regressions (Frisch and augh, 1933), the two residual-based IV estimands in the above expression for can be recovered from R Y:(X ;S ) j(z ;S ) and R :(X ;S ) j(z ;S ) as the coe cients associated with ~ X (i.e. as the coe cients on X in IV regressions of Y and respectively on (1; X ; S ) using instruments (1; Z ; S ) ). From Theorem 3.4, we have that j (s) is fully identi ed under conditional exogeneity (B j (s) = ) or conditional signed proportional confounding ( (s) = d(s), a vector of known or estimable functions of s). Otherwise, the expression for (s) can be used, along with sign and magnitude restrictions on elements of (s) 1 (s) Y (s), involving the conditional average direct e ects of U on Y and given S = s, to partially identify j (s). This yields the sharp 17

18 identi cation regions B j ( m h=1 D h(s) j sign (A jh (s))) for j (s) de ned as in Corollary 3.3 with h=1;:::;m D h (s) [d L;h (s); d H;h (s)] where d L;h (s) and d H;h (s) are known or estimable functions of s for h = 1; :::; m and with R Y:XjZ (s) replacing R Y:XjZ and A(s) R :XjZ (s) replacing A. Appendix A contains extensions of the results in Section 3 on identi cation of average coe cients under magnitude and sign restrictions on confounding. Section A.1 studies a panel structure with individual and time varying random coe cients without requiring xed e ects. Section A.2 studies cases where the proxies are a component X 1 of X, included in the Y equation. 4 Identi cation of Average Nonparametric E ects e extend the analysis in Section 3 by removing the linearity assumption S.2 and studying the identi cation of average nonparametric e ects, with Y = r(x; S; U; U Y ) as speci ed in S.1. As discussed in Section 2.1, here we study the identi cation of the conditional average direct e ect of X on Y at (x; x ) given X = x and S = s: (x; x jx ; s) E[r(x ; s; U; U Y ) r(x; s; U; U Y )jx = x ; S = s]; such as the conditional average e ect of the treatment on the treated (; 1j1; s) as well as the conditional average direct marginal e ect X on Y at x given X = x and S = s: (xjx; r(x; s; U; U Y )jx = x; S = s]; where we set k = 1 to label the cause of interest by X. This also enables studying the identi cation of averages of (x; x jx ; s) and (xjx; s) such as the average direct e ect of X on Y at (x; x ) given X = x, (x; x jx ) = E[ (x; x jx ; S)jX = x ], and the conditional average marginal e ect (s) = E[ (XjX; s)js = s]. In studying the identi cation of (x; x jx ; s) and (xjx; s), we nd it useful to employ the following notation for the di erence and derivative of a nonparametric regression. Speci cally, for generic random vectors A and B with E(A) nite and b and b in the support of B given covariates S = s, we let R N A:B(b; b ; s) E(A jb = b ; S = s) E(A jb = b; S = s): Further, when B is a scalar and the derivative exists, we de ne R N E(A jb = b; S = s): 18

19 4.1 Additive Separability e begin our analysis by studying the case in which U enters r separably. In particular, we impose the following assumption. e remove separability in Section 4.2. Assumption 3 (S.3) Additive Separability: Assume S.1 with lx Y = r(x; S; U; U Y ) = r(x; S; U Y ) + U g r g (S; U Y ) r(x; S; U Y ) + U Y ; g=1 and generated as in S.2: = + U ; and adjust the random coe cients ( ; vec( ) ; Y ) correspondingly. m1 lm l1 Under S.3, when Y = and U Y? XjS, we obtain the nonparametric speci cation for the Y equation, imposed e.g. in Altonji and Matzkin (25), Hoderlein and Mammen (27), and Imbens and Newey 12 (29), in which case certain average e ects of X on Y are point identi ed. Next, we maintain that U Y? XjS = s but allow U to freely (conditionally) depend on X, with Y possibly nonzero. e then employ proxies to fully or partially identify average e ects of X on Y under magnitude and sign restrictions on confounding. For s 2 S and x; x 2 X, note that S.3 and U Y? XjS = s give that (x; x jx ; s) = E[r(x ; s; U Y ) r(x; s; U Y )js = s] (x; x js): and (xjx; s) r(x; s; U Y )js = s] (xjs): Theorem 4.1 gives conditions under which (x; x js) and (xjs) depend on the unknown (s) 1 (s) Y (s), involving the conditional average direct e ects of U on Y and. Theorem 4.1 Assume S.3 with m = l and that, for s 2 S, (i) (s) is nonsingular, (ii) U Y? XjS = s and E( ~ Y (S)jU; X; S = s) =, (iii) E(~ (S)jX; S = s) = and E( ~ (S)jU; X; S = s) =. Let (s) 1 (s) Y (s). Then for x; x 2 X, (x; x js) = R N Y:X(x; x ; s) R N :X(x; x ; s) (s): 12 Similar to Imbens and Newey (29), one can consider covariates S 2 and a scalar unobserved S 1 recoverable from the choice equation X = q(z; S 2 ; S 1 ) with q monotonic in S 1, such that (U Y ; S 1 )? Z js 2, yielding U Y? XjS with S = (S 1 ; S 2). e allow but do not require this possibility. 19

20 Set k = 1 and suppose further E( jx = x; S = s) exists and is nite and (iv.b) for all x y nonempty open neighborhood r(xy ; s; u y ) exists for a.e N (x) X, a u y, and there is a function s (U Y ) with E( s (U Y )js = s) < 1 r(xy ; s; u y ) < s (u y ) for a.e. u y. Then (xjs) = RY:X(x; N s) R:X(x; N s) (s): Conditions (ii) and (iii) of Theorem 4.1 strengthen their counterparts in Theorem 3.4 with Z = X. They are implied by the stronger condition 14 (U ; U Y )? (U; X)jS. Note that (ii) allows for nonzero Y and thus weakens the assumption U Y? XjS with Y = often employed in the literature. Condition (iii) ensures that is an informative proxy so that the mean dependence of on X given S = s arises solely due to U. Condition (iv) ensures that derivatives exist E[r(x; s; U Y )js = s] r(x; s; U Y )js = s] (see e.g. hite and Chalak, 213). The conditional nonparametric regression biases for (x; x js) and (xjs) are B(x; x js) = R N :X(x; x ; s) (s) and B(xjs) = R N :X(x; s) (s): As in the linear case, (x; x js) and (xjs) are fully identi ed under conditional exogeneity or signed proportional confounding. In particular, if B(x; x js) = or B(xjs) = (conditional exogeneity) then (x; x js) = R N Y:X (x; x ; s) or (xjs) = R N Y:X (x; s). This obtains e.g. if Y (s) = or U is conditionally mean independent of X, E( ~ U(S)jX; S = s) =. Alternatively, if (s) = d(s) with d h (s), h = 1; :::; m, known or estimable (conditional signed proportional confounding) then (x; x js) = R N Y:X(x; x ; s) P m h=1 RN h :X(x; x ; s)d h (s) and (xjs) = R N Y:X(x; s) P m h=1 RN h :X(x; s)d h (s): From Theorem 4.1, observe that another avenue for full identi cation of (x; x js) or (xjs) is to impose m restrictions on (; js) or (js). For example, if m = l = 1 and one assumes that (x y ; x z js) = for x y ; x z 2 X, as occurs e.g. if a nondegenerate component of X is excluded from r and thus the Y equation, with R:X N (xy ; x z ; s) 6= then (s) = RN Y:X (xy ;x z ;s) and R:X N (xy ;x z ;s) (x; x js) is thus fully identi ed. Analogous restrictions fully identify (xjs). e don t require such restrictions. 13 In this context, the quali er almost every (a.e.) means that the condition can fail for u y belonging to a measurable set V having P [U Y 2 V j S = s] = (see e.g. hite and Chalak, 213). 14 Had U been observed, U Y? (U; X)jS would su ce to identify the average e ects of X on Y, as would (U; U Y )? XjS with U unobserved. Here U is unobserved and can depend on X. 2

Identification of Average Effects under Magnitude and Sign Restrictions on Confounding

Identification of Average Effects under Magnitude and Sign Restrictions on Confounding Identification of Average Effects under Magnitude and Sign Restrictions on Confounding Karim Chalak University of Virginia July 19, 2015 Abstract This paper studies measuring the average effects of X on

More information

Identification of Average Effects under Magnitude and Sign Restrictions on Confounding

Identification of Average Effects under Magnitude and Sign Restrictions on Confounding Identification of Average Effects under Magnitude and Sign Restrictions on Confounding Karim Chalak University of Virginia March 17, 2016 Abstract This paper studies measuring various average effects of

More information

Identification of Average Effects under Magnitude and Sign Restrictions on Confounding

Identification of Average Effects under Magnitude and Sign Restrictions on Confounding Identification of Average Effects under Magnitude and Sign Restrictions on Confounding Karim Chalak University of Virginia February 3, 2018 Abstract This paper studies measuring various average effects

More information

Economics 241B Estimation with Instruments

Economics 241B Estimation with Instruments Economics 241B Estimation with Instruments Measurement Error Measurement error is de ned as the error resulting from the measurement of a variable. At some level, every variable is measured with error.

More information

4.8 Instrumental Variables

4.8 Instrumental Variables 4.8. INSTRUMENTAL VARIABLES 35 4.8 Instrumental Variables A major complication that is emphasized in microeconometrics is the possibility of inconsistent parameter estimation due to endogenous regressors.

More information

Control Functions in Nonseparable Simultaneous Equations Models 1

Control Functions in Nonseparable Simultaneous Equations Models 1 Control Functions in Nonseparable Simultaneous Equations Models 1 Richard Blundell 2 UCL & IFS and Rosa L. Matzkin 3 UCLA June 2013 Abstract The control function approach (Heckman and Robb (1985)) in a

More information

Lecture Notes on Measurement Error

Lecture Notes on Measurement Error Steve Pischke Spring 2000 Lecture Notes on Measurement Error These notes summarize a variety of simple results on measurement error which I nd useful. They also provide some references where more complete

More information

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008 ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008 Instructions: Answer all four (4) questions. Point totals for each question are given in parenthesis; there are 00 points possible. Within

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

Identi cation of Positive Treatment E ects in. Randomized Experiments with Non-Compliance

Identi cation of Positive Treatment E ects in. Randomized Experiments with Non-Compliance Identi cation of Positive Treatment E ects in Randomized Experiments with Non-Compliance Aleksey Tetenov y February 18, 2012 Abstract I derive sharp nonparametric lower bounds on some parameters of the

More information

Testing for Regime Switching: A Comment

Testing for Regime Switching: A Comment Testing for Regime Switching: A Comment Andrew V. Carter Department of Statistics University of California, Santa Barbara Douglas G. Steigerwald Department of Economics University of California Santa Barbara

More information

Identifying Structural E ects in Nonseparable Systems Using Covariates

Identifying Structural E ects in Nonseparable Systems Using Covariates Identifying Structural E ects in Nonseparable Systems Using Covariates Halbert White UC San Diego Karim Chalak Boston College October 16, 2008 Abstract This paper demonstrates the extensive scope of an

More information

Nonparametric Identi cation of Regression Models Containing a Misclassi ed Dichotomous Regressor Without Instruments

Nonparametric Identi cation of Regression Models Containing a Misclassi ed Dichotomous Regressor Without Instruments Nonparametric Identi cation of Regression Models Containing a Misclassi ed Dichotomous Regressor Without Instruments Xiaohong Chen Yale University Yingyao Hu y Johns Hopkins University Arthur Lewbel z

More information

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43 Panel Data March 2, 212 () Applied Economoetrics: Topic March 2, 212 1 / 43 Overview Many economic applications involve panel data. Panel data has both cross-sectional and time series aspects. Regression

More information

ECONOMETRICS II (ECO 2401) Victor Aguirregabiria. Spring 2018 TOPIC 4: INTRODUCTION TO THE EVALUATION OF TREATMENT EFFECTS

ECONOMETRICS II (ECO 2401) Victor Aguirregabiria. Spring 2018 TOPIC 4: INTRODUCTION TO THE EVALUATION OF TREATMENT EFFECTS ECONOMETRICS II (ECO 2401) Victor Aguirregabiria Spring 2018 TOPIC 4: INTRODUCTION TO THE EVALUATION OF TREATMENT EFFECTS 1. Introduction and Notation 2. Randomized treatment 3. Conditional independence

More information

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Songnian Chen a, Xun Lu a, Xianbo Zhou b and Yahong Zhou c a Department of Economics, Hong Kong University

More information

IDENTIFICATION OF MARGINAL EFFECTS IN NONSEPARABLE MODELS WITHOUT MONOTONICITY

IDENTIFICATION OF MARGINAL EFFECTS IN NONSEPARABLE MODELS WITHOUT MONOTONICITY Econometrica, Vol. 75, No. 5 (September, 2007), 1513 1518 IDENTIFICATION OF MARGINAL EFFECTS IN NONSEPARABLE MODELS WITHOUT MONOTONICITY BY STEFAN HODERLEIN AND ENNO MAMMEN 1 Nonseparable models do not

More information

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria SOLUTION TO FINAL EXAM Friday, April 12, 2013. From 9:00-12:00 (3 hours) INSTRUCTIONS:

More information

Nonparametric Identi cation of Regression Models Containing a Misclassi ed Dichotomous Regressor Without Instruments

Nonparametric Identi cation of Regression Models Containing a Misclassi ed Dichotomous Regressor Without Instruments Nonparametric Identi cation of Regression Models Containing a Misclassi ed Dichotomous Regressor Without Instruments Xiaohong Chen Yale University Yingyao Hu y Johns Hopkins University Arthur Lewbel z

More information

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University Instructions: Answer all four (4) questions. Be sure to show your work or provide su cient justi cation for

More information

Identification with Imperfect Instruments

Identification with Imperfect Instruments Identification with Imperfect Instruments Aviv Nevo Adam Rosen The Institute for Fiscal Studies Department of Economics, UCL cemmap working paper CWP16/08 Identi cation with Imperfect Instruments Aviv

More information

Likelihood Ratio Based Test for the Exogeneity and the Relevance of Instrumental Variables

Likelihood Ratio Based Test for the Exogeneity and the Relevance of Instrumental Variables Likelihood Ratio Based est for the Exogeneity and the Relevance of Instrumental Variables Dukpa Kim y Yoonseok Lee z September [under revision] Abstract his paper develops a test for the exogeneity and

More information

We begin by thinking about population relationships.

We begin by thinking about population relationships. Conditional Expectation Function (CEF) We begin by thinking about population relationships. CEF Decomposition Theorem: Given some outcome Y i and some covariates X i there is always a decomposition where

More information

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails GMM-based inference in the AR() panel data model for parameter values where local identi cation fails Edith Madsen entre for Applied Microeconometrics (AM) Department of Economics, University of openhagen,

More information

Lecture 4: Linear panel models

Lecture 4: Linear panel models Lecture 4: Linear panel models Luc Behaghel PSE February 2009 Luc Behaghel (PSE) Lecture 4 February 2009 1 / 47 Introduction Panel = repeated observations of the same individuals (e.g., rms, workers, countries)

More information

1 Correlation between an independent variable and the error

1 Correlation between an independent variable and the error Chapter 7 outline, Econometrics Instrumental variables and model estimation 1 Correlation between an independent variable and the error Recall that one of the assumptions that we make when proving the

More information

Regressor Dimension Reduction with Economic Constraints: The Example of Demand Systems with Many Goods

Regressor Dimension Reduction with Economic Constraints: The Example of Demand Systems with Many Goods Regressor Dimension Reduction with Economic Constraints: The Example of Demand Systems with Many Goods Stefan Hoderlein and Arthur Lewbel Boston College and Boston College original Nov. 2006, revised Feb.

More information

PANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1

PANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1 PANEL DATA RANDOM AND FIXED EFFECTS MODEL Professor Menelaos Karanasos December 2011 PANEL DATA Notation y it is the value of the dependent variable for cross-section unit i at time t where i = 1,...,

More information

MC3: Econometric Theory and Methods. Course Notes 4

MC3: Econometric Theory and Methods. Course Notes 4 University College London Department of Economics M.Sc. in Economics MC3: Econometric Theory and Methods Course Notes 4 Notes on maximum likelihood methods Andrew Chesher 25/0/2005 Course Notes 4, Andrew

More information

Write your identification number on each paper and cover sheet (the number stated in the upper right hand corner on your exam cover).

Write your identification number on each paper and cover sheet (the number stated in the upper right hand corner on your exam cover). STOCKHOLM UNIVERSITY Department of Economics Course name: Empirical Methods in Economics 2 Course code: EC2402 Examiner: Peter Skogman Thoursie Number of credits: 7,5 credits (hp) Date of exam: Saturday,

More information

ECON0702: Mathematical Methods in Economics

ECON0702: Mathematical Methods in Economics ECON0702: Mathematical Methods in Economics Yulei Luo SEF of HKU January 14, 2009 Luo, Y. (SEF of HKU) MME January 14, 2009 1 / 44 Comparative Statics and The Concept of Derivative Comparative Statics

More information

Using Matching, Instrumental Variables and Control Functions to Estimate Economic Choice Models

Using Matching, Instrumental Variables and Control Functions to Estimate Economic Choice Models Using Matching, Instrumental Variables and Control Functions to Estimate Economic Choice Models James J. Heckman and Salvador Navarro The University of Chicago Review of Economics and Statistics 86(1)

More information

A Note on the Closed-form Identi cation of Regression Models with a Mismeasured Binary Regressor

A Note on the Closed-form Identi cation of Regression Models with a Mismeasured Binary Regressor A Note on the Closed-form Identi cation of Regression Models with a Mismeasured Binary Regressor Xiaohong Chen Yale University Yingyao Hu y Johns Hopkins University Arthur Lewbel z Boston College First

More information

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

1 The Multiple Regression Model: Freeing Up the Classical Assumptions 1 The Multiple Regression Model: Freeing Up the Classical Assumptions Some or all of classical assumptions were crucial for many of the derivations of the previous chapters. Derivation of the OLS estimator

More information

1. The Multivariate Classical Linear Regression Model

1. The Multivariate Classical Linear Regression Model Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The

More information

Simple Estimators for Semiparametric Multinomial Choice Models

Simple Estimators for Semiparametric Multinomial Choice Models Simple Estimators for Semiparametric Multinomial Choice Models James L. Powell and Paul A. Ruud University of California, Berkeley March 2008 Preliminary and Incomplete Comments Welcome Abstract This paper

More information

Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case

Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case Arthur Lewbel Boston College Original December 2016, revised July 2017 Abstract Lewbel (2012)

More information

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER By Donald W. K. Andrews August 2011 COWLES FOUNDATION DISCUSSION PAPER NO. 1815 COWLES FOUNDATION FOR RESEARCH IN ECONOMICS

More information

Limited Information Econometrics

Limited Information Econometrics Limited Information Econometrics Walras-Bowley Lecture NASM 2013 at USC Andrew Chesher CeMMAP & UCL June 14th 2013 AC (CeMMAP & UCL) LIE 6/14/13 1 / 32 Limited information econometrics Limited information

More information

11. Bootstrap Methods

11. Bootstrap Methods 11. Bootstrap Methods c A. Colin Cameron & Pravin K. Trivedi 2006 These transparencies were prepared in 20043. They can be used as an adjunct to Chapter 11 of our subsequent book Microeconometrics: Methods

More information

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Econometrics Week 8 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 25 Recommended Reading For the today Instrumental Variables Estimation and Two Stage

More information

Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case

Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case Arthur Lewbel Boston College December 2016 Abstract Lewbel (2012) provides an estimator

More information

Nonparametric Identi cation of a Binary Random Factor in Cross Section Data

Nonparametric Identi cation of a Binary Random Factor in Cross Section Data Nonparametric Identi cation of a Binary Random Factor in Cross Section Data Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College Original January 2009, revised March

More information

Part VII. Accounting for the Endogeneity of Schooling. Endogeneity of schooling Mean growth rate of earnings Mean growth rate Selection bias Summary

Part VII. Accounting for the Endogeneity of Schooling. Endogeneity of schooling Mean growth rate of earnings Mean growth rate Selection bias Summary Part VII Accounting for the Endogeneity of Schooling 327 / 785 Much of the CPS-Census literature on the returns to schooling ignores the choice of schooling and its consequences for estimating the rate

More information

Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation

Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation Xu Cheng y Department of Economics Yale University First Draft: August, 27 This Version: December 28 Abstract In this paper,

More information

Chapter 2. Dynamic panel data models

Chapter 2. Dynamic panel data models Chapter 2. Dynamic panel data models School of Economics and Management - University of Geneva Christophe Hurlin, Université of Orléans University of Orléans April 2018 C. Hurlin (University of Orléans)

More information

Regression. Econometrics II. Douglas G. Steigerwald. UC Santa Barbara. D. Steigerwald (UCSB) Regression 1 / 17

Regression. Econometrics II. Douglas G. Steigerwald. UC Santa Barbara. D. Steigerwald (UCSB) Regression 1 / 17 Regression Econometrics Douglas G. Steigerwald UC Santa Barbara D. Steigerwald (UCSB) Regression 1 / 17 Overview Reference: B. Hansen Econometrics Chapter 2.20-2.23, 2.25-2.29 Best Linear Projection is

More information

Instrumental Variables. Ethan Kaplan

Instrumental Variables. Ethan Kaplan Instrumental Variables Ethan Kaplan 1 Instrumental Variables: Intro. Bias in OLS: Consider a linear model: Y = X + Suppose that then OLS yields: cov (X; ) = ^ OLS = X 0 X 1 X 0 Y = X 0 X 1 X 0 (X + ) =)

More information

What do instrumental variable models deliver with discrete dependent variables?

What do instrumental variable models deliver with discrete dependent variables? What do instrumental variable models deliver with discrete dependent variables? Andrew Chesher Adam Rosen The Institute for Fiscal Studies Department of Economics, UCL cemmap working paper CWP10/13 What

More information

An Extended Class of Instrumental Variables for the Estimation of Causal E ects

An Extended Class of Instrumental Variables for the Estimation of Causal E ects An Extended Class of Instrumental Variables for the Estimation of Causal E ects Karim Chalak y Boston College Halbert White UC San Diego Current Draft: November 2009 First Draft: March 2005 Abstract We

More information

Chapter 1. GMM: Basic Concepts

Chapter 1. GMM: Basic Concepts Chapter 1. GMM: Basic Concepts Contents 1 Motivating Examples 1 1.1 Instrumental variable estimator....................... 1 1.2 Estimating parameters in monetary policy rules.............. 2 1.3 Estimating

More information

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations. Exercises for the course of Econometrics Introduction 1. () A researcher is using data for a sample of 30 observations to investigate the relationship between some dependent variable y i and independent

More information

Economics Introduction to Econometrics - Fall 2007 Final Exam - Answers

Economics Introduction to Econometrics - Fall 2007 Final Exam - Answers Student Name: Economics 4818 - Introduction to Econometrics - Fall 2007 Final Exam - Answers SHOW ALL WORK! Evaluation: Problems: 3, 4C, 5C and 5F are worth 4 points. All other questions are worth 3 points.

More information

Sensitivity checks for the local average treatment effect

Sensitivity checks for the local average treatment effect Sensitivity checks for the local average treatment effect Martin Huber March 13, 2014 University of St. Gallen, Dept. of Economics Abstract: The nonparametric identification of the local average treatment

More information

Non-parametric Identi cation and Testable Implications of the Roy Model

Non-parametric Identi cation and Testable Implications of the Roy Model Non-parametric Identi cation and Testable Implications of the Roy Model Francisco J. Buera Northwestern University January 26 Abstract This paper studies non-parametric identi cation and the testable implications

More information

Unconditional Quantile Regression with Endogenous Regressors

Unconditional Quantile Regression with Endogenous Regressors Unconditional Quantile Regression with Endogenous Regressors Pallab Kumar Ghosh Department of Economics Syracuse University. Email: paghosh@syr.edu Abstract This paper proposes an extension of the Fortin,

More information

Inference about Clustering and Parametric. Assumptions in Covariance Matrix Estimation

Inference about Clustering and Parametric. Assumptions in Covariance Matrix Estimation Inference about Clustering and Parametric Assumptions in Covariance Matrix Estimation Mikko Packalen y Tony Wirjanto z 26 November 2010 Abstract Selecting an estimator for the variance covariance matrix

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory

More information

When is it really justifiable to ignore explanatory variable endogeneity in a regression model?

When is it really justifiable to ignore explanatory variable endogeneity in a regression model? Discussion Paper: 2015/05 When is it really justifiable to ignore explanatory variable endogeneity in a regression model? Jan F. Kiviet www.ase.uva.nl/uva-econometrics Amsterdam School of Economics Roetersstraat

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Many economic models involve endogeneity: that is, a theoretical relationship does not fit

More information

Models, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley

Models, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley Models, Testing, and Correction of Heteroskedasticity James L. Powell Department of Economics University of California, Berkeley Aitken s GLS and Weighted LS The Generalized Classical Regression Model

More information

Markov-Switching Models with Endogenous Explanatory Variables. Chang-Jin Kim 1

Markov-Switching Models with Endogenous Explanatory Variables. Chang-Jin Kim 1 Markov-Switching Models with Endogenous Explanatory Variables by Chang-Jin Kim 1 Dept. of Economics, Korea University and Dept. of Economics, University of Washington First draft: August, 2002 This version:

More information

Multiple Linear Regression CIVL 7012/8012

Multiple Linear Regression CIVL 7012/8012 Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for

More information

Granger Causality and Dynamic Structural Systems

Granger Causality and Dynamic Structural Systems Granger Causality and Dynamic Structural Systems Halbert White and Xun Lu Department of Economics University of California, San Diego June 26, 2008 Abstract We analyze the relations between Granger (G)

More information

Econometrics Midterm Examination Answers

Econometrics Midterm Examination Answers Econometrics Midterm Examination Answers March 4, 204. Question (35 points) Answer the following short questions. (i) De ne what is an unbiased estimator. Show that X is an unbiased estimator for E(X i

More information

Nonparametric Welfare Analysis for Discrete Choice

Nonparametric Welfare Analysis for Discrete Choice Nonparametric Welfare Analysis for Discrete Choice Debopam Bhattacharya University of Oxford September 26, 2014. Abstract We consider empirical measurement of exact equivalent/compensating variation resulting

More information

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47 ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with

More information

DEPARTMENT OF ECONOMICS DISCUSSION PAPER SERIES

DEPARTMENT OF ECONOMICS DISCUSSION PAPER SERIES ISSN 1471-0498 DEPARTMENT OF ECONOMICS DISCUSSION PAPER SERIES DYNAMIC BINARY OUTCOME MODELS WITH MAXIMAL HETEROGENEITY Martin Browning and Jesus M. Carro Number 426 April 2009 Manor Road Building, Oxford

More information

The returns to schooling, ability bias, and regression

The returns to schooling, ability bias, and regression The returns to schooling, ability bias, and regression Jörn-Steffen Pischke LSE October 4, 2016 Pischke (LSE) Griliches 1977 October 4, 2016 1 / 44 Counterfactual outcomes Scholing for individual i is

More information

13 Endogeneity and Nonparametric IV

13 Endogeneity and Nonparametric IV 13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,

More information

1 Motivation for Instrumental Variable (IV) Regression

1 Motivation for Instrumental Variable (IV) Regression ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data

More information

An Instrumental Variable Model of Multiple Discrete Choice

An Instrumental Variable Model of Multiple Discrete Choice An Instrumental Variable Model of Multiple Discrete Choice Andrew Chesher y UCL and CeMMAP Adam M. Rosen z UCL and CeMMAP February, 20 Konrad Smolinski x UCL and CeMMAP Abstract This paper studies identi

More information

Chapter 6: Endogeneity and Instrumental Variables (IV) estimator

Chapter 6: Endogeneity and Instrumental Variables (IV) estimator Chapter 6: Endogeneity and Instrumental Variables (IV) estimator Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 15, 2013 Christophe Hurlin (University of Orléans)

More information

Parametric Inference on Strong Dependence

Parametric Inference on Strong Dependence Parametric Inference on Strong Dependence Peter M. Robinson London School of Economics Based on joint work with Javier Hualde: Javier Hualde and Peter M. Robinson: Gaussian Pseudo-Maximum Likelihood Estimation

More information

More on Specification and Data Issues

More on Specification and Data Issues More on Specification and Data Issues Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Specification and Data Issues 1 / 35 Functional Form Misspecification Functional

More information

IDENTIFICATION IN DIFFERENTIATED PRODUCTS MARKETS USING MARKET LEVEL DATA. Steven T. Berry and Philip A. Haile. January 2010 Revised May 2012

IDENTIFICATION IN DIFFERENTIATED PRODUCTS MARKETS USING MARKET LEVEL DATA. Steven T. Berry and Philip A. Haile. January 2010 Revised May 2012 IDENTIFICATION IN DIFFERENTIATED PRODUCTS MARKETS USING MARKET LEVEL DATA By Steven T. Berry and Philip A. Haile January 2010 Revised May 2012 COWLES FOUNDATION DISCUSSION PAPER NO. 1744R COWLES FOUNDATION

More information

Supplemental Material 1 for On Optimal Inference in the Linear IV Model

Supplemental Material 1 for On Optimal Inference in the Linear IV Model Supplemental Material 1 for On Optimal Inference in the Linear IV Model Donald W. K. Andrews Cowles Foundation for Research in Economics Yale University Vadim Marmer Vancouver School of Economics University

More information

Applied Health Economics (for B.Sc.)

Applied Health Economics (for B.Sc.) Applied Health Economics (for B.Sc.) Helmut Farbmacher Department of Economics University of Mannheim Autumn Semester 2017 Outlook 1 Linear models (OLS, Omitted variables, 2SLS) 2 Limited and qualitative

More information

The regression model with one stochastic regressor (part II)

The regression model with one stochastic regressor (part II) The regression model with one stochastic regressor (part II) 3150/4150 Lecture 7 Ragnar Nymoen 6 Feb 2012 We will finish Lecture topic 4: The regression model with stochastic regressor We will first look

More information

Sharp identified sets for discrete variable IV models

Sharp identified sets for discrete variable IV models Sharp identified sets for discrete variable IV models Andrew Chesher Konrad Smolinski The Institute for Fiscal Studies Department of Economics, UCL cemmap working paper CWP11/10 Sharp identi ed sets for

More information

Bias from Censored Regressors

Bias from Censored Regressors Bias from Censored Regressors Roberto Rigobon Thomas M. Stoker September 2005 Abstract We study the bias that arises from using censored regressors in estimation of linear models. We present results on

More information

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 Revised March 2012

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 Revised March 2012 SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER By Donald W. K. Andrews August 2011 Revised March 2012 COWLES FOUNDATION DISCUSSION PAPER NO. 1815R COWLES FOUNDATION FOR

More information

Econometrics II. Nonstandard Standard Error Issues: A Guide for the. Practitioner

Econometrics II. Nonstandard Standard Error Issues: A Guide for the. Practitioner Econometrics II Nonstandard Standard Error Issues: A Guide for the Practitioner Måns Söderbom 10 May 2011 Department of Economics, University of Gothenburg. Email: mans.soderbom@economics.gu.se. Web: www.economics.gu.se/soderbom,

More information

Föreläsning /31

Föreläsning /31 1/31 Föreläsning 10 090420 Chapter 13 Econometric Modeling: Model Speci cation and Diagnostic testing 2/31 Types of speci cation errors Consider the following models: Y i = β 1 + β 2 X i + β 3 X 2 i +

More information

Instrumental Variable Models for Discrete Outcomes. Andrew Chesher Centre for Microdata Methods and Practice and UCL. Revised November 13th 2008

Instrumental Variable Models for Discrete Outcomes. Andrew Chesher Centre for Microdata Methods and Practice and UCL. Revised November 13th 2008 Instrumental Variable Models for Discrete Outcomes Andrew Chesher Centre for Microdata Methods and Practice and UCL Revised November 13th 2008 Abstract. Single equation instrumental variable models for

More information

Empirical Methods in Applied Microeconomics

Empirical Methods in Applied Microeconomics Empirical Methods in Applied Microeconomics Jörn-Ste en Pischke LSE November 2007 1 Nonlinearity and Heterogeneity We have so far concentrated on the estimation of treatment e ects when the treatment e

More information

A New Approach to Robust Inference in Cointegration

A New Approach to Robust Inference in Cointegration A New Approach to Robust Inference in Cointegration Sainan Jin Guanghua School of Management, Peking University Peter C. B. Phillips Cowles Foundation, Yale University, University of Auckland & University

More information

Instrumental variable models for discrete outcomes

Instrumental variable models for discrete outcomes Instrumental variable models for discrete outcomes Andrew Chesher The Institute for Fiscal Studies Department of Economics, UCL cemmap working paper CWP30/08 Instrumental Variable Models for Discrete Outcomes

More information

Partial Identi cation in Monotone Binary Models: Discrete Regressors and Interval Data.

Partial Identi cation in Monotone Binary Models: Discrete Regressors and Interval Data. Partial Identi cation in Monotone Binary Models: Discrete Regressors and Interval Data. Thierry Magnac Eric Maurin y First version: February 004 This revision: December 006 Abstract We investigate identi

More information

Identification of Regression Models with Misclassified and Endogenous Binary Regressor

Identification of Regression Models with Misclassified and Endogenous Binary Regressor Identification of Regression Models with Misclassified and Endogenous Binary Regressor A Celebration of Peter Phillips Fourty Years at Yale Conference Hiroyuki Kasahara 1 Katsumi Shimotsu 2 1 Department

More information

NBER WORKING PAPER SERIES NONPARAMETRIC IDENTIFICATION OF MULTINOMIAL CHOICE DEMAND MODELS WITH HETEROGENEOUS CONSUMERS

NBER WORKING PAPER SERIES NONPARAMETRIC IDENTIFICATION OF MULTINOMIAL CHOICE DEMAND MODELS WITH HETEROGENEOUS CONSUMERS NBER WORKING PAPER SERIES NONPARAMETRIC IDENTIFICATION OF MULTINOMIAL CHOICE DEMAND MODELS WITH HETEROGENEOUS CONSUMERS Steven T. Berry Philip A. Haile Working Paper 15276 http://www.nber.org/papers/w15276

More information

STOCKHOLM UNIVERSITY Department of Economics Course name: Empirical Methods Course code: EC40 Examiner: Lena Nekby Number of credits: 7,5 credits Date of exam: Friday, June 5, 009 Examination time: 3 hours

More information

Quantitative Techniques - Lecture 8: Estimation

Quantitative Techniques - Lecture 8: Estimation Quantitative Techniques - Lecture 8: Estimation Key words: Estimation, hypothesis testing, bias, e ciency, least squares Hypothesis testing when the population variance is not known roperties of estimates

More information

Trimming for Bounds on Treatment Effects with Missing Outcomes *

Trimming for Bounds on Treatment Effects with Missing Outcomes * CENTER FOR LABOR ECONOMICS UNIVERSITY OF CALIFORNIA, BERKELEY WORKING PAPER NO. 5 Trimming for Bounds on Treatment Effects with Missing Outcomes * David S. Lee UC Berkeley and NBER March 2002 Abstract

More information

ECON2285: Mathematical Economics

ECON2285: Mathematical Economics ECON2285: Mathematical Economics Yulei Luo Economics, HKU September 17, 2018 Luo, Y. (Economics, HKU) ME September 17, 2018 1 / 46 Static Optimization and Extreme Values In this topic, we will study goal

More information

Identification in Nonparametric Limited Dependent Variable Models with Simultaneity and Unobserved Heterogeneity

Identification in Nonparametric Limited Dependent Variable Models with Simultaneity and Unobserved Heterogeneity Identification in Nonparametric Limited Dependent Variable Models with Simultaneity and Unobserved Heterogeneity Rosa L. Matzkin 1 Department of Economics University of California, Los Angeles First version:

More information

Estimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels.

Estimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels. Estimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels. Pedro Albarran y Raquel Carrasco z Jesus M. Carro x June 2014 Preliminary and Incomplete Abstract This paper presents and evaluates

More information

Nonparametric Estimation of Wages and Labor Force Participation

Nonparametric Estimation of Wages and Labor Force Participation Nonparametric Estimation of Wages Labor Force Participation John Pepper University of Virginia Steven Stern University of Virginia May 5, 000 Preliminary Draft - Comments Welcome Abstract. Model Let y

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information