Casuality and Programme Evaluation

Similar documents
A Practitioner s Guide to Cluster-Robust Inference

Wild Bootstrap Inference for Wildly Dierent Cluster Sizes

A Course in Applied Econometrics Lecture 7: Cluster Sampling. Jeff Wooldridge IRP Lectures, UW Madison, August 2008

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Bootstrap-Based Improvements for Inference with Clustered Errors

Inference in difference-in-differences approaches

Econometrics II. Nonstandard Standard Error Issues: A Guide for the. Practitioner

Econometrics of Panel Data

New Developments in Econometrics Lecture 11: Difference-in-Differences Estimation

Difference-in-Differences Estimation

Cluster-Robust Inference

BOOTSTRAPPING DIFFERENCES-IN-DIFFERENCES ESTIMATES

A Practitioner s Guide to Cluster-Robust Inference

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Heteroskedasticity-Robust Inference in Finite Samples

Robust Inference with Multi-way Clustering

Econometric Analysis of Cross Section and Panel Data

Economics 582 Random Effects Estimation

Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data

Non-linear panel data modeling

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Reliability of inference (1 of 2 lectures)

Applied Microeconometrics (L5): Panel Data-Basics

Econometrics of Panel Data

Advanced Econometrics

Inference in Differences-in-Differences with Few Treated Groups and Heteroskedasticity

Inference about Clustering and Parametric. Assumptions in Covariance Matrix Estimation

Econometrics of Panel Data

Lecture 7: Dynamic panel models 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

GLS and FGLS. Econ 671. Purdue University. Justin L. Tobias (Purdue) GLS and FGLS 1 / 22

Clustering as a Design Problem

A Direct Test for Consistency of Random Effects Models that Outperforms the Hausman Test

FNCE 926 Empirical Methods in CF

Small sample methods for cluster-robust variance estimation and hypothesis testing in fixed effects models

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

GLS. Miguel Sarzosa. Econ626: Empirical Microeconomics, Department of Economics University of Maryland

Lecture 4: Linear panel models

Topic 10: Panel Data Analysis

Recitation 5. Inference and Power Calculations. Yiqing Xu. March 7, 2014 MIT

Panel Data Models. Chapter 5. Financial Econometrics. Michael Hauser WS17/18 1 / 63

The Exact Distribution of the t-ratio with Robust and Clustered Standard Errors

Fixed Effects Models for Panel Data. December 1, 2014

Panel Data Models. James L. Powell Department of Economics University of California, Berkeley

8. Nonstandard standard error issues 8.1. The bias of robust standard errors

When Should We Use Linear Fixed Effects Regression Models for Causal Inference with Panel Data?

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

LECTURE 11. Introduction to Econometrics. Autocorrelation

Econ 582 Fixed Effects Estimation of Panel Data

EC327: Advanced Econometrics, Spring 2007

Ninth ARTNeT Capacity Building Workshop for Trade Research "Trade Flows and Trade Policy Analysis"

Econometrics - 30C00200

What s New in Econometrics? Lecture 14 Quantile Methods

Final Exam. Economics 835: Econometrics. Fall 2010

Short T Panels - Review

Applied Economics. Panel Data. Department of Economics Universidad Carlos III de Madrid

Multiple Equation GMM with Common Coefficients: Panel Data

A Course in Applied Econometrics Lecture 18: Missing Data. Jeff Wooldridge IRP Lectures, UW Madison, August Linear model with IVs: y i x i u i,

The Exact Distribution of the t-ratio with Robust and Clustered Standard Errors

CRE METHODS FOR UNBALANCED PANELS Correlated Random Effects Panel Data Models IZA Summer School in Labor Economics May 13-19, 2013 Jeffrey M.

Testing Random Effects in Two-Way Spatial Panel Data Models

Econometrics. Week 6. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Statistics 910, #5 1. Regression Methods

Econometrics of Panel Data

Robust Inference with Clustered Data

Small sample methods for cluster-robust variance estimation and hypothesis testing in fixed effects models

Introductory Econometrics

Anova-type consistent estimators of variance components in unbalanced multi-way error components models

Bootstrapping the Grainger Causality Test With Integrated Data

Modified Variance Ratio Test for Autocorrelation in the Presence of Heteroskedasticity

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

A Practitioner's Guide to Cluster-Robust Inference

When Should We Use Linear Fixed Effects Regression Models for Causal Inference with Longitudinal Data?

Econometrics of Panel Data

Statistical Inference with Regression Analysis

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

y it = α i + β 0 ix it + ε it (0.1) The panel data estimators for the linear model are all standard, either the application of OLS or GLS.

Lecture 8 Panel Data

1 Motivation for Instrumental Variable (IV) Regression

Reply to Manovskii s Discussion on The Limited Macroeconomic Effects of Unemployment Benefit Extensions

A Non-parametric bootstrap for multilevel models

11. Bootstrap Methods

ECO375 Tutorial 8 Instrumental Variables

Problem Set 2

Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation

Econometrics Summary Algebraic and Statistical Preliminaries

Efficiency of repeated-cross-section estimators in fixed-effects models

10 Panel Data. Andrius Buteikis,

Bootstrap Methods in Econometrics

Capital humain, développement et migrations: approche macroéconomique (Empirical Analysis - Static Part)

ECON Introductory Econometrics. Lecture 16: Instrumental variables

Microeconometrics: Clustering. Ethan Kaplan

Panel Threshold Regression Models with Endogenous Threshold Variables

A Simple, Graphical Procedure for Comparing. Multiple Treatment Effects

QED. Queen s Economics Department Working Paper No. 1355

ECON 312 FINAL PROJECT

Økonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning

Dynamic Panels. Chapter Introduction Autoregressive Model

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test

Transcription:

Casuality and Programme Evaluation Lecture V: Difference-in-Differences II Dr Martin Karlsson University of Duisburg-Essen Summer Semester 2017 M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 1 / 41

Outline 1 Recap of Last Lecture 2 Introduction 3 Test Size Within-Group Correlation Serial Correlation Brewer et al (2013) 4 Low Power 5 Feasible GLS: A Combined Approach 6 Summary and Conclusions M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 2 / 41

Recap of Last Lecture Recap of Last Lecture DID is a well-established, powerful and simple technique. Simplest case: common time trend is sufficient to achieve consistency. The basic 2 2 model can be extended in various directions: Multiple groups, multiple periods Models with covariates Multiple dimensions (triple difference etc). Extensions for panel data and limited dependent variables exist, but can be more tricky. Synthetic control methods are a convenient way to define credible control groups at the aggregate level. The changes-in-changes model relaxes assumptions from the standard DID model; bases identification on monotonicity and invariance in the distribution of unobservables. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 3 / 41

Introduction Introduction Recent literature on inference in DID designs focus on the problem of incorrect test size. In fact, such designs give rise to potential sources of correlation between observations. Two main issues: Treatment status varies only at the group level ( clustering problem ). Treatment status typically highly correlated over time ( policy autocorrelation ). If these issues are ignored, inference may be misleading. Most recent literature shifts the focus to low power issues. How to address the power-size trade-off? M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 4 / 41

Type I and Type II Errors Introduction Null Hypothesis (H 0 ) Alternative Hypothesis (H 1 ) 0 Type II t α/2 Type I t Figure 1. Type I and Type II Errors. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 5 / 41

Test Size Problems with Standard Errors Recall from lecture 2: grouped residuals inflate standard errors. Consider the simple bivariate case Y ig = α + βx ig + e ig where there are G groups and common group errors: e ig = υ g + η ig Component υ g captures that group members are exposed to the same environment: classroom, teacher, weather... The intraclass correlation coefficient thus given by ρ e = σ2 υ σ 2 υ + σ 2 η...but the OLS estimator assumes iid residuals (υ g = 0). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 6 / 41

The Moulton Factor Test Size The Moulton factor: ratio between correct sampling variance and OLS variance. V ( ˆβ ) [ ( ) ] V ( V ) = 1 + ng + n 1 ρ x ρ e (1) c ˆβ n where ρ x = g j i =j ( xig x ) ( x jg x ) V ( x ig ) n n g ( ng 1 ) (2) Hence, the standard errors get inflated whenever Intraclass correlation is high (ρ e ). Group size varies considerably (V ( n g ) ). High intraclass correlation also in x ig (ρ x ) At least two of these apply by design in a DID setting. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 7 / 41

Test Size Two Dimensions of the Problem The general conclusion: OLS underestimates standard errors correction needed. Two dimensions: (A) Within-group correlation. Shared environment leads to correlated shocks. (B) Serial correlation. Outcomes typically exhibit persistence (earnings, employment, health...). Additional complication: Number of groups and time periods typically small. Inference based on G or T approaching infinity. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 8 / 41

Test Size A. Within-Group Correlation Within-Group Correlation Donald and Lang (Rev. Econ. Statist. 2007) discuss inference in DID and related models. Focus on within-group correlation of outcomes. Problem: some explanatory variables (like the treatment indicator) are constant among all members of a group. Three traditional solutions: 1 RE FGLS estimation. Estimate covariance matrix, reweight. 2 Correcting standard errors using covariance matrix with common group errors (Moulton 1990). 3 Cluster (Liang and Zeger 1986). D&L: These procedures based on G. Consider instead aggregating and drawing inference using T G 2. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 9 / 41

Test Size B. Serial Correlation, an Example Serial Correlation Bertrand et al (2004) utilise a standard dataset (the Current Population Survey - CPS): where Y ist = A s + B t + τd st + X ist β + ɛ ist (3) Y ist Log weekly earnings of females between 25-50 at t 1979 to 2000 in state s. D st Treatment indicator = 1 if state s is affected in year t. A s State fixed effects. B t Year fixed effects. X ist Individual-level control variables. Residual variation. ɛ ist Y is exhibits strong positive serial correlation: ρ 1 = 0.51, ρ 2 = 0.44 and ρ 3 = 0.33. In total 50 21 = 1, 050 state-year cells. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 10 / 41

The Problem Test Size Serial Correlation OLS gives an unbiased and consistent estimate ˆτ of effect. Bertrand et al run Monte Carlo simulations using placebo law changes. With consistent standard errors, false treatment effect should be observed in roughly 5% of cases. But standard errors are often inconsistent. H 0 is rejected in 67.5% of cases when neither within-group correlation nor serial correlation are taken into account. Taking within-group correlation into account (cluster or aggregate): H 0 is rejected in 44% of cases. Serial correlation can matter a lot! Many approaches to address the problem; none is uniformly better. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 11 / 41

Test Size Serial Correlation: Solutions Serial Correlation To evaluate possible solutions to the serial correlation problem, Bertrand et compare the simulated performance of five different techniques: 1 Parametric methods (AR(p)): perform poorly. 2 Block bootstrap: (sample clusters and calculate t statistic) performs well when the no. of groups is large. 3 Aggregate (collapse) time series information: reliable also when the no. of groups is small, on the other hand power is relatively low. 4 Empirical variance-covariance matrix: performs well in panels with high no. of groups, but assumes cross-sectional homoskedasticity (cf. Hausman & Kuersteiner, 2008). 5 Arbitrary (clustered) variance-covariance matrix: allows for an arbitrary correlation patterns over time. Performs well for moderate no. of groups; for small no. of groups d.o.f. adjustment needed. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 12 / 41

Test Size Serial Correlation The Clustered Covariance Matrix Estimator The empirical VCV estimator is consistent only under homoskedasticity. A robust alternative is the Clustered Covariance Matrix estimator (CCM; cf. Arellano, 1987): ( N ) Σ = (Z Z) 1 e se s (Z Z) 1. s=1 where Z Matrix of independent variables (i.e. A s, B t and D st ) with NT vectors z st. e s T t=1 υ stz st. υ st Estimated residuals for state s at time t. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 13 / 41

CCM: Properties Test Size Serial Correlation The estimation procedure that uses SEs computed according to the CCM performs quite well in finite samples. Approximately correct size regardless of relationship btw. N and T. However, there is still overrejection to some extent when the number of states is small: Bertrand et al reject H 0 in 8% (11%) of cases using a sample from 10 (6) states. Much better than before, but still twice nominal test size. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 14 / 41

CCM: Properties II Test Size Serial Correlation Asymptotic properties of CCM estimator for N are well known. Even without restrictions on the serial dependence, ˆΣ is N-consistent and asymptotically normal. But in DID studies, we often have small samples, in which robust standard errors are downwards biased. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 15 / 41

Hansen Correction Test Size Serial Correlation Hansen (2007a) derives properties of ˆΣ for T, N fixed: Even if {z st, υ st } is a strong mixing sequence (i.e. temporal dependence decreases in distance), ˆΣ is no longer consistent. If Var (z s ) and Σ s are the same for all s, standard t-statistics will be scaled by a factor of (N 1) N. Thus, using ( N N 1) ˆΣ and a t N 1 distribution will provide asymptotically unbiased inference irrespective of dimension approaching infinity. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 16 / 41

Brewer et al (2013): Test Size Test Size Brewer et al (2013) Brewer et al (2013): correct size can be obtained quite easily even when G is low!. Consider Model 3. The benchmark is the OLS estimator of ˆβ s standard error, i.e. assuming that errors are i.i.d. To get cluster-robust standard errors (CRSE), they use Liang and Zeger s (1986) formula to compute a cluster-robust variance matrix. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 17 / 41

Test Size Brewer et al (2013) Brewer et al (2013): Test Size II The estimator is consistent and Wald statistics are asymptotically normal as the no. of groups G. But it is biased (SE downward biased). The bias can be substantial when G is small. One way to reduce such bias is to scale up the residuals by before plugging them into the CRSE estimator. An alternative is to recover empirically the distribution of the t-statistic using a bootstrap procedure. G G 1 The wild cluster bootstrap-t procedure by Cameron et al (2008) outperformed other bootstrap-based approaches and works well also with small G. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 18 / 41

Wild Cluster Bootstrap-t Test Size Brewer et al (2013) Cf. Cameron & Miller (2013) A Practitioner s Guide to Cluster-Robust Inference. 1 Estimate with OLS, imposing H 0 : β 1 = β 0 1 and recover residual vectors {û 1,..., û G }. 2 Generate pseudo-residuals as û g = û g or û g = û g ; each with probability 0.5 and the resulting pseudo-sample { (ŷ 1, X 1 ),..., (ŷ G, X G) }. 3 Generate OLS estimate ˆβ 1,b, standard error s ˆβ ( ) 1,b wb = ˆβ 1,b β0 1 /s ˆβ 1,b. 4 Repeat for b = 1,..., B. 5 Reject H 0 at level α if w / [w α/2, w 1 α/2 ]. and Wald statistic M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 19 / 41

Summary Test Size Brewer et al (2013) Brewer et al address both serial correlation and within-group correlation in the following steps: Aggregate data on state-year level. Apply a scaling factor to the residuals: G G 1. Plug the scaled residuals into the cluster-robust variance-covariance matrix to get cluster-robust standard errors (CRSE). Use critical values from a t distribution with d.o.f. correction: t G 1 instead of a standard normal. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 20 / 41

Experimental Design Test Size Brewer et al (2013) They use the same data as Betrand et al on the period 1979-2008 and placebo law changes with tests of nominal 5% size. Monte Carlo simulations to show that their procedure allows to build tests with the intended test size. Resample states with replacement; half of the states are treated. They use OLS and FGLS and compare rejection rates, assuming different inference methods and different number of groups: 1 Errors i.i.d. 2 CRSE, unscaled residuals and N(0, 1) 3 CRSE, unscaled residuals and t G 1 4 CRSE, scaled residuals and N(0, 1) 5 CRSE, scaled residuals and t G 1 6 Wild cluster bootstrap-t 6, 10, 20, 50 states resampled. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 21 / 41

Experimental Design II Test Size Brewer et al (2013) The purpose is to compare the performance of the different methods in terms of both Type I and Type II errors. Robustness checks: Robustness to mis-specification of the error process: State-time shocks simulated according to an AR(1) process with varying parameters. Vary the fraction of treated groups to check performance in unbalanced designs. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 22 / 41

Compare Methods Test Size Brewer et al (2013) Table 1. Rejection rates when the null is true. Tests of nominal 5% size with placebo treatments in log-earnings data, 5,000 replications. Equation 3 is estimated by OLS. Inference method G = 50 G = 20 G = 10 G = 6 i.i.d. errors 0.429 0.424 0.422 0.413 (0.007) (0.007) (0.007) (0.007) CRSE, N(0, 1) critical values 0.059 0.073 0.110 0.175 (0.003) (0.004) (0.004) (0.005) CRSE, t G 1 critical values 0.053 0.056 0.066 0.095 (0.003) (0.003) (0.004) (0.004) CRSE, G residuals, N(0, 1) (G 1) 0.049 0.056 0.071 0.113 (0.003) (0.003) (0.004) (0.004) CRSE, G (G 1) residuals, t G 1 0.045 0.041 0.042 0.052 (0.003) (0.003) (0.003) (0.003) Wild cluster bootstrap-t 0.044 0.041 0.048 0.059 (0.003) (0.003) (0.003) (0.003) Simulation standard errors in parentheses. The treatment parameter has a true coefficient of zero. G number of sampled states. Data from 1976 to 2008 inclusive (T = 30). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 23 / 41

Imbalance between Groups Test Size Brewer et al (2013) Table 2. Rejection rates when the null is true. Tests of nominal 5% size with placebo treatments in log-earnings data, 5,000 replications. Equation 3 is estimated by OLS. Inference method G1 = 5 G1 = 4 G1 = 3 G1 = 2 i.i.d. errors 0.422 0.408 0.409 0.405 (0.007) (0.007) (0.007) (0.007) CRSE, N(0, 1) critical values 0.110 0.125 0.150 0.241 (0.004) (0.005) (0.005) (0.006) CRSE, t G 1 critical values 0.066 0.079 0.105 0.191 (0.004) (0.004) (0.004) (0.006) CRSE, G residuals, N(0, 1) (G 1) 0.071 0.084 0.113 0.199 (0.004) (0.004) (0.004) (0.006) CRSE, G (G 1) residuals, t G 1 0.042 0.051 0.074 0.150 (0.003) (0.003) (0.004) (0.005) Wild cluster bootstrap-t 0.048 0.054 0.052 0.018 (0.003) (0.003) (0.003) (0.002) Simulation standard errors in parentheses. The treatment parameter has a true coefficient of zero. G1 the number of treated out of a total of 10 states. Data from 1976 to 2008 inclusive (T = 30). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 24 / 41

Size vs Power Low Power The proposed combined modifications (scaled CRSE and t S 1 critical values) yield good results in most cases: true test size is within about 1% of nominal test size. Large imbalance between the numbers of treatment and control groups wild cluster bootstrap-t procedure performs better. However, Brewer et al stress that it is relatively easy to obtain the correct test size. The main issue is that the power to detect real treatment effects with tests of the correct size is low. It is extremely low when S is small. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 25 / 41

(Low) Power with OLS Low Power Table 3. Rejection rates of H 0 : no treatment effect when β is the true value of the treatment parameter. Tests of nominal 5% size with placebo treatments in log-earnings data, 5,000 replications. Comparison of different inference methods. True effect Inference G = 50 G = 20 G = 10 G = 6 β = 0.02 G (G 1) CRSE, t G 1 critical values 0.238 0.134 0.088 0.074 (0.006) (0.005) (0.004) (0.004) wild cluster bootstrap-t 0.225 0.125 0.093 0.074 (0.006) (0.005) (0.004) (0.004) β = 0.05 G (G 1) CRSE, t G 1 critical values 0.822 0.513 0.273 0.168 (0.005) (0.007) (0.006) (0.005) wild cluster bootstrap-t 0.799 0.490 0.283 0.167 (0.006) (0.007) (0.006) (0.005) β = 0.10 G (G 1) CRSE, t G 1 critical values 1.000 0.919 0.718 0.448 (0.000) (0.004) (0.006) (0.007) wild cluster bootstrap-t 0.999 0.898 0.712 0.429 (0.000) (0.004) (0.006) (0.007) β = 0.15 G (G 1) CRSE, t G 1 critical values 1.000 0.995 0.904 0.755 (.) (0.001) (0.004) (0.006) wild cluster bootstrap-t 1.000 0.992 0.896 0.700 (.) (0.001) (0.004) (0.006) Simulation standard errors in parentheses. G number of sampled states. Data from 1976 to 2008 inclusive (T = 30). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 26 / 41

Low Power Minimum Detectable Effect The power of the two inference methods is similar Power is documented more comprehensively when looking at the minimum effect that would be detected (Minimum Detectable Effect - MDE). Recall: the MDE is defined as MDE(κ) = ŜE( ˆβ) [ c u + p1 κ t ] where κ Level of power. SE( ˆβ) Scaled CRSE estimate. c u Upper critical value of the t S 1 distribution. p1 κ t (1 κ)th percentile of the t-statistic under H 0 : no treatment effect. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 27 / 41

Low Power Minimum Detectable Effect: Illustration Figure 2 shows the proportion of times for which H 0 : no treatment effect is rejected when the treatment parameter β has a coefficient between 0 Figure 1: Minimum detectable eects on log-earnings using and 0.3. G/(G 1)-CRSEs and tg 1 critical values and tests of size 0.05 Figure 2. MDE on log-earnings with scaled residuals, t G 1 critical values and The Figure shows the proportion of the time that the null hypothesis of no treatment eect is rejected when the treatment parameter has a true coecient ranging from 0 to 0.3. Numbers are computed using the results of 100000 Monte Carlo simulations combined tests of 5% size, 100,000 with equation replications. 4, as described in the text. The simulations resample states from CPS-MORG data, having imposed the sampling restrictions described in the text. Data from 1979 to 2008 inclusive are sampled (i.e. T = 30). Regressions are run on aggregated M Karlsson (University of Duisburg-Essen) state-year data. Casuality and Programme Evaluation Summer Semester 2017 28 / 41

Feasible GLS: A Combined Approach Solution: Increasing power with FGLS Hansen (2007b) proposes to use a FGLS estimation under the assumption that the state-year shock follows a stationary AR(p) process. Coefficients of the AR(p) process can be biased in panel data if the time dimension is short and fixed effects are included (incidental parameters problem). He also introduces a bias correction to account for this. He finds that the FGLS estimation clearly dominates OLS also when inference is based on CRSE. The FGLS procedure with bias correction (BC-FGLS) is consistent as S. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 29 / 41

Feasible GLS: A Combined Approach Solution: Increasing power with FGLS II Brewer et al use simulations to show that it is possible to retain the correct test size and achieve gains in power by using FGLS instead of OLS. Combine BC-FGLS with robust inference technique (scaled CRSE and critical values from t with d.o.f. adjustment). In fact, the size of the test can be controlled using robust inference, even for small S. In this way tests have the correct size and FGLS improves power considerably. Procedure also robust to mis-specifications of the error process. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 30 / 41

Feasible GLS: A Combined Approach A Performance Comparison Table 4. Rejection rates for tests of nominal 5% size with placebo treatments in log-earnings data, 5,000 replications, comparison of different estimation and inference methods. Estimation Inference G = 50 G = 20 G = 10 G = 6 OLS G (G 1) CRSE, t G 1 critical values 0.045 0.041 0.042 0.052 (0.003) (0.003) (0.003) (0.003) FGLS (no correction) 0.106 0.101 0.120 0.124 (0.004) (0.004) (0.005) (0.005) FGLS G (G 1) CRSE, t G 1 critical values 0.049 0.045 0.054 0.061 (0.003) (0.003) (0.003) (0.003) BC-FGLS 0.073 0.070 0.087 0.096 (0.004) (0.004) (0.004) (0.004) BC-FGLS G (G 1) CRSE, t G 1 critical values 0.049 0.045 0.058 0.065 (0.003) (0.003) (0.003) (0.003) Simulation standard errors in parentheses. The treatment parameter has a true coefficient of zero. G number of sampled states. Data from 1976 to 2008 inclusive (T = 30). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 31 / 41

Feasible GLS: A Combined Approach Assumptions about the Serial Correlation Process Table 5. Rejection rates for tests of nominal 5% size with placebo treatments in log-earnings data, 5,000 replications, 10 groups. Empirical regression residuals (CPS) replaced by a simulated error term generated according to an AR(2) and a MA(1) process. Estimation Inference CPS residuals Heterogeneous AR(2) MA(1) OLS G (G 1) CRSE, t G 1 0.049 0.040 0.052 (0.003) (0.002) (0.002) FGLS (no correction) 0.114 0.101 0.088 (0.004) (0.003) (0.003) FGLS G (G 1) CRSE, t G 1 0.054 0.055 0.051 (0.003) (0.002) (0.002) BC-FGLS 0.081 0.072 0.072 (0.004) (0.003) (0.003) BC-FGLS G (G 1) CRSE, t G 1 0.056 0.059 0.052 (0.003) (0.002) (0.002) Simulation standard errors in parentheses. The treatment parameter has a true coefficient of zero. G number of sampled states. Data from 1976 to 2008 inclusive (T = 30). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 32 / 41

Gains in Power Feasible GLS: A Combined Approach Table 6. Rejection rates for tests of nominal 5% size with a treatment effect of +0.05 in log earnings data Estimation Inference G = 50 G = 20 G = 10 G = 6 OLS G (G 1) CRSE, t G 1 0.810 0.467 0.252 0.168 (0.006) (0.007) (0.006) (0.005) FGLS (no correction) 0.985 0.799 0.573 0.434 (0.002) (0.006) (0.007) (0.007) FGLS G (G 1) CRSE, t G 1 0.957 0.670 0.401 0.255 (0.003) (0.007) (0.007) (0.006) BC-FGLS 0.978 0.763 0.513 0.384 (0.002) (0.006) (0.007) (0.007) BC-FGLS G (G 1) CRSE, t G 1 0.955 0.696 0.423 0.286 (0.003) (0.007) (0.007) (0.006) Simulation standard errors in parentheses. Data from 1976 to 2008 inclusive (T = 30). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 33 / 41

Summary and Conclusions Summary and Conclusions In DID designs: great risk of underestimating standard errors: Dependent variables tend to be positively serially correlated. Treatment tends to be serially correlated as well. Various methods allow to correct for this problem, e.g. bootstrapping or robust covariance matrix estimators (clustered covariance matrix estimator). However, even when test size is correct, one issue is low power. Brewer et al (2013) get accurate size (easy), then maximise power: Robust inference (CRSE with scaled residuals & t S 1 critical values)...coupled with a FGLS estimation procedure. Performs quite well also when S is small and is robust to mis-specifications of the error process. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 34 / 41

1. Parametric Methods Appendix Alternative Solutions to Serial Correlation Parametric methods specify an autocorrelation structure, which is then estimated: It may be either individual-specific or uniform. This was traditionally the common approach to deal with the problem. OLS residuals used to estimate autocorrelation parameters (ρ). Finally, employ ρ s in an FGLS regression. Problem: In short time series, autocorrelation parameters estimated by OLS are biased downwards! Consequence: Over-rejection remains a problem. Hansen s bias correction is consistent for S (T fixed). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 35 / 41

2. Boostrapping Appendix Alternative Solutions to Serial Correlation Simple Bootstrapping. Boostrapping: a technique used when we are unable or unwilling to derive the distribution of our estimator. A simple bootstrapping scheme draws R samples of size N from our original sample. On each of these samples, we run our main regression. Our R estimated parameters ˆτ r will mimic the distribution of ˆτ. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 36 / 41

Block Boostrap Appendix Alternative Solutions to Serial Correlation Block bootstrap preserves the autocorrelation structure by using series of observations instead of individual observations. 1 Bootstrap sample is generated by drawing N s matrices (Y s, V s ): Y s is the entire series of observations for state s. V s is the matrix of D, state & time dummies for state s. 2 Run OLS on each sample, obtain estimate ˆβ and absolute t statistic t r = ˆβ r ˆβ SE ( ˆβ r ) 3 t r approaches the sampling distribution of t as R increases. Assessment: Significant improvement over parametric techniques, but many groups required. Implemented in Stata by xtreg yvar treatvar xvars, i(id) fe vce(bootstrap, seed(1234)). M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 37 / 41

Appendix 3. Ignoring Time Series Information Alternative Solutions to Serial Correlation Simpler alternative: ignore time series information. For laws implemented at the same time in all treated groups, we can simply compute pre- and post-reform averages for each group. If not, proceed as follows: 1 Start with a regression leaving treatment indicator out Y st = A s + B t + X st β + υ st 2 Calculate before and after averages for treated groups only: ˆῡ 0 s = T t=1 1 (D st = 0) ˆυ st T t=1 1 (D st = 0) ˆῡ 1 s = T t=1 1 (D st = 1) ˆυ st T t=1 1 (D st = 1) = T t=1 1 (D st = 0) ( Y st  s ˆB t X st ˆβ ) t=1 T 1 (D st = 0) = T t=1 1 (D st = 1) ( Y st  s ˆB t X st ˆβ ) t=1 T 1 (D st = 1) M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 38 / 41

Appendix Ignoring Time Series Information II Alternative Solutions to Serial Correlation 3 Run the regression ˆῡ st = τd st + u st When S is small, need to make a correction to the t statistic. Simple aggregation performs well, and residual aggregation has reasonable rejection rates as well. But power tends to be very low! M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 39 / 41

4. Empirical VCV Matrix Appendix Alternative Solutions to Serial Correlation Parametric corrections unnecessarily inflexible: S > 1, so we can estimate the covariance matrix more flexibly......if we are willing to assume autocorrelation structure is the same......and homoskedastic (Kiefer, 1980; Hausman and Kuersteiner, 2008). Thus, we express the dataset in vector form, where Y s is the T 1 vector of outcome observations. We want to estimate the T T matrix Σ. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 40 / 41

Empirical VCV Matrix II Appendix Alternative Solutions to Serial Correlation Consider the empirical covariance matrix = 1 N N [Q (Y s τd s X s β)] [Q (Y s τd s X s β)] s=1 where Q performs a within transformation. And then use the estimated matrix to compute standard errors: Var ( β ) = [ N ] Z sq ( ) 1 1 QZs s=1 The matrix has rank T 1: we need a generalised inverse such as suggested by Hsiao (2003). This method performs well when S is large. M Karlsson (University of Duisburg-Essen) Casuality and Programme Evaluation Summer Semester 2017 41 / 41