AN EVALUATION OF PARAMETRIC AND NONPARAMETRIC VARIANCE ESTIMATORS IN COMPLETELY RANDOMIZED EXPERIMENTS. Stanley A. Lubanski. and. Peter M.

Similar documents
Online Appendix to Yes, But What s the Mechanism? (Don t Expect an Easy Answer) John G. Bullock, Donald P. Green, and Shang E. Ha

Econometrics of Panel Data

Notes on causal effects

Controlling for latent confounding by confirmatory factor analysis (CFA) Blinded Blinded

A COMPARISON OF HETEROSCEDASTICITY ROBUST STANDARD ERRORS AND NONPARAMETRIC GENERALIZED LEAST SQUARES

12E016. Econometric Methods II 6 ECTS. Overview and Objectives

The Economics of European Regions: Theory, Empirics, and Policy

Causal Inference Lecture Notes: Causal Inference with Repeated Measures in Observational Studies

ESTIMATION OF TREATMENT EFFECTS VIA MATCHING

Simple Linear Regression

Robust Standard Errors in Small Samples: Some Practical Advice

On the Use of Linear Fixed Effects Regression Models for Causal Inference

Clustering as a Design Problem

When Should We Use Linear Fixed Effects Regression Models for Causal Inference with Longitudinal Data?

Quantitative Economics for the Evaluation of the European Policy

NBER WORKING PAPER SERIES ROBUST STANDARD ERRORS IN SMALL SAMPLES: SOME PRACTICAL ADVICE. Guido W. Imbens Michal Kolesar

New Developments in Econometrics Lecture 11: Difference-in-Differences Estimation

A Course in Applied Econometrics Lecture 18: Missing Data. Jeff Wooldridge IRP Lectures, UW Madison, August Linear model with IVs: y i x i u i,

Categorical Predictor Variables

When Should We Use Linear Fixed Effects Regression Models for Causal Inference with Panel Data?

The Model Building Process Part I: Checking Model Assumptions Best Practice

Combining multiple observational data sources to estimate causal eects

Plausible Values for Latent Variables Using Mplus

Andreas Steinhauer, University of Zurich Tobias Wuergler, University of Zurich. September 24, 2010

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator

Randomized trials for policy

What s New in Econometrics. Lecture 1

Instrumental Variables

On Variance Estimation for 2SLS When Instruments Identify Different LATEs

When Should We Use Linear Fixed Effects Regression Models for Causal Inference with Longitudinal Data?

IsoLATEing: Identifying Heterogeneous Effects of Multiple Treatments

The Exact Distribution of the t-ratio with Robust and Clustered Standard Errors

Inference in difference-in-differences approaches

Applied Econometrics Lecture 1

Variable Selection and Model Building

Inferences for Regression

Heteroskedasticity-Robust Inference in Finite Samples

Bayesian Interpretations of Heteroskedastic Consistent Covariance Estimators Using the Informed Bayesian Bootstrap

Estimation of the Conditional Variance in Paired Experiments

On dealing with spatially correlated residuals in remote sensing and GIS

Small-sample cluster-robust variance estimators for two-stage least squares models

Chapter 11 Specification Error Analysis

The Exact Distribution of the t-ratio with Robust and Clustered Standard Errors

A Non-parametric bootstrap for multilevel models

What s New in Econometrics? Lecture 14 Quantile Methods

Introduction to Regression Analysis. Dr. Devlina Chatterjee 11 th August, 2017

An Introduction to Causal Mediation Analysis. Xu Qin University of Chicago Presented at the Central Iowa R User Group Meetup Aug 10, 2016

Reliability of inference (1 of 2 lectures)

Econometrics of Panel Data

Multiple Linear Regression

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

Outline. Possible Reasons. Nature of Heteroscedasticity. Basic Econometrics in Transportation. Heteroscedasticity

19 Effect Heterogeneity and Bias in Main-Effects-Only Regression Models

Comparing Change Scores with Lagged Dependent Variables in Models of the Effects of Parents Actions to Modify Children's Problem Behavior

Effect Heterogeneity and Bias in Main-Effects- Only Regression Models

A better way to bootstrap pairs

arxiv: v2 [math.st] 14 Oct 2017

The Multiple Linear Regression Model

1 Estimation of Persistent Dynamic Panel Data. Motivation

Bootstrap Approach to Comparison of Alternative Methods of Parameter Estimation of a Simultaneous Equation Model

AGEC 661 Note Fourteen

Flexible Estimation of Treatment Effect Parameters

ECON The Simple Regression Model

Bayesian Interpretations of Heteroskedastic Consistent Covariance. Estimators Using the Informed Bayesian Bootstrap

Econometrics II. Nonstandard Standard Error Issues: A Guide for the. Practitioner

Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis

An Alternative Assumption to Identify LATE in Regression Discontinuity Designs

Inference for Average Treatment Effects

Correlation and Simple Linear Regression

New Developments in Econometrics Lecture 16: Quantile Estimation

The Prediction of Monthly Inflation Rate in Romania 1

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

MS&E 226: Small Data

An Alternative Assumption to Identify LATE in Regression Discontinuity Design

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

The Model Building Process Part I: Checking Model Assumptions Best Practice (Version 1.1)

Test Yourself! Methodological and Statistical Requirements for M.Sc. Early Childhood Research

Lecture 2 Simple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: Chapter 1

Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case

Correlation Analysis

Discussion: Can We Get More Out of Experiments?

Small sample methods for cluster-robust variance estimation and hypothesis testing in fixed effects models

A Practitioner s Guide to Cluster-Robust Inference

9/26/17. Ridge regression. What our model needs to do. Ridge Regression: L2 penalty. Ridge coefficients. Ridge coefficients

Implementing Matching Estimators for. Average Treatment Effects in STATA

Causal Mechanisms Short Course Part II:

The Simulation Study to Test the Performance of Quantile Regression Method With Heteroscedastic Error Variance

Some Monte Carlo Evidence for Adaptive Estimation of Unit-Time Varying Heteroscedastic Panel Data Models

FAQ: Linear and Multiple Regression Analysis: Coefficients

Telescope Matching: A Flexible Approach to Estimating Direct Effects

11. Bootstrap Methods

Stat/F&W Ecol/Hort 572 Review Points Ané, Spring 2010

Exam D0M61A Advanced econometrics

Introduction to Econometrics. Heteroskedasticity

Version: September, 2013

Homoskedasticity. Var (u X) = σ 2. (23)

Transcription:

AN EVALUATION OF PARAMETRIC AND NONPARAMETRIC VARIANCE ESTIMATORS IN COMPLETELY RANDOMIZED EXPERIMENTS by Stanley A. Lubanski and Peter M. Steiner UNIVERSITY OF WISCONSIN-MADISON 018

Background To make valid inferences about population parameters, researchers must be able to accurately estimate those parameters, but also be able to estimate the variability of those estimates so that they may rule out sampling uncertainty as a reasonable explanation for the results observed in their sample. However, even in the simple, well-understood context of the completely randomized experiment, advice for selecting a variance estimator can vary: some recommend nonparametric, design-based estimators (Imbens & Rubin, 015; Schochet, 010), while others recommend catch-all corrections to the conventional parametric variance estimators (Angrist & Pischke, 008; Long & Ervin, 000; MacKinnon & White, 1985). In this paper, an answer is sought to the following question: do nonparametric variance estimators perform better than conventional or corrected variance estimators in completely randomized experiments (CREs) and, if so, under what conditions? Objective This paper focuses on evaluating the performance of four variance estimators in a CRE with two experimental groups. The first estimator will be the conventional regression variance estimator which assumes homoscedasticity of variance (i.e., constant error variance between the groups; hence this estimator is labeled Constant ). Two design-based estimators labeled Neyman and Fixed Sample Upper Bound ; FSUB for short (Imbens & Rubin, 015; Schochet, 015) and one corrected parametric estimator the HC3 estimator (MacKinnon & White, 1985) all of which relax the homoscedasticity assumption, will be compared to the conventional estimator. Of interest is how each variance estimator will handle different sources and types of variability. Formulae are presented in Table 1 below. By design, point estimates from a CRE will have random variability two sources: random assignment or random sampling. If a researcher wishes to generalize back to a larger target population, then they often use the super population model where both random assignment and random sampling are present. If a researcher is only concerned with the single sample they have drawn, then they often use what we will call the fixed sample model where only random assignment is present. Additional factors that impact the variability of the point estimators are also considered, including: heteroscedasticity, treatment group proportion, and covariance adjustment.

Covariance Adjusted (X) Unadjusted Table 1 Variance Estimators to be evaluated. No. Model type Name Notation Point Estimator Formula 1 Heteroscedasticity- Consistent Covariance Estimator 3 (HC3) Constant V Con τ V HC3 τ s t 3 Neyman V Ney τ s t 4 Fixed Sample Upper Bound (FSUB) 5 Constant w/covariates* (m 1) ( s t n + s (n m 1) n c ) n m(n m) s c m 1 + n m 1 s c m + n m V FSUB τ s t m + s c n m (s t s c ) n X X V Con τ MSE(M M) 1 8 HC3 w/covariates* X X V HC3 τ (M M) 1 M diag [ (Y i Y i) (1 h ii ) ] M(M M) 1 6 Neyman w/covariates X X V Ney τ MSE t m + MSE c n m 7 FSUB w/covariates X X V FSUB τ MSE t m + MSE c n m ( MSE t MSE c ) n *formulas given for these estimators are in matrix notation, so the value of interest would be the (, ) element of the resulting matrix Note 1. M is the design matrix with n rows and k+1 columns, where k is the number of predictors Note. MSE is the estimated mean square error of the residuals Note 3. MSE t and MSE c are the sum of squares of the residuals for the treatment and control group, respectively, divided by the appropriate degrees of freedom Note 4. For simplicity, Covariance Adjusted can refer to either the main effects or interaction model; variance estimator equations for both models are identical, but differ only in the value of the residuals, hat values, or design matrix

Table Research Design In R, we ran a simulation with 10,000 iterations for each of the 4 combinations of the three following factors: inference population (four levels), heteroscedasticity (three levels), and treatment group proportion ( levels). The base simulation used normally-distributed error terms and sample size of 1,000, but the full results will include non-normally distributed error terms and sample size of 100. For each iteration, point and variance estimates are computed for three regression models: (1) unadjusted model the only predictor is the binary treatment indicator, Z, () main effects model in a addition to Z, four covariates (Xs) are included as predictors, (3) interaction model in addition to the main effects of Z and the Xs, the interaction of each of the Xs with Z are included as predictors. In all regression models, the predictors were centered at their sample means. Detail of the full data generating model are presented in Table below. Data Generating Model Variables Explanation Value/Model τ X β y Population Average Treatment Effect (PATE) Covariates matrix Vector of coefficients relating X s to Y(0) 5 1 1. 4. 0. 4 1 0. 1 ~N 4 μ x = ( ), Σ 1 x = ( ). 0 1 0 0. 1 0 1 ( ) β 1 = 4 β ( = ) β 3 = 1 β 4 = 3 ε y Y(0) error term ~N(0, 5) Y(0) Control Potential Outcome Xβ y + ε y ϵran Y(1)ran error term ~N(0, σ Y(0) ) Y(1)con Treatment Potential Outcome (constant) Y(0) + τ Y(1)ran Treatment Potential Outcome (random) Y(0) + τ + ε ran Y(1)sys Treatment Potential Outcome (systematic) Y(0) + τx 1

Data Analysis To determine if the variance estimators were biased, the mean variance estimates were compared to the empirical variance of the point estimate across simulations, in a ratio of the estimated variance over the true variance. However, to account for the finiteness of the simulation, the variance ratios were bootstrapped 1,000 times and then 99% bootstrap confidence intervals were computed. If the confidence intervals contained 1, then the estimator was considered unbiased, else, it was considered biased. Results The plot of the results for the fixed sample inference population are presented in Figure 1; results for other inference populations will be discussed in relation to this plot. Taking the figure row-by-row, first is the case of no treatment effect heterogeneity. In the left plot the proportion of treated units, p, equals 0.5 and all estimators are unbiased. In the right plot, where p = 0.1, the only bias occurs in the interaction model where small biases exist for the nonparametric and HC3 estimators. Super population results for this row are essentially identical. In the second row with random treatment effect heterogeneity, in the left plot, all estimators are upwardly biased because, for fixed samples, the presence of treatment effect heterogeneity reduces point estimator variability (Imbens & Rubin, 015; Schochet, 010). However, only the FSUB estimator has some correction to reflect this and thus its bias is less severe. In the right plot where p = 0.1, the Constant estimator becomes severely biased in all three regression models, underestimating the variance by more than 50% in the covariance adjusted models. Some of the other estimators are slightly upwardly biased. The upward bias is not present in the super population model, so there the FSUB estimator is downwardly biased; the rest of the results are nearly identical. Finally, in the third row with systematic treatment effect heterogeneity, when p = 0.5, the estimators differ little, but some upward bias is present, as in the previous row of plots. None of the estimators are biased when correctly modeling the heterogeneity. When p = 0.1, again, the Constant estimator is downwardly biased (except when the model is correct), but the other estimators are similar. Again, the upward bias is not present in the super population, but due to sample mean centering, there is downward bias for the covariance adjusted model with interactions.

Figure 1. Variance estimator bias as variance ratios (estimated variance over true variance) for four variance estimators. Rows indicate level of treatment effect heterogeneity (top-tobottom: none, random, systematic). Columns indicate treated unit proportion (left-toright: p = 0.5, p = 0.1). X-axis of plots refer to variance estimators; first four estimates from unadjusted regression model; next four from covariance adjusted regression model (main effects only); last four from covariance adjusted model (main effects plus interactions). Estimators with red triangles indicate bias, i.e., that the 99% bootstrap interval does not include 1. Conclusions If the CRE is balanced and there is little or no heteroscedasticity, then all variance estimators work well, but if the design is unbalanced, then the conventional parametric estimator is best as it pools the variance from both treatment groups to give the most efficient variance estimate. However, if there is random, idiosyncratic variability causing heteroscedasticity, then the nonparametric estimators are best, especially if the design is unbalanced or if working under the fixed sample model. Finally, if the heteroscedasticity is caused by systematic variation, then the choice of variance estimator matters little and it is most important to model the variation appropriately and to avoid sample-mean-centering if working under the super population model.

References Angrist, J. D., & Pischke, J.-S. (008). Mostly harmless econometrics: An empiricist s companion. Princeton university press. Imbens, G. W., & Rubin, D. B. (015). Causal inference in statistics, social, and biomedical sciences. Cambridge University Press. Long, J. S., & Ervin, L. H. (000). Using Heteroscedasticity Consistent Standard Errors in the Linear Regression Model. American Statistician, 54(3), 17 4. https://doi.org/10.307/685594 MacKinnon, J. G., & White, H. (1985). Some heteroskedasticity-consistent covariance matrix estimators with improved finite sample properties. Journal of Econometrics, 9(3), 305 35. Schochet, P. Z. (010). Is regression adjustment supported by the Neyman model for causal inference? Journal of Statistical Planning and Inference, 140(1), 46 59. https://doi.org/10.1016/j.jspi.009.07.008 Schochet, P. Z. (015). Statistical Theory for the RCT-YES Software: Design-Based Causal Inference for RCTs, (June). Retrieved from https://files.eric.ed.gov/fulltext/ed556496.pdf