Discussant: Lawrence D Brown* Statistics Department, Wharton, Univ. of Penn.

Similar documents
Model Selection, Estimation, and Bootstrap Smoothing. Bradley Efron Stanford University

Lawrence D. Brown* and Daniel McCarthy*

Higher-Order von Mises Expansions, Bagging and Assumption-Lean Inference

Mallows Cp for Out-of-sample Prediction

Why Do Statisticians Treat Predictors as Fixed? A Conspiracy Theory

9 Correlation and Regression

Construction of PoSI Statistics 1

Simultaneous Inference: An Overview

Objectives Simple linear regression. Statistical model for linear regression. Estimating the regression parameters

Remedial Measures for Multiple Linear Regression Models

Stat 101 L: Laboratory 5

Business Statistics. Lecture 10: Course Review

Finite Population Correction Methods

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

BIOMETRICS INFORMATION

Inference for Regression Simple Linear Regression

Supporting Information for Estimating restricted mean. treatment effects with stacked survival models

Applied Regression Modeling: A Business Approach Chapter 3: Multiple Linear Regression Sections

STAT 704 Sections IRLS and Bootstrap

Resampling and the Bootstrap

Experimental Design and Data Analysis for Biologists

Measurement error as missing data: the case of epidemiologic assays. Roderick J. Little

Inference for the Regression Coefficient

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

PoSI and its Geometry

L. Brown. Statistics Department, Wharton School University of Pennsylvania

PoSI Valid Post-Selection Inference

Chapter 10 Regression Analysis

Covariance and Correlation

Problem Set 3: Bootstrap, Quantile Regression and MCMC Methods. MIT , Fall Due: Wednesday, 07 November 2007, 5:00 PM

Regression. Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables X and Y.

Confidence Estimation Methods for Neural Networks: A Practical Comparison

Biostatistics Advanced Methods in Biostatistics IV

STATISTICS 141 Final Review

Statistical Inference

Fast and robust bootstrap for LTS

Chapter 7: Simple linear regression

Marginal Screening and Post-Selection Inference

Data Set 8: Laysan Finch Beak Widths

Response Surface Methodology

Recitation 5. Inference and Power Calculations. Yiqing Xu. March 7, 2014 MIT

Subject CS1 Actuarial Statistics 1 Core Principles

Post-Selection Inference for Models that are Approximations

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

Introduction to bivariate analysis

Bootstrapping, Randomization, 2B-PLS

STAT5044: Regression and Anova. Inyoung Kim

Topic 22 Analysis of Variance

Confidence Intervals, Testing and ANOVA Summary

Introduction to bivariate analysis

Generalized Cp (GCp) in a Model Lean Framework

Lecture 3: Just a little more math

Simple Linear Regression for the Climate Data

1 Introduction to Minitab

Inference for Regression Inference about the Regression Model and Using the Regression Line, with Details. Section 10.1, 2, 3

Empirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss;

SUPPLEMENT TO PARAMETRIC OR NONPARAMETRIC? A PARAMETRICNESS INDEX FOR MODEL SELECTION. University of Minnesota

STA 101 Final Review

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

TA: Sheng Zhgang (Th 1:20) / 342 (W 1:20) / 343 (W 2:25) / 344 (W 12:05) Haoyang Fan (W 1:20) / 346 (Th 12:05) FINAL EXAM

Response Surface Methodology:

Multiple Linear Regression for the Supervisor Data

Assumptions, Diagnostics, and Inferences for the Simple Linear Regression Model with Normal Residuals

General Linear Model (Chapter 4)

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility

DESIGN AND ANALYSIS OF EXPERIMENTS Third Edition

Prentice Hall Stats: Modeling the World 2004 (Bock) Correlated to: National Advanced Placement (AP) Statistics Course Outline (Grades 9-12)

Regression Diagnostics for Survey Data

The assumptions are needed to give us... valid standard errors valid confidence intervals valid hypothesis tests and p-values

Introduction to Linear regression analysis. Part 2. Model comparisons

MODEL SELECTION, ESTIMATION, AND BOOTSTRAP SMOOTHING. Bradley Efron. Division of Biostatistics STANFORD UNIVERSITY Stanford, California

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models

Business Statistics. Lecture 10: Correlation and Linear Regression

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

Inference Tutorial 2

Estimation of AUC from 0 to Infinity in Serial Sacrifice Designs

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk

Psychology 282 Lecture #4 Outline Inferences in SLR

STATS DOESN T SUCK! ~ CHAPTER 16

Data Analysis III. CU- Boulder CHEM-4181 Instrumental Analysis Laboratory. Prof. Jose-Luis Jimenez Spring 2007

Bootstrap confidence levels for phylogenetic trees B. Efron, E. Halloran, and S. Holmes, 1996

Bootstrap Approach to Comparison of Alternative Methods of Parameter Estimation of a Simultaneous Equation Model

Correlation and Linear Regression

Data Analysis and Machine Learning Lecture 12: Multicollinearity, Bias-Variance Trade-off, Cross-validation and Shrinkage Methods.

Bootstrap, Jackknife and other resampling methods

Stat 5101 Lecture Notes

Regression, Ridge Regression, Lasso

y response variable x 1, x 2,, x k -- a set of explanatory variables

Low Bias Bagged Support Vector Machines

Bayesian Inference and the Parametric Bootstrap. Bradley Efron Stanford University

arxiv: v1 [stat.me] 30 Dec 2017

Chapter 27 Summary Inferences for Regression

Bootstrap Confidence Intervals

Inference for Regression Inference about the Regression Model and Using the Regression Line

sphericity, 5-29, 5-32 residuals, 7-1 spread and level, 2-17 t test, 1-13 transformations, 2-15 violations, 1-19

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.

Applied Regression Modeling: A Business Approach Chapter 2: Simple Linear Regression Sections

review session gov 2000 gov 2000 () review session 1 / 38

Sociology 593 Exam 1 Answer Key February 17, 1995

Transcription:

Discussion of Estimation and Accuracy After Model Selection By Bradley Efron Discussant: Lawrence D Brown* Statistics Department, Wharton, Univ. of Penn. lbrown@wharton.upenn.edu JSM, Boston, Aug. 4, 2014 * Joint author: Dan McCarthy (Univ. of Penn.) 1

2 Discussion is about analysis of Cholesterol Data 2 1 0 1 2 0 50 100 Cholesterol data, n=164 subjects: cholesterol decrease plotted versus adjusted compliance; Green curve is OLS cubic regression; Red points indicate 5 featured subjects compliance cholesterol decrease 1 2 3 4 5

Focus on: Initial comment about Post Selection Inference The bagged estimator with Cp selection, vs a SURE mixture of polynomials. Confidence intervals for the predictive mean via a direct double bootstrap 3

Post Selection Inference Efron points out it is bad practice to: (a) look at data (b) choose model (c) fit estimates using chosen model (d) analyze [get CIs] as if pre- chosen as Efron notes Berk,, Zhao (2013), Ann Stat, 802-813 [ PoSI ] make a similar point. But the problem considered there is different from that here: 4

Differences PoSI paper Efron 1 Inference about slopes Inference about E(Y x k ) 2 Model algorithm not pre- Model via C p specified 3 Fixed design Random design for covariates 4 Conventional assumption No assumption on residuals on residuals #1 is important to POSI since the error of slope estimates does not depend on parameters However, #2 for Efron allows narrower CIs. & It s suitable for bootstrapping. #3 for Efron allows for the paired bootstrap. So Efron implements a pairs- bootstrap and gets: 5

6 1 2 3 4 5 6 40 30 20 10 0 10 20 30 Boxplot of Cp boot estimates for Subject 1; B=4000 bootreps; Red bars indicate selection proportions for Models 1 6 only 1/3 of the bootstrap replications chose Model 3 selected model subject 1 estimates Model3 ********* MODEL 3

Embarrassing because: 1. Raises issue of what is true target for the bootstrap. 2. Suggests Efron s estimate from the data may be badly biased. 3. Calls into question the integrity of Cp selection as basis for estimation + CI. 7

SO 4. Efron recommends bagging as a way to produce a more stable and smoother estimator, which may yield more satisfactory estimate and bootstrap CIs. My two main topics A. An alternate method to directly produce an estimate that is an average of polynomials. B. An alternate bootstrap methodology. (Applied to the bagged estimator, but could also apply to the estimator in A.) 8

Alternate Methodology, The SURE- weighted average of polynomials Let v 1, v 2,... be the sequentially orthogonalized versions of x, " " x 2,.... (ie, v 1 1, v 2 1& v 1, ) A weighted average of L.S. polynomials is ˆf (x) = ˆ 0 + " 1 ˆ 1 +...+ " p ˆ p (*) 1 = 1 " 2 "... " p Then ( ) ˆ ( ) 2 ˆ SURE = popsse + 2 % j "1 # 2 + % j "1. Minimize this subject to the monotonicity (*). This yields estimates for the weights $ j 2 9

The unconstrained minimizer of SURE is j;uncon = 1" 1# 1 $ ˆ2 j ( ( )) but the constraint (*) may require a PAV operation that pools adjacent coordinates and produces J- S shrinkage among them. The result is very similar to the bagged estimator, but not exactly the same. (It s nearly indistinguishable on the very benign Cholesterol data.) 10

The Double Bootstrap This is similar to what is suggested in DiCiccio and Efron (1996) Stat Sci, and elsewhere But without any BCa/ABC step. We find it to work well here, and in (the few) other simple and multiple regression examples we ve so far tried it on. Here is a schematic: 11

ŷ i ( ˆB) ŷ i ( ˆB 1 ) ŷ i( ˆB B ) ŷ i ( ˆB 1,1 ) ŷ i( ˆB 1,B 2 ) ŷ i ( ˆB B,1 ) ŷ i( ˆB B,B 2 ) Figure 1: Two Levels of Bootstrap Conditional Mean Estimates for Observation i ŷ( ˆB ) denotes the original bagged estimator. The other levels are paired- bootstraps followed by bagging estimators. Second bootstrap results are used to calibrate the first: 12

For equal- tail intervals the first level bootstrap yields CIs of the form L 2,U 1" 2 that putatively cover with probability 1 ". The second level calibrates (adjusts) the value of so that the actual coverage is estimated to be the desired 1 ". The resulting CIs are equal- tail but need not be symmetric about the point estimate. Two types of 95% CIs were computed (1) CIs for estimates of E(Y xk) at Efron s 5 points (2) Parallel bands that have marginal probability 95% of covering E(Y X). 13

14 3 2 1 0 1 2 0 50 100 Cholesterol data, n=164 subjects: cholesterol decrease plotted versus adjusted compliance; Confidence Intervals Added; Red points indicate 5 featured subjects compliance 1 2 3 4 5 observed data test points single boot point est sngl boot CI quantile dbl boot CI quantile sngl boot CI band dbl boot CI band

Notes: What s labeled single boot point estimate is actually the bagging estimate from the original data. The differences here between the CIs is pretty small. This is attributable to the well- behaved data. For less benign data the differences can be much more notable. Typically (but not always) the double boot CIs are wider than single boot ones. That s so here, but not too noticeable. (But look at the right- most of the 5 points.) The double boot routine is computationally intensive in its own right and much more so because the bagged estimator itself requires 500 bootstrap samples each time it is computed; for the SURE estimator the routine would be quite fast to compute (minutes instead of 4 hours on a parallel array). 15