Better Bootstrap Confidence Intervals

Similar documents
Bootstrap Confidence Intervals

The Nonparametric Bootstrap

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Nonparametric Methods II

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Bootstrap, Jackknife and other resampling methods

ITERATED BOOTSTRAP PREDICTION INTERVALS

11. Bootstrap Methods

The automatic construction of bootstrap confidence intervals

The bootstrap. Patrick Breheny. December 6. The empirical distribution function The bootstrap

Math 494: Mathematical Statistics

Bootstrap and Parametric Inference: Successes and Challenges

Mathematical statistics

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

One-Sample Numerical Data

Confidence intervals for kernel density estimation

MS&E 226: Small Data

Post-exam 2 practice questions 18.05, Spring 2014

The formal relationship between analytic and bootstrap approaches to parametric inference

A Very Brief Summary of Statistical Inference, and Examples

5601 Notes: The Sandwich Estimator

STAT 830 Non-parametric Inference Basics

A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling

Bayesian Inference and the Parametric Bootstrap. Bradley Efron Stanford University

Statistics: Learning models from data

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

Reliable Inference in Conditions of Extreme Events. Adriana Cornea

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

A note on multiple imputation for general purpose estimation

STAT 135 Lab 3 Asymptotic MLE and the Method of Moments

Introduction to Estimation Methods for Time Series models Lecture 2

Mathematical statistics

Elements of statistics (MATH0487-1)

Interval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean

Indian Agricultural Statistics Research Institute, New Delhi, India b Indian Institute of Pulses Research, Kanpur, India

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

A better way to bootstrap pairs

Bootstrap (Part 3) Christof Seiler. Stanford University, Spring 2016, Stats 205

Chapter 3: Maximum Likelihood Theory

Nonparametric confidence intervals. for receiver operating characteristic curves

Statistical Inference

Loglikelihood and Confidence Intervals

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Institute of Actuaries of India

Chapters 9. Properties of Point Estimators

Chapter 8 - Statistical intervals for a single sample

Regression Estimation Least Squares and Maximum Likelihood

Module 22: Bayesian Methods Lecture 9 A: Default prior selection

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

Chapter 2: Resampling Maarten Jansen

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY

Stat 5101 Lecture Notes

simple if it completely specifies the density of x

5.2 Fisher information and the Cramer-Rao bound

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data

UNIVERSITÄT POTSDAM Institut für Mathematik

Finite Population Correction Methods

Bootstrap Confidence Intervals for the Correlation Coefficient

A General Overview of Parametric Estimation and Inference Techniques.

Stat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota

Likelihood-based inference with missing data under missing-at-random

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That

Bootstrap Testing in Econometrics

Outline of GLMs. Definitions

STAT 461/561- Assignments, Year 2015

Monte Carlo Simulations

ST745: Survival Analysis: Nonparametric methods

The assumptions are needed to give us... valid standard errors valid confidence intervals valid hypothesis tests and p-values

Uncertainty Quantification for Inverse Problems. November 7, 2011

GARCH Models Estimation and Inference

STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method

Principles of Statistical Inference

ST495: Survival Analysis: Hypothesis testing and confidence intervals

Statistical techniques for data analysis in Cosmology

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

9 Asymptotic Approximations and Practical Asymptotic Tools

Advanced Econometrics

Primer on statistics:

The Bootstrap Small Sample Properties. F.W. Scholz Boeing Computer Services Research and Technology

Application of Variance Homogeneity Tests Under Violation of Normality Assumption

Preliminaries The bootstrap Bias reduction Hypothesis tests Regression Confidence intervals Time series Final remark. Bootstrap inference

Preliminaries The bootstrap Bias reduction Hypothesis tests Regression Confidence intervals Time series Final remark. Bootstrap inference

Lecture 7 Introduction to Statistical Decision Theory

Interval Estimation III: Fisher's Information & Bootstrapping

Some Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables?

Hypothesis Testing. A rule for making the required choice can be described in two ways: called the rejection or critical region of the test.

3.3 Estimator quality, confidence sets and bootstrapping

Spring 2012 Math 541B Exam 1

Theory of Maximum Likelihood Estimation. Konstantin Kashin

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X).

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

Can we do statistical inference in a non-asymptotic way? 1

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Statistics 3858 : Maximum Likelihood Estimators

Transcription:

by Bradley Efron University of Washington, Department of Statistics April 12, 2012

An example Suppose we wish to make inference on some parameter θ T (F ) (e.g. θ = E F X ), based on data We might suppose that X i i.i.d. F F. F = {F : E F X 2 < }, or we might let F = {F θ : θ Θ R p }. We often concern ourselves with two tasks: Point estimates ˆθ Interval estimates (ˆθ[α], ˆθ[1 α]) where ˆθ[α] denotes an α level endpoint of a confidence interval.

An example In our frequentist hats, the two tasks usually start (and end) with maximum likelihood estimation. We turn our frequentist crank, finding The MLE ˆθ MLE The observed Fisher information Î n. We then use the relationship n(ˆθ MLE θ) d N (0, Î 1 n to pivot around θ and generate an interval estimate (α) Î 1 ˆθ MLE ± z n n )

An example But what can go wrong? n(ˆθmle θ) d N (0, Î 1 n is just an approximation. For small sample sizes, we have no guarantees. Strong skewness or heavy tails in data distribution may distort the sampling distribution of ˆθ. Result: Badly calibrated confidence intervals! )

An example So what can we do about that? Ad-hoc solutions: Find some monotone transformation g that makes g(ˆθ) really approximately normal Get lucky: identify the exact distribution of ˆθ or some transformation thereof This can be hard!

An example For example, consider X i i.i.d. N(µ, θ) ˆθ = 1 n 1 (x i x) 2. Suppose θ = 1. Normal theory yields that ˆθ χ2 (n 1) (n 1). i Sampling distribution of sample variance f(x) 0.0 0.2 0.4 0.6 0.8 1.0 1.2 θ 0.0 0.5 1.0 1.5 2.0 2.5 θ^

An example When the data are normal we also have that Var(ˆθ) = 2σ4 n 1. Applying asymptotic approximations, an approximate 95% confidence interval would be 2ˆθ ˆθ ± 1.96 2 n 1 and an exact 95% confidence interval for θ is ( 1 1 (n 1)ˆθ, χ 2 n 1,.975 χ 2 n 1,.025 For n = 20, the approximate intervals have 88% coverage. )

An example Suppose that last two ad-hoc solutions are hard, and our interval estimates perform badly. What else can we do? Could we adjust for the skewness or kurtosis that mucks up our intervals? This is the road taken by Abramovitch & Singh (1985), Bartlett (1953), Hall (1983), to name a few. The general idea is that the true confidence limits often exist in the form of ˆθ[α] = ˆθ + ˆσ(z (α) + A(α) n + B(α) n n n + ). These methods generally boil down to estimating some function of higher order moments to correct the interval.

An example Are we in the clear yet? No. These methods can become exceedingly cumbersome to apply in multiparameter families, especially for more exotic transformations of parameters (e.g. ratios of means, ratios of regression parameters, products of parameters...). What are we to do?

Bootstrap overview BC a method You can pull yourself up by your bootstraps and you don t need anything else -Efron

Bootstrap overview BC a method The bootstrap is a data resampling technique that allows us to use the empirical distribution function F n to estimate the sampling distribution of our estimator ˆθ. Recall F n (t) = 1 n n i=1 1 [t x i ]. For b = 1,..., B: Sample X (b) i i.i.d. F n i = 1,..., n. Compute ˆθ (b) from data (X (b) 1,..., X (b) n ). Use the bootstrap distribution of ˆθ to compute quantities of interest. E.g. (ˆθ (Bα), ˆθ (B(1 α)) ) would give an approximate 100(1-2α)% confidence interval for θ.

Bootstrap overview BC a method The bootstrap technique applies to a huge variety of parameters in both parametric and nonparametric families. Since its inception the technique has seen various refinements. The primary contribution of this talk s namesake is the accelerated bootstrap confidence interval (BC a ). The method s goal: Automatically create confidence intervals that adjust for underlying higher order effects.

Bootstrap overview BC a method For simplicity, we start by assuming that ˆθ f θ estimates θ. We make the following assumptions regarding our estimator. For our family f θ, there exists a monotone increasing transformation g and constants z 0 and a such that ˆφ = g(ˆθ) φ = g(θ) satisfy with ˆφ = φ + σ φ (Z z 0 ) Z N(0, 1) σ φ = 1 + aφ The constant z 0 is the bias correction constant The constant a is the acceleration constant

Bootstrap overview BC a method Let Ĝ(s) = P{ˆθ < s} denote the bootstrap distribution function. We can either estimate this via Monte Carlo or use ˆθ fˆθ. The BC a interval is defined as below. Lemma Under the conditions in the previous slide the correct central confidence interval of level 1-2α for θ is [Ĝ 1 (Φ(z[α])), Ĝ 1 (Φ(z[1 α]))] where z[α] = z 0 + (z 0 + z (α) ) 1 a(z 0 + z (α) )

Bootstrap overview BC a method Note that the form of the transformation g never comes into play. In a sense, the method automatically selects a transformation that brings ˆθ to normality, computes an exact 95% interval, and then transforms backwards to reach the θ scale again. We require estimates of a and z 0 to make this work: z 0 Φ 1 (Ĝ(ˆθ)) a 1 6 SKEW θ=ˆθ ( l θ ) I.e. the skew of the score transformation of ˆθ

Bootstrap overview BC a method A main theorem of the paper is that this interval is second-order correct in the sense that the endpoints of the BC a confidence intervals are very close to the true exact endpoints, ˆθ BC [α] ˆθ EX [α] = O p (n 3 2 ). That is to say, the BC a method successfully captures A(α) n n below: ˆθ[α] = ˆθ + ˆσ(z (α) + A(α) n + B(α) n n n + ). In the previous sample variance example, BC a intervals have estimated coverage essentially identical with the exact intervals.

Bootstrap overview BC a method Looking ahead: This method is also second order correct in multiparameter families, with different estimates of a and z 0 The method also works in nonparametric settings, via an extension from multiparameter families Lots of math