Bootstrap Confidence Intervals

Similar documents
Better Bootstrap Confidence Intervals

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

The bootstrap. Patrick Breheny. December 6. The empirical distribution function The bootstrap

The Nonparametric Bootstrap

Bootstrap tests. Patrick Breheny. October 11. Bootstrap vs. permutation tests Testing for equality of location

Empirical Likelihood

4 Resampling Methods: The Bootstrap

Relative efficiency. Patrick Breheny. October 9. Theoretical framework Application to the two-group problem

Bootstrap (Part 3) Christof Seiler. Stanford University, Spring 2016, Stats 205

Nonparametric Methods II

ITERATED BOOTSTRAP PREDICTION INTERVALS

11. Bootstrap Methods

The automatic construction of bootstrap confidence intervals

Loglikelihood and Confidence Intervals

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

One-Sample Numerical Data

Bootstrap, Jackknife and other resampling methods

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf

Supporting Information for Estimating restricted mean. treatment effects with stacked survival models

Indian Agricultural Statistics Research Institute, New Delhi, India b Indian Institute of Pulses Research, Kanpur, India

ST745: Survival Analysis: Nonparametric methods

Inference for P(Y<X) in Exponentiated Gumbel Distribution

STAT 830 Non-parametric Inference Basics

Model Selection, Estimation, and Bootstrap Smoothing. Bradley Efron Stanford University

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Analytical Bootstrap Methods for Censored Data

The t-test Pivots Summary. Pivots and t-tests. Patrick Breheny. October 15. Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/18

Bootstrap confidence levels for phylogenetic trees B. Efron, E. Halloran, and S. Holmes, 1996

Semiparametric Regression

This produces (edited) ORDINARY NONPARAMETRIC BOOTSTRAP Call: boot(data = aircondit, statistic = function(x, i) { 1/mean(x[i, ]) }, R = B)

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Finding a Better Confidence Interval for a Single Regression Changepoint Using Different Bootstrap Confidence Interval Procedures

Bayesian Inference and the Parametric Bootstrap. Bradley Efron Stanford University

Permutation tests. Patrick Breheny. September 25. Conditioning Nonparametric null hypotheses Permutation testing

STAT Section 2.1: Basic Inference. Basic Definitions

SOLUTION FOR HOMEWORK 4, STAT 4352

EXPLICIT NONPARAMETRIC CONFIDENCE INTERVALS FOR THE VARIANCE WITH GUARANTEED COVERAGE

STAT 6350 Analysis of Lifetime Data. Probability Plotting

STATISTICAL INFERENCE IN ACCELERATED LIFE TESTING WITH GEOMETRIC PROCESS MODEL. A Thesis. Presented to the. Faculty of. San Diego State University

Permutation Tests. Noa Haas Statistics M.Sc. Seminar, Spring 2017 Bootstrap and Resampling Methods

Finite Population Correction Methods

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions

Preliminaries The bootstrap Bias reduction Hypothesis tests Regression Confidence intervals Time series Final remark. Bootstrap inference

Preliminaries The bootstrap Bias reduction Hypothesis tests Regression Confidence intervals Time series Final remark. Bootstrap inference

The Weibull Distribution

Contents 1. Contents

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions

Censoring mechanisms

Lecture 32: Asymptotic confidence sets and likelihoods

INTERVAL ESTIMATION AND HYPOTHESES TESTING

inferences on stress-strength reliability from lindley distributions

A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling

Kernel density estimation

Chapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics

On the Accuracy of Bootstrap Confidence Intervals for Efficiency Levels in Stochastic Frontier Models with Panel Data

The formal relationship between analytic and bootstrap approaches to parametric inference

The comparative studies on reliability for Rayleigh models

Estimation of Quantiles

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X).

Estimation of Stress-Strength Reliability Using Record Ranked Set Sampling Scheme from the Exponential Distribution

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Spring 2012 Math 541B Exam 1

Higher-Order Confidence Intervals for Stochastic Programming using Bootstrapping

P Values and Nuisance Parameters

The Bootstrap Small Sample Properties. F.W. Scholz Boeing Computer Services Research and Technology

36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

Ridge regression. Patrick Breheny. February 8. Penalized regression Ridge regression Bayesian interpretation

Hypothesis Testing with the Bootstrap. Noa Haas Statistics M.Sc. Seminar, Spring 2017 Bootstrap and Resampling Methods

Reliable Inference in Conditions of Extreme Events. Adriana Cornea

Bayesian Inference. Chapter 2: Conjugate models

Selection should be based on the desired biological interpretation!

Chapter 6. Estimation of Confidence Intervals for Nodal Maximum Power Consumption per Customer

Probability and Statistics

The assumptions are needed to give us... valid standard errors valid confidence intervals valid hypothesis tests and p-values

NONINFORMATIVE NONPARAMETRIC BAYESIAN ESTIMATION OF QUANTILES

Stat 710: Mathematical Statistics Lecture 31

STAT440/840: Statistical Computing

A Simulation Comparison Study for Estimating the Process Capability Index C pm with Asymmetric Tolerances

Bootstrap. Director of Center for Astrostatistics. G. Jogesh Babu. Penn State University babu.

Chapter 2: Resampling Maarten Jansen

The Bootstrap Suppose we draw aniid sample Y 1 ;:::;Y B from a distribution G. Bythe law of large numbers, Y n = 1 B BX j=1 Y j P! Z ydg(y) =E

ECE 275A Homework 7 Solutions

Estimation of AUC from 0 to Infinity in Serial Sacrifice Designs

Resampling and the Bootstrap

Stat 5101 Lecture Notes

Rank conditional coverage and confidence intervals in high dimensional problems

An analogy from Calculus: limits

Statistical Inference on Constant Stress Accelerated Life Tests Under Generalized Gamma Lifetime Distributions

Outline. Confidence intervals More parametric tests More bootstrap and randomization tests. Cohen Empirical Methods CS650

University of California San Diego and Stanford University and


Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

1 Glivenko-Cantelli type theorems

5601 Notes: The Subsampling Bootstrap

Double Bootstrap Confidence Interval Estimates with Censored and Truncated Data

Post-exam 2 practice questions 18.05, Spring 2014

ECO Class 6 Nonparametric Econometrics

The Bootstrap: Theory and Applications. Biing-Shen Kuo National Chengchi University

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

Transcription:

Bootstrap Confidence Intervals Patrick Breheny September 18 Patrick Breheny STA 621: Nonparametric Statistics 1/22

Introduction Bootstrap confidence intervals So far, we have discussed the idea behind the bootstrap and how it can be used to estimate standard errors Standard errors are often used to construct confidence intervals based on the estimate having a normal sampling distribution: ˆθ ± z 1 α/2 SE; alternatively, the interval could be based on the t distribution The bootstrap SE can be used in this way as well Patrick Breheny STA 621: Nonparametric Statistics 2/22

Introduction Bootstrap confidence intervals However, recall that the bootstrap can also be used to estimate the CDF G of ˆθ Thus, with the bootstrap, we do not need to make assumptions/approximations concerning the sampling distribution of ˆθ we can estimate it as part of the confidence interval procedure This has been an active area of theoretical research into the bootstrap Patrick Breheny STA 621: Nonparametric Statistics 3/22

For example, suppose that the standard error of an estimate varies with the size of the estimate If this is so, then our confidence interval should be wider on the right side than it is on the left One way to implement this idea is to estimate the SE separately for each bootstrap replication this is the idea behind the bootstrap-t interval Patrick Breheny STA 621: Nonparametric Statistics 4/22

: Procedure The procedure of the bootstrap-t interval is as follows: (1) For each bootstrap sample, compute z b = ˆθ b ˆθ ŜE b where ŜE b is an estimate of the standard error of ˆθ based on the data in the bth bootstrap sample (2) Estimate the αth percentile of z by the value ˆt α such that B 1 b I(z b ˆt α ) = α (3) A 1 α confidence interval for θ is then (ˆθ ˆt 1 α/2 ŜE, ˆθ ˆt α/2 ŜE) Patrick Breheny STA 621: Nonparametric Statistics 5/22

: Example As a small example, the survival times of 9 rats were 10, 27, 30, 40, 46, 51, 52, 104, and 146 days Consider estimating the mean; the point estimates are ˆθ = 56.2 and ŜE = 14.1 The percentile points for a 95% confidence interval: Normal -1.96 1.96 t -2.31 2.31 Bootstrap-t -4.86 1.61 This translates into the following confidence intervals: Normal 28.5 84.0 t 23.6 88.9 Bootstrap-t 33.4 125.1 Patrick Breheny STA 621: Nonparametric Statistics 6/22

: R If you want to implement this confidence interval in R, your function that you pass to boot will need to return two things: ˆθ b and ˆV(ˆθ b ) For example: mean.boot <- function(x, ind) { c(mean(x[ind]), var(x[ind])/length(x)) } out <- boot(x, mean.boot, 999) boot.ci(out) The function boot.ci returns the bootstrap-t interval (which it calls the studentized interval) along with the normal interval and some other intervals which we will talk about next Patrick Breheny STA 621: Nonparametric Statistics 7/22

Double bootstrap Estimation of ŜE b is straightforward for the mean (and a few other statistics), but what about in general? Once again, the functional delta method and jackknife are options The bootstrap is also an option here we use the bootstrap within a bootstrap replication to estimate the standard error of a bootstrap replication and is called the double bootstrap The obvious drawback to the double bootstrap is that it requires B 2 bootstrap replications, and is therefore time-consuming to compute Patrick Breheny STA 621: Nonparametric Statistics 8/22

Confidence intervals and sampling distributions The next interval we will discuss is not based on the usual pivoting ideas we encounter when constructing confidence intervals Instead, it is based on the observation that, if ˆθ N(ˆθ, ŜE2 ), ˆθ ± ŜEz 1 α/2 = (ˆθ α/2, ˆθ 1 α/2 ) is a 1 α level confidence interval for θ, where ˆθ α is the αth percentile of the distribution of ˆθ Letting Ĝ the empirical CDF of ˆθ, another way of writing the above interval is [Ĝ 1 (α/2), Ĝ 1 (1 α/2)] Thus, there is a connection (at least, if ˆθ Normal) between the resampling distribution of ˆθ and confidence intervals for θ Patrick Breheny STA 621: Nonparametric Statistics 9/22

This interval called the percentile interval is much simpler to construct than the bootstrap-t, and turns out to work surprisingly well in a wide range of examples has an additional justification beyond approximation to the normal: it is invariant to transformations of θ For example, suppose we wish to construct a confidence interval for φ = m(θ), where m is a monotone transformation We do not need to generate new bootstrap replications: ˆφ b = m(ˆθ b ), because ( ˆφα/2, ˆφ ) [ ] 1 α/2 = m(ˆθ α/2 ), m(ˆθ 1 α/2 ) Patrick Breheny STA 621: Nonparametric Statistics 10/22

Percentile interval theorem In the 20th century, statisticians explored dozens of transformations of various statistics designed to make the sampling distribution of the statistic more normal, from simple ones like the log-odds to more complex ones like the Fisher and Anscombe transformations One compelling justification for the percentile interval is that, if such a transformation exists, the percentile interval will find it automatically Theorem: Suppose there exists φ = m(θ) such that ( ˆφ ) N ˆφ, τ 2 for some τ. Then the percentile interval (ˆθ α/2, ˆθ 1 α/2 ) is equivalent to the optimal pivot-based interval based on the above relationship. Patrick Breheny STA 621: Nonparametric Statistics 11/22

Example: Automatic transformation As an example, suppose X 1,..., X 10 iid N(0, 1), and we are interested in estimating θ = e µ In this case, there does exist a normalizing transformation φ = log(θ) such that ˆφ N(φ, τ 2 ) Suppose, however, that we didn t know any of this and applied the ordinary nonparametric bootstrap: ˆθ L ˆθU Exact 0.49 2.29 Normal bootstrap 0.15 1.81 Percentile interval 0.50 2.10 Patrick Breheny STA 621: Nonparametric Statistics 12/22

Percentile intervals in R Percentile intervals are also returned by boot.ci They can also be obtained by applying quantile to the output of boot For the mouse survival data from earlier, Normal 28.5 84.0 t 23.6 88.9 Bootstrap-t 33.4 125.1 Percentile 32.1 87.7 Patrick Breheny STA 621: Nonparametric Statistics 13/22

Standard vs. bootstrap approach The standard approach is based on taking literally the following asymptotic relationship ˆθ θ σ N(0, 1) The bootstrap percentile interval relaxes that assumption; instead of requiring normality of ˆθ, it requires only that m : φ = m(θ) satisfies ˆφ φ N(0, 1) τ However, this still requires that there exists a single transformation that is both normalizing and variance-stabilizing often, such a transformation does not exist Patrick Breheny STA 621: Nonparametric Statistics 14/22

Generalizing the bootstrap assumptions Thus, in a brilliant 1987 paper, Efron considers a generalization of the bootstrap assumptions Suppose that for some monotone increasing transformation m, some bias constant z 0, and some acceleration constant a, the following relationship holds for φ = m(θ): ˆφ φ N( z 0, 1), σ σ = 1 + aφ Efron named the confidence interval based on this assumption the BC a interval, because it corrects for both bias and acceleration of the variance Patrick Breheny STA 621: Nonparametric Statistics 15/22

The BC a interval Theorem: If the relationship on the previous slide holds, then the following interval is correct, in the sense of being the optimal pivot-based interval: [Ĝ 1 θ {Φ(z[α])}, Ĝ 1 {Φ(z[1 α])}] where z[α] = z 0 + z 0 + z (α) 1 a(z 0 + z (α) ). Note that, in order to construct the BC a interval, one has to estimate z 0 and a, but, like the percentile interval, it is not necessary to know m Patrick Breheny STA 621: Nonparametric Statistics 16/22

Example revisited Unfortunately, we don t have time to cover in detail the derivation of this interval and the estimation of z 0 and a, but it is available from boot.ci When applied to our mouse survival data, we have ẑ 0 = 0.026 and â = 0.066, which gives us the adjusted percentage points (0.037, 0.986) and the resulting interval Normal 28.5 84.0 t 23.6 88.9 Bootstrap-t 33.4 125.1 Percentile 32.1 87.7 BC a 37.3 90.9 Patrick Breheny STA 621: Nonparametric Statistics 17/22

Asymptotic accuracy of confidence intervals Let us denote a confidence interval procedure by its confidence points ˆθ[α], where ideally, P(θ ˆθ[α]) = α A confidence point is called first-order accurate if and second-order accurate if P(θ ˆθ[α]) = α + O(n 1/2 ) P(θ ˆθ[α]) = α + O(n 1 ) Patrick Breheny STA 621: Nonparametric Statistics 18/22

Second-order accuracy of bootstrap confidence intervals It can be shown (our textbook has some details on this and the regularity conditions that are required) that the standard interval and the percentile interval are first-order accurate, while the bootstrap-t and are second-order accurate regardless of the true distribution, F This is a powerful theoretical justification for the bootstrap-t and Arguments can be made for either interval in different situations The BC a interval is also transformation-invariant and range-preserving, meaning that, if it is not possible for a statistic or a function of a statistic to lie outside a certain range [a, b], then the BC a interval will be contained in [a, b] (the bootstrap-t intervals are neither transformation-invariant nor range-preserving) Patrick Breheny STA 621: Nonparametric Statistics 19/22

Bootstrap failure #1 Bootstrap confidence intervals are unquestionably a tremendous methodological advance, and the BC a interval represents the state of the art as far as nonparametric confidence intervals goes However, bootstrap intervals are limited by the accuracy of ˆF The empirical CDF is generally poorly estimated at the tails of a distribution Consequently, it is difficult to produce nonparametric confidence intervals for statistics that are highly influenced by distribution tails Patrick Breheny STA 621: Nonparametric Statistics 20/22

Bootstrap failure #2 Furthermore, the bootstrap still requires conditions such as Hadamard differentiability and can fail to produce accurate intervals when those conditions do not hold We have already encountered the example of density estimation As an additional example, consider ˆθ = max(x i ) The BC a interval can never exceed Ĝ 1 (1) = max(x i ), and therefore provides a 0% confidence interval for any continuous distribution Patrick Breheny STA 621: Nonparametric Statistics 21/22

Homework Bootstrap confidence intervals In an earlier homework, we examined the coverage probabilities of parametric and nonparametric confidence intervals for the variance Your assignment now is to compare three nonparametric methods: the functional delta method, the bootstrap percentile interval, and the BC a interval Homework: Conduct a simulation study to determine how the coverage probability and average interval width of these two intervals varies with the sample size n, when (i) the data is truly normally distributed, and (ii) the data follows an exponential distribution. For each distribution, produce a plot of coverage probability versus sample size, with lines representing the various methods, as well as a corresponding plot for interval width. Patrick Breheny STA 621: Nonparametric Statistics 22/22