Symmetric Tests and Condence Intervals for Survival Probabilities and Quantiles of Censored Survival Data Stuart Barber and Christopher Jennison Depar

Similar documents
A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky

Analytical Bootstrap Methods for Censored Data

11. Bootstrap Methods

Power Calculations for Preclinical Studies Using a K-Sample Rank Test and the Lehmann Alternative Hypothesis

Supporting Information for Estimating restricted mean. treatment effects with stacked survival models

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

11 Survival Analysis and Empirical Likelihood

Group Sequential Tests for Delayed Responses. Christopher Jennison. Lisa Hampson. Workshop on Special Topics on Sequential Methodology


Empirical likelihood ratio with arbitrarily censored/truncated data by EM algorithm

Double Bootstrap Confidence Interval Estimates with Censored and Truncated Data

A comparison study of the nonparametric tests based on the empirical distributions

Mantel-Haenszel Test Statistics. for Correlated Binary Data. Department of Statistics, North Carolina State University. Raleigh, NC

University of California, Berkeley

Monitoring clinical trial outcomes with delayed response: incorporating pipeline data in group sequential designs. Christopher Jennison

The Design of a Survival Study

ST745: Survival Analysis: Nonparametric methods

Quantile Regression for Residual Life and Empirical Likelihood

Statistics and Probability Letters. Using randomization tests to preserve type I error with response adaptive and covariate adaptive randomization

Comparisons of Approximate Confidence Interval Procedures for Type I Censored Data

Typical Survival Data Arising From a Clinical Trial. Censoring. The Survivor Function. Mathematical Definitions Introduction

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances

Chapter 2 Inference on Mean Residual Life-Overview

Analysis of incomplete data in presence of competing risks

UNIVERSITY OF CALIFORNIA, SAN DIEGO

Adaptive Designs: Why, How and When?

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints

PROD. TYPE: COM ARTICLE IN PRESS. Computational Statistics & Data Analysis ( )

Group Sequential Designs: Theory, Computation and Optimisation

The Nonparametric Bootstrap

Product-limit estimators of the survival function with left or right censored data

arxiv: v3 [stat.me] 24 Mar 2018

Introduction to Statistical Analysis

Optimising Group Sequential Designs. Decision Theory, Dynamic Programming. and Optimal Stopping

TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOZIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississip

ON THE CONSEQUENCES OF MISSPECIFING ASSUMPTIONS CONCERNING RESIDUALS DISTRIBUTION IN A REPEATED MEASURES AND NONLINEAR MIXED MODELLING CONTEXT

Group Sequential Tests for Delayed Responses

Goodness-of-fit tests for randomly censored Weibull distributions with estimated parameters

Power and Sample Size Calculations with the Additive Hazards Model

Discussion Paper Series

Can we do statistical inference in a non-asymptotic way? 1

Heteroskedasticity-Robust Inference in Finite Samples

A nonparametric two-sample wald test of equality of variances

Interim Monitoring of Clinical Trials: Decision Theory, Dynamic Programming. and Optimal Stopping

Lecture 3. Truncation, length-bias and prevalence sampling

GROUPED SURVIVAL DATA. Florida State University and Medical College of Wisconsin

Reports of the Institute of Biostatistics

Comparing Adaptive Designs and the. Classical Group Sequential Approach. to Clinical Trial Design

Likelihood ratio confidence bands in nonparametric regression with censored data

Group-Sequential Tests for One Proportion in a Fleming Design

Statistical Methods for Alzheimer s Disease Studies

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions

Power Comparison of Exact Unconditional Tests for Comparing Two Binomial Proportions

Approximate and Fiducial Confidence Intervals for the Difference Between Two Binomial Proportions

For right censored data with Y i = T i C i and censoring indicator, δ i = I(T i < C i ), arising from such a parametric model we have the likelihood,

Nonparametric Hypothesis Testing and Condence Intervals with. Department of Statistics. University ofkentucky SUMMARY

AFT Models and Empirical Likelihood

Confidence intervals for kernel density estimation

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

A Simulation Study on Confidence Interval Procedures of Some Mean Cumulative Function Estimators

COMPUTE CENSORED EMPIRICAL LIKELIHOOD RATIO BY SEQUENTIAL QUADRATIC PROGRAMMING Kun Chen and Mai Zhou University of Kentucky

ADJUSTED POWER ESTIMATES IN. Ji Zhang. Biostatistics and Research Data Systems. Merck Research Laboratories. Rahway, NJ

Estimation of AUC from 0 to Infinity in Serial Sacrifice Designs

Empirical Likelihood in Survival Analysis

Interval Estimation for Parameters of a Bivariate Time Varying Covariate Model

COMPUTATION OF THE EMPIRICAL LIKELIHOOD RATIO FROM CENSORED DATA. Kun Chen and Mai Zhou 1 Bayer Pharmaceuticals and University of Kentucky

Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring

MAS3301 / MAS8311 Biostatistics Part II: Survival

Part III Measures of Classification Accuracy for the Prediction of Survival Times

GOODNESS-OF-FIT TESTS FOR ARCHIMEDEAN COPULA MODELS

Optimum Hybrid Censoring Scheme using Cost Function Approach

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials

FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH

Distribution Fitting (Censored Data)

Inference via Kernel Smoothing of Bootstrap P Values

Estimation and sample size calculations for correlated binary error rates of biometric identification devices

TMA 4275 Lifetime Analysis June 2004 Solution

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST

Meei Pyng Ng 1 and Ray Watson 1

ingestion of selenium tablets for plasma levels to rise; another explanation may be that selenium aects only initiation not promotion of tumors, so tu

COMPARISON OF THE ESTIMATORS OF THE LOCATION AND SCALE PARAMETERS UNDER THE MIXTURE AND OUTLIER MODELS VIA SIMULATION

BIAS OF MAXIMUM-LIKELIHOOD ESTIMATES IN LOGISTIC AND COX REGRESSION MODELS: A COMPARATIVE SIMULATION STUDY

In Defence of Score Intervals for Proportions and their Differences

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA

Smooth nonparametric estimation of a quantile function under right censoring using beta kernels

Performance of Low Density Parity Check Codes. as a Function of Actual and Assumed Noise Levels. David J.C. MacKay & Christopher P.

and Comparison with NPMLE

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints and Its Application to Empirical Likelihood

Part III. Hypothesis Testing. III.1. Log-rank Test for Right-censored Failure Time Data

Investigation of goodness-of-fit test statistic distributions by random censored samples

Step-Stress Models and Associated Inference

Prediction of ordinal outcomes when the association between predictors and outcome diers between outcome levels

M O N A S H U N I V E R S I T Y

Normal Probability Plot Probability Probability

Early selection in a randomized phase II clinical trial

A Note on Bayesian Inference After Multiple Imputation

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations

Sample size re-estimation in clinical trials. Dealing with those unknowns. Chris Jennison. University of Kyoto, January 2018

Published online: 10 Apr 2012.

Introduction to Empirical Processes and Semiparametric Inference Lecture 01: Introduction and Overview

Transcription:

Symmetric Tests and Condence Intervals for Survival Probabilities and Quantiles of Censored Survival Data Stuart Barber and Christopher Jennison Department of Mathematical Sciences, University of Bath, Bath, BA2 7AY, U.K. Summary We describe existing tests and introduce two new tests concerning the value of a survival function. These tests may be used to construct a condence interval for the survival probability atagiven time or for a quantile of the survival distribution. Simulation studies show that error rates can dier substantially from their nominal values, particularly at survival probabilities close to 0 or 1. We recommend our new \constrained bootstrap" test for its good overall performance. 1. Introduction We consider inference on the survivor function, S(t) = pr(t >t), from survival times subject to independent right censoring. We present size tests of a hypothesis = p, noting that a family of such tests can be inverted to obtain 100(1, )% condence intervals for the survival probability at a specic time or for a quantile of S(t): the condence interval for is the set of values p for which = p is accepted and the condence interval for S,1 (p ) is the set of values t for which S(t) =p is accepted. The condence interval for lies above p if H 0 : = p is rejected in favour of > p and below p if H 0 is rejected in favour of < p. Under H 0, these these two errors usually have diering practical consequences, thus it is appropriate to require a 100(1, )% condence interval to lie entirely above or below the true survival probability with probability =2 each. This requires a two-tailed parent test with type I error divided equally between both tails. Key words: Beta distribution; Bootstrap sampling; Censoring; Condence interval; Greenwood's formula; Hypothesis test; Kaplan{Meier estimate; Likelihood ratio test; Survival data. 1

Our objective is, therefore, to nd a test of H 0 : =p, achieving type I error close to, and rejecting H 0 in favour of >p and <p with probability approximately =2 each. Previous work in this area has concentrated on the median survival time; see Slud, Byar and Green (1984) and Strawderman, Parzen and Wells (1997). We have found additional diculties in testing =p for values of p away from 0.5 and so seek a test which performs well over a broad range of p values. We rst consider tests based on the Kaplan-Meier estimate S(t b ) which refer the distribution of b S(t ) to normal, beta and bootstrap approximations. In all these cases, the estimated distribution of b S(t ) under H 0 : = p is greatly improved by making explicit use of the fact = p. When testing whether a binomial probability is equal to p, it is standard practice to take the null variance of bp from a sample of size n to be p (1, p )=n rather than bp(1, bp)=n, but analogous methods have often been overlooked in the related problem of testing a hypothesised survival probability for right-censored data. The two new tests which we introduce, based on a beta approximation and a \constrained" bootstrap, prove to be the most successful tests of this type. Lastly, we consider Thomas and Grunkemeier's (1975) likelihood ratio test. Despite being an early proposal, this test has received little attention and we are not aware of any small sample results beyond Thomas and Grunkemeier's limited simulations. We have found this test to be superior to many more recent proposals and well worth further consideration. 2. Methods for testing H 0 : =p 2.1 Normal approximations to the distribution of S(t b ) We assume failure times are from a distribution with survivor function S(t) and the observed data consist of times t i and indicator variables i ;i = 1;:::;n; with i = 0 if t i is an exact failure time and i = 1 if t i is a censoring time. Let t (1) < ::: < t (r) be the distinct ordered failure times and denote the number of subjects at risk just before t (j) 2

by n j = #fi : t i t (j) g;j = 1;:::;r. Dening d j to be the number of failures at time t (j), the Kaplan{Meier estimate of S(t) (Kaplan and Meier, 1958) is bs(t) = Y j:t (j) t n j, d j n j ; 0 t max i ft i g: Meier (1975) showed the distribution of b S(t) to be approximately normal with mean S(t) and variance consistently estimated by Greenwood's (1926) formula b 2 G(t) =f b S(t)g 2 X j:t (j) t d j n j (n j, d j ) : Hence, we obtain the \Greenwood test" which rejects H 0 : = p if b S(t ) 62 fp z =2 b G (t )g, where z =2 is the value exceeded with probability =2 by a N(0; 1) variate. In approximating the distribution of S(t b ) under H 0 : =p, it is reasonable to use the fact =p. The maximum likelihood estimate of S(t) for 0 t t with probability masses only at observed failure times and subject to the constraint S(t b )=p is S(t; t ;p )= Y n j +, d j j:t (j) t n j + ; 0 t t ; (1) where is the unique solution to S(t; t ;p ) = p. Using S(t; t ;p ) instead of b S(t) in the derivation of Greenwood's formula, we obtain the constrained estimate of varf b g, b 2 C(t ; p )=(p ) 2 X j:t (j) t b S(t,(j) ) q j n j S(t, (j) ; t ;p )p j ; where p j = (n j +, d j )=(n j + ), q j = 1, p j and b S(t,) and S(t,; t ;p ) denote values of b S(t) and S(t; t ;p ) just before time t. This estimate is due to Thomas and Grunkemeier (1975). It is used by Jennison and Turnbull (1985), although the formula they state (p.620) is in error as t j occurs in two places where t, j is needed. If =p, b 2 C(t ; p ) and b 2 G(t ) are asymptotically equivalent, but b 2 C has the advantage of reducing to the correct binomial variance, p (1, p )=n, in the absence of censoring. 3

The \constrained variance" test rejects H 0 : = p if S(t b ) 62 fp z =2 b C (t ; p )g. However, if no exact failures are observed before time t, S(t; t ;p ) is undened and we test H 0 directly but conservatively. Let n c denote the number of observations censored before t. Placing the failure probability 1, p just before t, the number of subjects surviving past t, conditional on n c being censored before t, is Bin(n, n c ;p ). In the observed data all n, n c uncensored cases survived past t, and we reject H 0 if (p ) (n,n c) <=2. Figure 1 shows empirical type I error rates for the Greenwood and constrained variance tests in 5,000 simulated data sets. In each case, sample size was 100, failure times were exponentially distributed with mean 10, denoted Exp(10), and censoring times were Exp(50), so the probability of censoring was 1/6. In this and subsequent gures, results are for p = 0:1; 0:15;:::;0:9 and t such that = p, the left and right bars of each pair represent rejection of H 0 : =p due to low and high values of S(t b ) respectively. The intended two-sided type I error is =0:05 and the horizontal line at =2 =0:025 indicates the target height for each bar. The empirical one-sided error rate from a sample of 5,000 will lie within the dotted lines at 0:025 0:006 with probability 0.95 if the true rate is 0.025. Figure 1 shows that the Greenwood test performs well when = 0:5 but poorly at high or low values of. The results are strongly asymmetric, most errors arising with b too high for high and too low for low. of the constrained variance test are closer to their target values over the whole range of, although there are still noticeable discrepancies for values of above 0.7 or below 0.35. We have also simulated other survival and censoring distributions with similar results. Other estimates of varfs(t b )g have been proposed. We have used Peto et al.'s (1977) suggestion of b 2 P (t)=f b S(t)g 2 f1, b S(t)g=n k for t (k) t<t (k+1), but found this test to fare worse than the Greenwood test, especially at the high or low values of for which this variance estimate is intended. Rothman (1978) approximates varfs(t b )g under H 0 :=p by 4

b 2 R(t ; p )= p (1, p ) bf1, b g b 2 G(t ): This estimate equals b 2 G(t )if b =p and, like the binomial variance for uncensored data, is proportional to p (1,p ). In fact, Rothman's estimate gives numerical values close to the constrained variance estimate, b 2 C(t ; p ), and tests based on these two variance estimates have similar properties, the Rothman test doing slightly worse for below 0.2. 2.2 Other approximations to the distribution of b Further study of simulation results helps explain the empirical error rates seen so far. Barber and Jennison (1998) demonstrate that the estimate b 2 G(t )varies systematically with b, over-estimating varf b g when b is closer to 0.5 than and under-estimating when b is farther from 0.5, and this causes the asymmetric error rates in the two tails of the Greenwood test. In contrast b 2 C(t ; p ) and b 2 R(t ; p ) estimate varf b g accurately with no systematic dependence on b. We conclude that any shortcomings of the constrained variance and Rothman tests are due to inadequacies of the normal approximation to the distribution of b. Anderson, Bernstein and Pike (1982) have suggested applying logit, arcsin or log(, log) transformations to b to produce a variable which is more nearly normally distributed. We found the third of these transformations, in which log[, logf b g] is treated as approximately normal with estimated variance b 2 G(t )=[ b logf b g] 2, to give the best results. However, Figure 2 shows this transformed test to be inferior to the constrained variance test. In addition, problems arise when b = 0 or 1 since log[, logf b g] and its estimated variance are undened. Replacing b G(t 2 ) by b P 2 (t ) or b C(t 2 ; p ) in the variance estimate led to markedly worse results. In our simulations we found the distribution of S(t b ) to resemble a beta distribution, prompting us to use a beta approximation directly. A Beta(r;s), random variable has 5

expectation r=(r + s) and variance rs=f(r + s) 2 (r + s + 1)g. Equating these to the hypothesised probability p and constrained variance estimate b 2 C(t ; p ) gives the pair of equations r=(r + s) = p and rs=f(r + s) 2 (r + s + 1)g = b 2 C(t ; p ). We use the solutions, r and s, of these equations as the parameters of our beta approximation to the distribution of b and reject H 0 :=p if b 62 f(r;s;=2);(r;s;1,=2)g, where prfbeta(r;s) <(r;s;p) = pg. Abramowitz and Stegun (1972, p.945) provide the accurate approximation fr;s;p) r=(r + se 2w ) where w = z p p h + 1 1, h 2s, 1, + 5 2 2r, 1 6, ; 3h h =2f(2r,1),1 +(2s,1),1 g,1 and =(zp 2,3)=6. The beta approximation is asymptotically equivalent to the normal approximation; see Bickel and Doksum (1977, p. 53). Its small sample accuracy is demonstrated in Figure 2. The cause of the one serious failing when = 0:1 lies in the lower tail of the distribution of which exhibits several spikes, corresponding to zero, one, and two uncensored survivors at time t, and it is not surprising that the beta distribution should approximate this tail poorly. 2.3 Bootstrap condence intervals and a new bootstrap test The bootstrap oers an alternative approach to approximating the distribution of b S(t ). Endpoints of Efron's (1981) \percentile" condence intervals are simply percentiles of the bootstrap sampling distribution of the quantity of interest. In testing H 0 : = p, bootstrap data sets are obtained by sampling with replacement from the observed pairs of survival times and censoring indicators, f(t i ; i ); i = 1;:::;ng. Kaplan-Meier estimates bs B (t ) of the survival function at time t for M bootstrap samples are written in order as bs B [1] (t ) ::: b S B [M ] (t ) and H 0 is rejected if p < b S B [M=2] (t ) or p > b S B [M (1,=2)] (t ): If b S(t ), is approximately pivotal, its distribution can be approximated by that of the bootstrap quantity b S B (t ), b, leading to a dierent form of bootstrap condence interval; see Hall (1992, Example 1.2) and Efron and Tibshirani (1993, Section 6

13.4). The related \pivotal" bootstrap test rejects H 0 : = p if p < 2 b, bs B [M (1,=2)] (t ) or p > 2 b, b S B [M=2] (t ): The rst two plots in Figure 3 show empirical type I error probabilities for percentile and pivotal bootstrap tests, using bootstrap samples of 1,000 replicates. The results are poor and qualitatively similar to those of the Greenwood test. Percentile-t intervals (Hall, 1992, Section 3.2) performed poorly and bias corrected and accelerated percentile intervals (Efron, 1987) were inferior to the \constrained bootstrap" intervals which we now present. The most successful parametric approximations to the distribution of b S(t ) under H 0 : = p make use of the information in H 0. We use this information in our new bootstrap method by sampling from a distribution satisfying =p to provide a reference distribution for b under H 0. We thus avoid the assumption that b,ispivotal or can be transformed to a normality, or the need to correct inaccuracies in such assumptions. In our constrained bootstrap we generate n survival times, fi B, from S(t; t ;p ) as dened by (1) with potential censoring times, c B i, from C, b the Kaplan-Meier estimate of the censoring time distribution, failure being treated as right-censoring for this purpose. We calculate the Kaplan-Meier estimate, b S B (t; p ), up to time t for each bootstrap data-set, order the M bootstrap estimates of as b S B [1] (t ; p ) ::: b S B [M ] (t ; p ), and reject H 0 :=p if b S(t ) < b S B [M=2] (t ; p ) or b > b S B [M (1,=2)] (t ; p ): In fact, we right censor each f B i just beyond t since S(t; t ;p ) is only dened for t t but this suces for determining bs B (t ; p ). If no failures are observed before t, S(t; t ;p ) is undened and, as for the constrained variance test, we reject H 0 if (p ) (n,n c) < =2, where n c is the number of observations censored before t. If all observations in a bootstrap data set are less than t and the largest one is censored, S bb (t ; p ) is undened and we take it to be zero. Results for the constrained bootstrap test, shown in Figure 3, are clearly the best so far. The standard justication of Monte Carlo tests (Barnard, 1963) implies that the number of 7

bootstrap samples, 1000 in our case, has little eect on the empirical rejection rate if the bootstrap provides a good approximation to the null distribution of the test statistic. Despite the well-known connection between hypothesis tests and condence intervals, little common ground is apparent between work on bootstrap condence intervals and bootstrap hypothesis tests. Precedents in which simulations are conducted under a \closest element" of the null hypothesis are Silverman's (1981, 1983) test for the number of modes of a probability density and the bootstrap sampling methods proposed by Zhang (1996, 1997). In Owen's (1988) empirical likelihood method, the bootstrap sampling distribution satises anull hypothesis and Jennison's (1992) bootstrap test for a hazard ratio,, uses bootstrap values of the score statistic generated under the hypothesis = 0. Coupling bootstrap samples for dierent tests can help in computing a constrained bootstrap condence interval. Consider the general setting where bootstrap values T B () of a test statistic are generated under the parameter value. Suppose the value t obs has been observed and let g() = prft B () t obs g: If the distribution of T, and so of T B (), is stochastically increasing in, a 100(1, )% bootstrap condence interval for is ( l ; u ) where g( l ) = =2 and g( u )=1, =2. Finding l and u is complicated by the fact that simulation based estimates of g() may not be monotonic but this problem disappears if bootstrap realisations can be coupled so that the estimate of g() is monotonic in. When computing a constrained bootstrap condence interval for, such coupling is achieved by using a common sequence of uniform random variates for eachhypothesised p in determining survival past the possible failure times t (1) ;:::;t (r). 2.4 Thomas and Grunkemeier's Likelihood Ratio test Thomas and Grunkemeier (1975) derived the constrained estimate of S(t), S(t; t ;p ), for use in constructing a likelihood ratio condence interval for. The equivalent test of H 0 :=p rejects H 0 at the 100% level if,2r > 2 1; where R is the log-likelihood ratio 8

of the observed data under (a) the constrained estimate and (b) the Kaplan-Meier estimate of S(t). Thomas and Grunkemeier show that R = X i:t (i) t ( (n i, d i ) log 1+ n i, d i!, n i log 1+ n i!) and outline a proof that,2r has asymptotically a 2 1 distribution. Li (1995) gives a rigorous proof of this result, and Murphy (1995) further justies the use of this and another extension of the likelihood ratio test to obtain condence intervals for the cumulative hazard function. The results in Figure 3 show Thomas and Grunkemeier's test to perform well over the range of. It attains the nominal error probabilities in each tail more accurately than the beta test and only a little less accurately than the constrained bootstrap test. Some asymmetry is apparent, rejection of H 0 tending to occur in favour of higher survival when is high and in favour of lower survival when islow. 3. Further simulations and conclusions We have applied the constrained variance, beta, constrained bootstrap and likelihood ratio tests to more demanding examples with higher censoring and smaller sample sizes. Figure 4 shows empirical error rates from 5,000 simulations using Exp(10) survival times and Exp(25) censoring times. The sample size is 100 for results on the left and 50 for those on the right. As sample size is decreased or censoring increased, it becomes increasingly likely that max i ft i g < t and the Kaplan-Meier estimate of is undened. The results presented are conditional error rates given that S(t b ) is dened but we have omitted completely cases for which S(t b )was undened in 10 or more of the 5,000 data sets. All four tests continue to do well at values of between about 0.7 and 0.4. The performances of the constrained variance and beta test suer most at high and low values of, while there is not a great deal to choose between the constrained bootstrap and Thomas & Grunkemeier's likelihood ratio test which both fare better. We have seen similar results in further simulations not reported in detail here. 9

Discreteness is apparent when is high and there is signicant probability of zero, one or two failures by time t. The constrained bootstrap only rejects H 0 if S(t b ) is strictly greater than or strictly less than a proportion =2 of the bootstrap replicates. Consequently, this test errs on the side of conservatism where the continuous approximations used by the other three tests can lead to either liberal or conservative error rates. We recommend the constrained bootstrap test for its accuracy and its conservative treatment of the discreteness problem. However, this method requires the greatest amountof computation and, if suciently many uncensored failure times are available, the constrained variance, beta or likelihood ratio tests may bepreferred for their computational simplicity. Acknowledgement S. Barber was supported by an EPSRC studentship. The authors thank Dr Andrew Wood for discussions concerning bootstrap methods and referees for their useful comments. References Abramowitz, M. & Stegun, I.A. (1972). Handbook of mathematical functions with formulas, graphs and mathematical tables. New York: Dover Publications. Anderson, J.R., Bernstein, L. & Pike, M.C. (1982). Approximate condence intervals for probabilities of survival and quantiles in life-table analysis. Biometrics 38, 407{416. Barber, S. & Jennison, C. (1998). A Review of Inferential Methods for the Kaplan-Meier Estimator. Statistics Group Research Report 98:02, University of Bath, UK. Barnard, G. A. (1963). Discussion of paper by M. S. Bartlett. Journal of the Royal Statistical Society, B 25, 294. Bickel, P.J. & Doksum, K.A. (1977) Mathematical statistics: basic ideas and selected topics. San Francisco: Holden{Day. Efron, B. (1981). Censored data and the bootstrap. Journal of the American Statistical Association 76, 312{319. 10

Efron, B. (1987). Better bootstrap condence intervals. (with discussion). Journal of the American Statistical Association 82, 171{200. Efron, B. & Tibshirani R. J. (1993). An Introduction to the Bootstrap. New York: Chapman and Hall. Greenwood, M. (1926). The errors of sampling of the survivorship tables. Reports on Public Health and Statistical Subjects. 33, 1{26. Hall, P. (1992). The Bootstrap and Edgeworth Expansion. New York, Springer-Verlag. Jennison, C. (1992). Bootstrap tests and condence intervals for a hazard ratio when the number of observed failures is small, with applications to group sequential survival studies. In Computing Science and Statistics, Volume 22, (Ed. C. Page and R. LePage), pp. 89{97. New York: Springer-Verlag. Jennison, C. & Turnbull, B.W. (1985). Repeated condence intervals for the median survival time. Biometrika 72, 619{625. Kaplan, E.L. & Meier, P. (1958). Non-parametric estimation from incomplete observations. Journal American Statistical Association 53, 457{481. Li, G. (1995). On nonparametric likelihood ratio estimation of survival probabilities for censored data. Statistics & Probability Letters 25, 95{104. Meier, P. (1975). Estimation of a distribution function from incomplete observations. In Perspectives in Probability and Statistics (Ed J.Gani), pp. 67{87. London: Academic Press. Murphy, S. A. (1995). Likelihood Ratio-Based Condence Intervals in Survival Analysis. Journal of the American Statistical Association 90, 1399{1405. Owen, A. B. (1988). Empirical likelihood ratio condence intervals for a single functional. Biometrika, 75, 237{49. Peto, R., Pike, M.C., Armitage, P., Breslow, N.E., Cox, D.R., Howard, S.V., Mantel, N., 11

McPherson, K., Peto, J. & Smith, P.G. (1977). Design and analysis of randomised clinical trials requiring prolonged observation of each patient, II. British Journal of Cancer 35, 1{39. Rothman, K.J. (1978). Estimation of condence levels for the cumulative probability of survival in life-table analysis. Journal of Chronic Diseases 31, 557{560. Silverman, B. W. (1981). Using kernel density estimates to investigate multimodality. Journal of the Royal Statistical Society B 43, 97{99. Silverman, B. W. (1983). Some properties of a test for multimodality based on kernel density estimates. In Probability, Statistics and Analysis (Ed. J. F. C. Kingman and G. E. H. Reuter), pp. 248{260, Cambridge University Press. Slud, E.V., Byar, D.P. & Green, S.B. (1984). A comparison of reected versus test-based condence intervals for the median survival time, based on censored data. Biometrics 40, 587{600. Strawderman, R.L., Parzen, M.I. & Wells, M.T. (1997). Accurate condence intervals for quantiles under random censoring. Biometrics 53, 1399{1415. Thomas, D.R. & Grunkemeier, G.L. (1975). Condence interval estimation of survival probabilities for censored data. Journal of the American Statistical Association 70, 865{871. Zhang, B. (1996). Estimating a population variance with a known mean. International Statistics review 64, 215{229. Zhang, B. (1997). Quantile processes in the presence of auxiliary information. Annals of the institute of Statistical Mathematics 49, 35{55. 12

Greenwood test Constrained variance test 0.0 0.02 0.04 0.06 0.05 Figure 1: Empirical error rates from 5,000 simulations of Greenwood and constrained variance tests with n = 100, survival times Exp(10) and censoring times Exp(50) Log-log transform test Beta approximation 0.05 0.05 Figure 2: Empirical error rates from 5,000 simulations of tests based on the log(-log) transform and the beta approximation of b. Here n = 100, survival times Exp(10) and censoring times Exp(50) 13

Percentile bootstrap test Pivotal bootstrap test 0.0 0.02 0.04 0.06 0.08 0.10 0.0 0.02 0.04 0.06 0.08 0.10 Constrained bootstrap test Likelihood ratio test 0.05 0.05 Figure 3: Empirical error rates from 5,000 simulations of the \percentile", \pivotal" and \constrained" bootstrap tests and Thomas and Grunkemeier's likelihood ratio test. Here n = 100, survival times Exp(10) and censoring times Exp(50) 14

Constrained variance tests Beta tests Constrained bootstrap tests Likelihood ratio tests Figure 4: from 5,000 simulations of four tests with survival times Exp(10) and censoring times Exp(25). Results on the left are for n = 100 and those on the right for n =50. Cases where b was undened in more than 10 trials out of 5,000 have been omitted. 15