A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints and Its Application to Empirical Likelihood

Similar documents
A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints

Empirical likelihood ratio with arbitrarily censored/truncated data by EM algorithm

COMPUTATION OF THE EMPIRICAL LIKELIHOOD RATIO FROM CENSORED DATA. Kun Chen and Mai Zhou 1 Bayer Pharmaceuticals and University of Kentucky

COMPUTE CENSORED EMPIRICAL LIKELIHOOD RATIO BY SEQUENTIAL QUADRATIC PROGRAMMING Kun Chen and Mai Zhou University of Kentucky

Empirical Likelihood in Survival Analysis

Quantile Regression for Residual Life and Empirical Likelihood

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky

11 Survival Analysis and Empirical Likelihood

and Comparison with NPMLE

EMPIRICAL ENVELOPE MLE AND LR TESTS. Mai Zhou University of Kentucky

AFT Models and Empirical Likelihood

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

Full likelihood inferences in the Cox model: an empirical likelihood approach

FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH

University of California, Berkeley

Package emplik2. August 18, 2018

Likelihood Construction, Inference for Parametric Survival Distributions

Empirical Likelihood

TMA 4275 Lifetime Analysis June 2004 Solution

Jackknife Empirical Likelihood for the Variance in the Linear Regression Model

Nonparametric rank based estimation of bivariate densities given censored data conditional on marginal probabilities

Multistate Modeling and Applications

Nonparametric Bayes Estimator of Survival Function for Right-Censoring and Left-Truncation Data

Multivariate Survival Analysis

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

Empirical Likelihood Inference for Two-Sample Problems

A note on profile likelihood for exponential tilt mixture models

Goodness-of-fit tests for randomly censored Weibull distributions with estimated parameters

Semiparametric Regression

ST495: Survival Analysis: Hypothesis testing and confidence intervals

1 Glivenko-Cantelli type theorems

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY

Pooling multiple imputations when the sample happens to be the population.

Estimation for nonparametric mixture models

Maximum Smoothed Likelihood for Multivariate Nonparametric Mixtures

Logistic regression model for survival time analysis using time-varying coefficients

Analytical Bootstrap Methods for Censored Data

Other Survival Models. (1) Non-PH models. We briefly discussed the non-proportional hazards (non-ph) model

Estimation of Conditional Kendall s Tau for Bivariate Interval Censored Data

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

STAT 6350 Analysis of Lifetime Data. Probability Plotting

Let us use the term failure time to indicate the time of the event of interest in either a survival analysis or reliability analysis.

Eco517 Fall 2004 C. Sims MIDTERM EXAM

LSS: An S-Plus/R program for the accelerated failure time model to right censored data based on least-squares principle

Analysis of Gamma and Weibull Lifetime Data under a General Censoring Scheme and in the presence of Covariates

18.465, further revised November 27, 2012 Survival analysis and the Kaplan Meier estimator

Proportional hazards regression

Modified Simes Critical Values Under Positive Dependence

MAS3301 / MAS8311 Biostatistics Part II: Survival

Chapter 4: Factor Analysis

Constrained estimation for binary and survival data

Nonparametric Hypothesis Testing and Condence Intervals with. Department of Statistics. University ofkentucky SUMMARY

exclusive prepublication prepublication discount 25 FREE reprints Order now!

Stochastic Analogues to Deterministic Optimizers

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon

Biostat 2065 Analysis of Incomplete Data

Estimating Bivariate Survival Function by Volterra Estimator Using Dynamic Programming Techniques

A general mixed model approach for spatio-temporal regression data

In contrast, parametric techniques (fitting exponential or Weibull, for example) are more focussed, can handle general covariates, but require

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Hacettepe Journal of Mathematics and Statistics Volume 45 (5) (2016), Abstract

Group Sequential Designs: Theory, Computation and Optimisation

Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring

Product-limit estimators of the survival function with left or right censored data

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations

36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

Survival Analysis for Case-Cohort Studies

Bootstrap Procedures for Testing Homogeneity Hypotheses

TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOZIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississip

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Exam C Solutions Spring 2005

Lecture 3. Truncation, length-bias and prevalence sampling

Approximation of Average Run Length of Moving Sum Algorithms Using Multivariate Probabilities

Chapter 2 Inference on Mean Residual Life-Overview

Censoring mechanisms

Optimal Treatment Regimes for Survival Endpoints from a Classification Perspective. Anastasios (Butch) Tsiatis and Xiaofei Bai

Likelihood ratio confidence bands in nonparametric regression with censored data

Stat 5101 Lecture Notes

Approximation of Survival Function by Taylor Series for General Partly Interval Censored Data

Lecture 22 Survival Analysis: An Introduction

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Survival Regression Models

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

ICSA Applied Statistics Symposium 1. Balanced adjusted empirical likelihood

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Testing Equality of Two Intercepts for the Parallel Regression Model with Non-sample Prior Information

NONPARAMETRIC BAYES ESTIMATOR OF SURVIVAL FUNCTIONS FOR DOUBLY/INTERVAL CENSORED DATA

14.30 Introduction to Statistical Methods in Economics Spring 2009

Masters Comprehensive Examination Department of Statistics, University of Florida

Master s Written Examination - Solution

A Model for Correlated Paired Comparison Data

Some Approximations of the Logistic Distribution with Application to the Covariance Matrix of Logistic Regression

Transcription:

Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints and Its Application to Empirical Likelihood Mai Zhou Yifan Yang Received: date / Accepted: date Abstract The Kaplan-Meier estimator is very popular in analysis of survival data. However, it is not easy to compute the constrained Kaplan-Meier. Current computational method uses expectationmaximization algorithm to achieve this, but can be slow at many situations. In this note we give a recursive computational algorithm for the constrained Kaplan-Meier estimator. The constraint is assumed given in linear estimating equations or mean functions. We also illustrate how this leads to the empirical likelihood ratio test with right censored data. Speed comparison to the EM based algorithm favours the current procedure. Keywords Wilks test Empirical Likelihood Ratio Right Censored Data NPMLE 1 Introduction Suppose that X 1,X 2,...,X n are i.i.d. non-negative random variables denoting the lifetimes with a continuousdistributionfunctionf 0.IndependentofthelifetimestherearecensoringtimesC 1,C 2,...,C n which are i.i.d. with a distribution G 0. Only the censored observations, (T i,δ i ), are available to us: T i = min(x i,c i ) and δ i = I[X i C i ] for i = 1,2,...,n. (1) Here I[A] is the indicator function of of A. The empirical likelihood (EL) of the censored data in terms of distribution F is defined as EL(F) = = n [ F(T i ))] δi [1 F(T i )] 1 δi n [ F(T i ))] δi { Mai Zhou Department of Statistics,University of Kentucky Tel.: (859) 257-6912 Fax: (859) 323-1973 E-mail: mai@ms.uky.edu Yifan Yang Department of Statistics,University of Kentucky Tel.: (859) 494-4072 E-mail: yifan.yang@uky.edu j:t j>t i F(T j )} 1 δi

2 Mai Zhou, Yifan Yang where F(t) = F(t+) F(t ) is the jump of F at t. See for example Kaplan and Meier (1958) and Owen (2001). The second line above assumes a discrete F( ). It is well known that the constrained or unconstrained maximum of the empirical likelihood are both obtained by discrete F (Zhou, 2005). Let w i = F(T i ) for i = 1,2,...,n. The likelihood at this F can be written in term of the jumps and the log likelihood is logel = EL = n n n [w i ] δi { w j I[T j > T i ]} 1 δi, δ ilogw i +(1 δ i )log n w j I[T j > T i ]. (2) If we maximize the log EL above without extra constraint (the probability constraints w i 0, and w i = 1 are always imposed), it is well known (Kaplan and Meier, 1958) that the Kaplan-Meier estimator w i = ˆF KM (T i ) will achieve the maximum value of the log EL. Empirical likelihood ratio method was first proposed by Thomas and Grunkemeier (1975) in the context of a Kaplan-Meier estimator. This method has been studied by Owen (1988, 2001); Li (1995); Murphy and van der Vaart (1997) and Pan and Zhou (1999) among many others. When using the empirical likelihood with right censored data in testing a general hypothesis, Zhou (2005) gave an EM algorithm to compute the likelihood ratio. He compared the EM algorithm to the sequential quadratic programming method (Chen and Zhou, 2007), and concluded that the EM was better. Though quite stable, the EM can be slow in certain data settings. See examples in section 3. We shall give a new recursive computation procedure for the constrained maximum of the log empirical likelihood above, which leads to the computation of the empirical likelihood ratio for testing. Section two contains the derivation of the recursive algorithm, as well as its application to the censored data empirical likelihood test. Section three reports the simulation results and a real data example. Finally we end the paper with a discussion of further issues. 2 The mean constrained Kaplan-Meier estimator In order to compute the empirical likelihood ratio, we need two empirical likelihoods: one with constraints, one without. The maximum of the empirical likelihood without constraint is achieved by an F equals to the Kaplan-Meier estimator, as is well known. It remains to find the maximum of logel under constraints, or equivalently, the Kaplan-Meier under constraints. Using an argument similar to in Owen (1988) will show that we may restrict our attention in the EL analysis, i.e. search max under constrains, to those discrete CDF F that are dominated by the Kaplan-Meier: F(t) ˆF KM (t). Owen (1988) restricted his attention to those distribution functions that are dominated by the empirical distribution. The first step in our analysis is to find a discrete CDF that maximizes the log EL(F) under the mean constraints (3), which are specified as follows: 0 0 0 g 1 (t)df(t) = µ 1 g 2 (t)df(t) = µ 2 (3) g p (t)df(t) = µ p

Title Suppressed Due to Excessive Length 3 0.0 0.2 0.4 0.6 0.8 1.0 Kaplan Meier Curve H 0: 0 τxdf(x)=2.2 H 0:S(3)=1 0.4 0 1 2 3 4 5 6 7 Fig. 1 Kaplan-Meier plot with and without mean constraint for Stanford 5 data discussed in Section 3, time scaled by 500. where g i (t)(i = 1,2,...,p) are given functions satisfy some moment conditions (specified later), and µ i (i = 1,2,...,p) are given constants. Without loss of generality, we shall assume all µ i = 0. Examples of such constrained Kaplan-Meier are shown in Figure 1. The constraints (3) can be written as (for discrete CDFs, and in terms of w i = F(T i ) = F(T i ) F(T i )) n g 1 (T i )w i = 0 (4) n g p (T i )w i = 0. We must find the maximum of the logel(f) under these constraints. We shall use the Lagrange multiplier to find this constrained maximum. Since F(t) ˆF KM (t), w i is only positive at the uncensored observation T i, except may be the last observation. Without loss of generality assume the observations are already ordered according to T and the smallest observation is an uncensored one (δ 1 = 1). To see this, suppose the first observation is right censored and second one is uncensored. In this case, δ 1 = 0, and w 1 = 0. Hence w 1 = 0, n w i = 1, and δ 1 logw 1 0. (5) i=2 The i-th term in log empirical likelihood is δ i logw i +(1 δ i )log n j=i+1 This is true as observations are sorted according to T. Since δ 1 logw 1 0, the log empirical likelihood only depends on w 2,...,w n. Additionally, the first observation with δ = 0 has no contribution to the constraints. w j.

4 Mai Zhou, Yifan Yang Now, assume the sample of n observations of are already ordered according to T and δ 1 = 1. Let I = {i 1,...,i k } be the index set of censored observations among the n s such that T i1 < < T ik, k is number of elements in I. Thus we have only n k positive probability w i s. Introduce k new variables { S 1,..., S k }, one for each censored T observation, i.e. assume j I, i.e. δ ij = 0, and let: = w i = 1 w i. (6) i:t i>t ij i:t i T ij This adds k new constraints to the optimization problem. We write the vector of those k constraints as i:t i>t ij w i = 0. With these k new variables S i, the log empirical likelihood in section 1 can be written simply as logel = n δ i logw i + k log. (7) Remark 1 Obviously, if two censored observations T ij and T im are such that there are no uncensored observations sandwiched in between them, then = S m. This is obvious from (6) and the fact w i > 0 only when δ i = 1. We comment that this do not change our final recursive formula (12) and (6). The Lagrangian function for constrained maximum is ( ) ( n ) G = logel(w, S)+λ i w i g(t i ) η w i 1 δ δ γ ( S ) W. Here, λ R p, g(t i ) = (g 1 (T i ),...,g p (T i )) is a vector corresponding to the constraints (4); η R is a scalar; γ, S,W R k, S is a vector for those s defined in (6), and the j-th entry of W is i:t i>t ij w i. Next we shall take the partial derivatives and set them to zero. We shall show that η = n. First we compute G = 1 δ j γ j Setting the derivative to zero, we have Furthermore, γ j = (1 δ j )/. (8) G w l = δ l w l +δ l λ g(t l ) η +γ U (l), whereu (l) {0,1} k isavectorwith the j-th entry tobe anindicatori[t ij < T l ] (1 δ ij ) = I[T ij < T l ]. Then set the derivative to zero and write l as i: Multiply w i on both sides and sum, w i η = i i η = δ i w i +δ i λ g(t i )+γ U (i). δ i + i Make use the other constraints, this simplifies to δ i w i λ g(t i )+( i w i γ U (i) ). η = (n k)+0+ i w i γ U (i). (9) We now focus on the last term above. Plug in the γ j expression we obtained in (8) above,

Title Suppressed Due to Excessive Length 5 switch order of summation and it is not hard to see that n w iγ U (i) = ( n w k ) i γ ju (i) j = k = k. n w ii[t ij <T i] I[δ ij =0] = k 1I[δ i j = 0] Therefore equation (9) becomes η = (n k)+0+k. Therefore η = n. Thus, we have δ i w i = (11) n λ δ i g(t i ) γ U (i) where we further note (plug in the γ, and δ ij = 0): γ U (i) = This finally gives rise to k (1 δ ij ) I[T ij < T i,δ ij = 0] = k I[T ij < T i ]. (10) w i = w i (λ) = δ i n λ δ i g(t i ) (12) k I[T ij <T i] which, together with (6), provides a recursive computation method for the probabilities w i, provided λ is given: 1. Starting from the left most observation, and without loss of generality (as noted above) we can assume it is an uncensored data point: δ 1 = 1. Thus w 1 = 1 n λ g(t 1 ). 2. Once we have w i for all i l, we also have all Sj where T ij < T l+1 and δ ij = 0, by using ( = 1 i I[T i T ij ]w i ), then we can compute w l+1 = δ l+1 n λ g(t l+1 ). (13) k I[T ij <T l+1 ] So this recursive calculation will give us w i and as a function of λ. Remark 2 In the special case of no constraint of mean, then there is no λ (or λ = 0) and we have w l+1 = δ l+1 n, k I[T ij <T l+1 ] which is actually the jump of the Kaplan-Meier estimator. Proof Use the identity (1 ˆF)(1 Ĝ) = 1 Ĥ, we first get a formula for the jump of the Kaplan-Meier estimator: w i = 1/n 1/(1 Ĝ). This works for ˆF as well as for Ĝ. We next show that 1 1/n k I[T ij < T k+1 ] = (1 Ĝ(T k+1)) since the left hand side is just equal to the summation of jumps of Ĝ before T k+1.

6 Mai Zhou, Yifan Yang Finally the λ value needs to be determined from the constraint equation 0 = i δ i w i (λ)g(t i ) = i More formally, the recursion proceeds as follows: δ i g(t i ) n λ δ i g(t i ). (14) k I[T j<t i] (1) Initialization: Pick a λ value near zero, or identical to zero. (Recall λ = 0 gives the Kaplan-Meier). (2) Updating w and S: With this λ find all the w i s and s by the recursive formula (13) and (6). (3) Updating λ: Plug those w i into the right hand side of equation (14) above and call it θ. The w i s obtained in step (2) are actually the constrained Kaplan-Meier with the constraint being these θ instead of zero. (4) Checking λ and repeat: Check if θ is zero. If not, change the λ value and repeat, until you find a λ which gives rise to w i and that satisfy θ = 0. For one dimensional λ this is going to be easily handled by any function that computes the root of an univariate function such as the uniroot function in R. For multi-dimensional λ this calls for a Newton type iteration. The empirical likelihood ratio is then obtained as 2logELR = 2{logEL(w i,s j ) logel(w i = ˆF KM (T i ))} ; for the first log EL inside the curly bracket above we use the expression (7) with the w i,s j computed fromtherecursivemethodinthissection,andthesecondlogelisobtainedby(2)withw i = ˆF KM (T i ), the regular Kaplan-Meier estimator. Observations (T i,δ i ) are first ordered according to their T i values. If there are several observations with identical T i value, we then order according to their δ i value. That is, an uncensored T i is considered as come before a censored T j even when T i = T j. This is the usual convention in the calculation of the Kaplan-Meier estimators. When there are observations with identical T i and δ i value, we can either merge the tied observations and record the number of tie in another vector u i ; or we may just leave the tied observations as is, in the order of their input. When there are substantial(extensive) tied observations, it may save computational time to first merge the tied data record the number of tied in u i before the recursive computation. However, in most applications of survival analysis the Kapan-Meier estimator will be computed along with some covariates, as in regression analysis. So the actual data likely will look like (T i,δ i,x i ) where x i are the covariates, e.g. treatment, gender, age, blood pressure, etc. of the i-th patient. In this case, even if two observations have identical T i and δ i, they should not be merged because they have different covariates. Therefore in the current implementation of kmc package, we choose not to merge any tied data. Under the assumption that the variance of g(t)dˆf KM (t) is finite, we have a chi square limiting distribution for the above -2 log empirical likelihood ratio, under null hypothesis as stated in the Wilks Theorem. Therefore we reject the null hypothesis if the computed -2 empirical likelihood ratio exceeds the chi square 95% percentile with p degrees of freedom. See Zhou (2010) for a proof of this theorem. 3 Simulation If the Jacobian matrix of mean zero requirement (14) is not singular, Newton method could be used to find the root of (14). As mentioned previously, the null parameter space which has no constraint corresponds to λ = 0. We shall call the recursive computation for w i s plus the Newton iteration for λ as Kaplan-Meierconstrained(KMC) method, which is also the name of the R package kmc available on the comprehensive R archive network (CRAN).

Title Suppressed Due to Excessive Length 7 In each Newton iterationwhen we try to find the root,it is helpful to know the bounds of solution, or socalled, feasible region.in the KMC computation, it is obviousthat those λ s suchthat n δ i λ g(t i ) k I[T ij <T i] = 0 will lead (14) to. Denote those roots as λ i Rp,i = 1,...,n. Any i-th entry of the desired λ root for (14) must satisfy j such that λ i ji < 0,λ i < λ i j+1i i = 1,...,p where λ ij is the j-th entry of vector λ i such that λ i <... < 1i λ i, and ni λ 0,i =, λ n+1,i = +. So, our strategy is to start at 0 and try to stay within the feasible region at all times when carry out the Newton iterations. We could also calculate the analytical derivatives used in the Newton iteration. Denote the right hand side of (14) as f(λ), i.e. f(λ) = δ i w i (λ)g(t i ). i To compute λf(λ), we only need to calculate λ w i and S ( λ j = λ 1 ) j k=1 w k = j λ k=1 w k. There are no closed forms of such derivatives, but it could again be derived recursively. (1) w 1 = 1 n λg(t 1), and λ w 1 = w 2 1 g(t 1) (2) λ w k+1(λ) = δ k+1 (w k+1 ) 2 ( g(t k+1 )+ n ( I[T ij < T k+1 ]( ) 2 ) ij λ s=1 λ w s To evaluate the performance of this algorithm, a series of simulations had been done. We compared with standard EM algorithm (Zhou, 2005). Without further statement, all simulation works have been repeat 5,000times and implemented in R language(r Core Team, 2014). R-2.15.3is used on awindows 7 (64-bits) computer with 2.4 GHz Intel(R) i7-3630qm CPU. Experiment 1 : Consider a right censored data with only one constraint: { X Exp(1) (15) C Exp(β) Censoring percentage of the data are determined by different β s. Three models are included in the experiments: (1) β = 1.5, then 40% data are uncensored; (2) β = 0.7, then 58.9% data are uncensored; (3) β = 0.2, then 83.3% data are uncensored. The common hypothesis is (4), where g(x) = (1 x)1 (0 x 1) e 1. We could verify that the true expectation is zero: g(x)df(x) = (1 x)1 (0 x 1) e x dx e 1 =0. To compare the performances of KMC and EM algorithm, we use four different sample sizes, i.e. 200, 1,000, 2,000, and 5,000 in the experiments. To make fair comparisons, f (t) f (t+1) l2 10 9 is used as the convergence criterion for EM algorithm, and f (t) l1 10 9 is used for KMC. Average spending time is reported to compare the computation efficiency in Table 1. The no censored case is included for reference, this is equivalent to Newton solving λ without recursion. In all cases in our study, EM and KMC reported almost the same χ 2 test statistics. As shown in Table 1, we observed the following phenomenons: (1) KMC always outperformed EM algorithm in speed at different simulation settings. (2) Computation complexity of EM increased sharply with the percentage of censored data increasing. This is reasonable, since more censored data needs more E-step computation. But censored rate did not affect KMC much. (3) Sample size is related to the computation complexity. We could see the running time of both EM and KMC increased along with the sample size. ).

8 Mai Zhou, Yifan Yang Table 1 Average running time of EM/KMC (in second). No Censor column refers to time spend on solving empirical likelihood without censoring in R el.test(emplik). We use this as a comparison reference. Censoring Rate N EM KMC(nuDev) KMC(AnalyDev) No Censor 200 0.175 0.011 0.028 0.005 60% 1000 3.503 0.106 0.211 0.007 β = 1.5 2000 13.935 0.349 0.692 0.033 5000 73.562 1.801 3.663 0.036 200 0.064 0.010 0.029 0.000 41% 1000 1.058 0.115 0.268 0.010 β = 0.7 2000 4.104 0.385 0.836 0.020 5000 22.878 2.367 4.693 0.037 200 0.014 0.008 0.029 0.002 17% 1000 0.117 0.071 0.240 0.009 β = 0.2 2000 0.425 0.240 0.694 0.018 5000 2.702 1.220 3.282 0.026 (4) Another phenomenon is that, the computation of numeric derivative and analytic derivative of KMC is similar. This fact is straightforward, as KMC depends on solving (14) which is defined iteratively. To summarize, when sample size is small and censored rate is low, the performance of EM and KMC is similar. But either in the large sample case or heavily censored case, KMC far outperformed EM algorithm with the same stopping criterion. Experiment 2 : Consider a right censored data setting with two constraints. The i.i.d. right censored data are generated from (1) using { X Exp(1) (16) C Exp(.7) with the following hypothesis: H 0 : i g j (T i )w i = 0; j = 1,2 where { g1 (x) = (1 x)1 (0 x 1) e 1 g 2 (x) = 1 (0 x 1) 1+e 1 (17) It is straightforward to verify that both g functions have expectation zero. In this simulation study, we Table 2 Average running time of EM/KMC (in second) Censoring Rate N EM KMC(nuDev) 41% 200 3.055 0.033 41% 500 55.601 0.083 observed that EM spent great amount of time (3s 55s per case) to meet the converge criterion, while the average running time of KMC was considerable shorter (0.03s 0.08s). This dramatic result shows in multi-dimensional case, KMC runs much faster than EM algorithm. Only numerical derivatives were used in our simulations, one could implement the analytic ones using iteration shown previously. But in multi-dimensional case, the iterative type of derivatives do not have advantage over numeric ones. We recommend to use KMC with numeric derivative if one has more than one hypothesis even the sample size is small. Experiment 3 : Non-exponential survival setting. Consider a right censored data with one constraints: { X Γ(3,2) (18) C U(0,η) with hypothesis H 0 : g(t i )w i = 0, with g(x) = x 1.5 i we carried out two experiments on different censoring time C i from uniform distribution:

0.02 Title Suppressed Due to Excessive Length 9 0.1 0.17 Intercept 3.48 3.50 3.52 3.54 3.56 0.02 0.05 0.11 0.04 0.09 0.09 0.15 0.14 0.06 0.03 0.04 0.17 0.12 0.12 + 0.07 0.05 0.16 0.13 0.1 0.14 0.08 0.13 0.07 0.01 0.08 0.14 0.03 0.15 0.17 0.16 0.11 0.06 0.0210 0.0205 0.0200 0.0195 0.0190 0.0185 Fig. 2 Contour plot of-2 log likelihood ratio corresponding to intercept and slope of age for the Stanford Heart Transplants Data. β Age (1) η = 5, then 70.00% data are uncensored; (2) η = 3, then 51.34% data are uncensored. Table 3 Average running time of EM/KMC (in second) of one constraint and Uniform distributed censored time Censoring Rate N EM KMC(nuDev) KMC(AnalyDev) No Censor 30.00% 200 0.124 0.018 0.044 0.005 η = 5 2000 11.237 0.725 1.197 0.112 48.66% 200 0.075 0.019 0.045 0.004 η = 3 2000 4.141 1.068 1.528 0.139 We found that the result shown in Table 3 is very similar to Table 1, which shows that different distribution of censored time will not affect the simulation result too much. Real Data Example : The speed advantage of KMC algorithm is more apparent in time consuming analysis such as drawing contour plot. In this real data example, we illustrate the proposed algorithm to analyze the Stanford Heart Transplants Data described in Miller and Halpern (1982) and considered regression with intercept and slope term of age. There were 157 patients who received transplantation, among which 55 were still alive and 102 were deceased. We deleted cases that the survival times are less than 5 days, and used only 152 cases in the analysis, as suggested by Miller and Halpern. To draw such contour plot 51 51 = 2601 empirical likelihood ratios were calculated. In this example, we used KMC to calculate the empirical likelihood instead of EM described in Zhou and Li (2008). 4 Discussion and Conclusion In this manuscript, we proposed a new recursive algorithm, KMC, to calculate mean constrained Kaplan-Meier estimator and log empirical likelihood ratio statistics of right censored data with linear constraints. Our algorithm used Lagrange multiplier method directly, and recursively computes the constrained Kaplan-Meier and test statistics.

10 Mai Zhou, Yifan Yang Numerical simulations show this new method has an advantage against traditional EM algorithm in the sense of computational complexity. Our simulation work also shows that the performance of KMC does not depend on the censoring rate, and outperformed EM algorithm at every simulation setting. We recommend to use KMC in all cases but particular large gain are expected in the following cases: (1) Sample size is large (e.g. > 1000 observations); (2) Data are heavily censored (e.g. censored rate > 40%); (3) There are more than one constraints. On the other hand, and somewhat surprisingly, the analytic derivative did not help speed up computation in our simulation study. Besides, since KMC with numeric derivative could be extended to more than one constraints case, we highly recommend using numeric derivative in KMC rather than analytical one. One of the issues of KMC is the initial value choosing, as is the case for most Newton algorithms. The performance of root solving relies on the precision of numerical derivative and Newton method. Our current strategy uses the M-step output of EM algorithm with only two iterations. Other better initial values are certainly possible. In addition, current KMC only works on right censored data, while EM algorithm works for right-, left- or doubly censored data; or even interval censored data. We were unable to find a proper way to derive such recursive computation algorithm in other censoring cases. Software is available to download at http://cran.r-project.org/web/packages/kmc as a standard R package. Acknowledgements Mai Zhou s research was supported in part by US NSF grant DMS 1007666. References Chen K, Zhou M (2007) Computation of the empirical likelihood ratio from censored data. Journal of Statistical Computation and Simulation 77(12):1033 1042 Kaplan EL, Meier P (1958) Nonparametric estimation from incomplete observations. Journal of the American statistical association 53(282):457 481 Li G (1995) On nonparametric likelihood ratio estimation of survival probabilities for censored data. Statistics & Probability Letters 25(2):95 104 Miller R, Halpern J (1982) Regression with censored data. Biometrika 69(3):521 531 Murphy SA, van der Vaart AW (1997) Semiparametric likelihood ratio inference. The Annals of Statistics pp 1471 1509 Owen AB (1988) Empirical likelihood ratio confidence intervals for a single functional. Biometrika 75(2):237 249 Owen AB (2001) Empirical likelihood. CRC press Pan XR, Zhou M (1999) Using one-parameter sub-family of distributions in empirical likelihood ratio with censored data. Journal of statistical planning and inference 75(2):379 392 R Core Team (2014) R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria, URL http://www.r-project.org/ Thomas DR, Grunkemeier GL (1975) Confidence interval estimation of survival probabilities for censored data. Journal of the American Statistical Association 70(352):865 871 Zhou M (2005) Empirical likelihood ratio with arbitrarily censored/truncated data by em algorithm. Journal of Computational and Graphical Statistics 14(3):643 656 Zhou M (2010) A wilks theorem for the censored empirical likelihood of means. URL www.ms.uky.edu/~mai/research/note3.pdf Zhou M, Li G (2008) Empirical likelihood analysis of the buckley james estimator. Journal of multivariate analysis 99(4):649 664