Comparison of Predictive Accuracy of Neural Network Methods and Cox Regression for Censored Survival Data

Similar documents
Prerequisite: STATS 7 or STATS 8 or AP90 or (STATS 120A and STATS 120B and STATS 120C). AP90 with a minimum score of 3

Reading Group on Deep Learning Session 1

Lecture 5: Logistic Regression. Neural Networks

Vasil Khalidov & Miles Hansard. C.M. Bishop s PRML: Chapter 5; Neural Networks

Statistical NLP for the Web

Data Mining Part 5. Prediction

Computational statistics

Neural Networks and the Back-propagation Algorithm

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

BIAS OF MAXIMUM-LIKELIHOOD ESTIMATES IN LOGISTIC AND COX REGRESSION MODELS: A COMPARATIVE SIMULATION STUDY

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Unit III. A Survey of Neural Network Model

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Artificial Intelligence

CSC 578 Neural Networks and Deep Learning

Statistical aspects of prediction models with high-dimensional data

An artificial neural networks (ANNs) model is a functional abstraction of the

Artificial Neural Networks (ANN) Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE GRAVIS 2016 BASEL. Logistic Regression. Pattern Recognition 2016 Sandro Schönborn University of Basel

AI Programming CS F-20 Neural Networks

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

Integer weight training by differential evolution algorithms

A comparison of inverse transform and composition methods of data simulation from the Lindley distribution

Neural Networks and Deep Learning

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Confidence Estimation Methods for Neural Networks: A Practical Comparison

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Artificial Neural Networks. MGS Lecture 2

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky

18.6 Regression and Classification with Linear Models

R-squared for Bayesian regression models

Artificial Neural Networks

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Semiparametric Regression

Data splitting. INSERM Workshop: Evaluation of predictive models: goodness-of-fit and predictive power #+TITLE:

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

4. Multilayer Perceptrons

Neural Networks. Bishop PRML Ch. 5. Alireza Ghane. Feed-forward Networks Network Training Error Backpropagation Applications

COMPUTATIONAL INTELLIGENCE (INTRODUCTION TO MACHINE LEARNING) SS16

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

Neural Network Approach to Estimating Conditional Quantile Polynomial Distributed Lag (QPDL) Model with an Application to Rubber Price Returns

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Lecture 10. Neural networks and optimization. Machine Learning and Data Mining November Nando de Freitas UBC. Nonlinear Supervised Learning

Survival SVM: a Practical Scalable Algorithm

Neural networks (not in book)

Feed-forward Networks Network Training Error Backpropagation Applications. Neural Networks. Oliver Schulte - CMPT 726. Bishop PRML Ch.

Package comparec. R topics documented: February 19, Type Package

Pattern Recognition and Machine Learning

Introduction to Machine Learning

Artificial Neural Network

C4 Phenomenological Modeling - Regression & Neural Networks : Computational Modeling and Simulation Instructor: Linwei Wang

Extensions of Cox Model for Non-Proportional Hazards Purpose

Logistic Regression & Neural Networks

Survival Regression Models

10-701/ Machine Learning, Fall

Lecture 4 Discriminant Analysis, k-nearest Neighbors

Artificial Neural Networks Examination, June 2005

STAT331. Cox s Proportional Hazards Model

From perceptrons to word embeddings. Simon Šuster University of Groningen

The exam is closed book, closed notes except your one-page cheat sheet.

Introduction to Statistical Analysis

Different Criteria for Active Learning in Neural Networks: A Comparative Study

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

epochs epochs

Extensions of Cox Model for Non-Proportional Hazards Purpose

Artificial Neural Networks

LOGISTIC REGRESSION Joseph M. Hilbe

Survival Prediction Under Dependent Censoring: A Copula-based Approach

Multi-state Models: An Overview

Forecasting of Rain Fall in Mirzapur District, Uttar Pradesh, India Using Feed-Forward Artificial Neural Network

Math for Machine Learning Open Doors to Data Science and Artificial Intelligence. Richard Han

ECLT 5810 Linear Regression and Logistic Regression for Classification. Prof. Wai Lam

Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters

Artificial Neural Networks

A Note on Bayesian Inference After Multiple Imputation

More Accurately Analyze Complex Relationships

Multilayer Perceptron = FeedForward Neural Network

Statistics in medicine

Machine Learning Linear Classification. Prof. Matteo Matteucci

Stable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems

Stat 642, Lecture notes for 04/12/05 96

STA 4273H: Statistical Machine Learning

Robustifying Trial-Derived Treatment Rules to a Target Population

Neural Network Based Response Surface Methods a Comparative Study

On a connection between the Bradley-Terry model and the Cox proportional hazards model

Course in Data Science

Classification. Chapter Introduction. 6.2 The Bayes classifier

Part III Measures of Classification Accuracy for the Prediction of Survival Times

Small n, σ known or unknown, underlying nongaussian

49th European Organization for Quality Congress. Topic: Quality Improvement. Service Reliability in Electrical Distribution Networks

Learning features by contrasting natural images with noise

Technical Report - 7/87 AN APPLICATION OF COX REGRESSION MODEL TO THE ANALYSIS OF GROUPED PULMONARY TUBERCULOSIS SURVIVAL DATA

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Missing covariate data in matched case-control studies: Do the usual paradigms apply?

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

UNIVERSITY OF CALIFORNIA, SAN DIEGO

Transcription:

Comparison of Predictive Accuracy of Neural Network Methods and Cox Regression for Censored Survival Data Stanley Azen Ph.D. 1, Annie Xiang Ph.D. 1, Pablo Lapuerta, M.D. 1, Alex Ryutov MS 2, Jonathan Buckley Ph.D. 2 1 Statistical Consultation and Research Center, Department of Preventive Medicine University of Southern California School of Medicine, 1540 Alcazar Street, CHP 218, Los Angeles CA 90033; email: sazen@rcf.usc.edu 2 Department of Outcomes Research, Pharmaceutical Research Institute, Bristol-Meyers Squibb, Princeton NJ ABSTRACT Strategies that have been developed to extend NN prediction methods to accommodate right-censored data include methods due to Faraggi-Simon, Liestol- Andersen-Andersen, and a modification of the Buckley-James method. In a Monte Carlo simulation study, we evaluated the performance of all three NN methods with that of Cox regression. Using the EPILOG PLUS PROC NEURAL utility developed by our group, feed-forward back-propagation networks were examined under 8 designs representing a variety of experimental conditions. Minimization methods were implemented that efficiently determined optimal parameters. For the training phase, the Faraggi-Simon and Liestol- Andersen-Andersen methods outperformed Cox regression for at least one of the study designs. For the testing phase of the study, the Farragi-Simon method performed as well as Cox regression for all 8 designs. The Liestol-Andersen-Andersen method performed as well as Cox regression for 6 of the 8 designs. Finally, the Buckley-James method performed as well as Cox regression for 5 of the 8 designs. The results of our study suggest that NNs can serve as effective methods for modeling rightcensored data. However, the performance of the NN is somewhat variable, depending on the method that is used. INTRODUCTION Neural networks (NNs) are sophisticated computer programs, which model the capabilities of the human brain by mimicking the structure and function of neurons in the brain. 1,2 Utilizing principles of artificial intelligence, NNs permit the modeling of complex functional relationships. 3 A NN consists of layers of nodes (analogous to neurons) linked by inter-connections (axons/dendrites), together with rules that specify how the output of each node is determined by input values from all nodes at the level below. A layered architecture of neurons in the brain can be used to provide progressively more abstract representation of input stimuli as the information is filtered through successive layers. Neural networks attempt to reproduce this effect, although most networks are limited in practice to three or four layers in total. Theoretical work suggests that NNs can consistently match or exceed the performance of statistical methods. 1,2 NNs are being used in the areas of prediction and classification of outcomes in medicine -- areas where regression models have traditionally been used. In most applications, outcomes (termed outputs by users of NNs) are characterized by the presence or absence of an event. In this situation, a NN is regarded as an alternative to traditional logistic regression methods. However, in the situation where outcomes are characterized by the time to an event, the application of NNs to predict clinical events requires development of a strategy to address the time course of a disease process. In cases where the event of interest does not occur, outcomes are regarded as (right)-censored. Simple exclusion of censored observations from the available training set would limit the amount of data available for network development and could lead to significant biases in event predictions. Several strategies have been developed to extend NN prediction methods to accommodate right-censored data. These are methods due to Farragi and Simon, 4 Liestol and co-workers, 5 and a modification of the Buckley-James method. 6 Because Cox regression analysis is an accepted solution to the problem of analyzing censored data, the performance of Cox regression models, when compared to those of NNs, can provide a useful perspective on the success of a NN approach. In this paper, we compare and evaluate all three NN methods with Cox regression in a Monte Carlo simulation study. Using the EPILOG PLUS PROC NEURAL utility developed by our group, feed-

forward back-propagation networks are examined under a variety of experimental conditions. Minimization methods are implemented that efficiently determine optimal parameters. To evaluate the performance of the NN methods relative to Cox regression analysis, a generalized version of the receiver operating characteristic (ROC) curve, the C index, is used as a measure of predictive accuracy. METHODS Neural Network Methods for Censored Data. Four strategies for applying neural networks to censored data were compared. These methods are summarized as follows: )DUUDJL6LPRQ 0HWKRG The method due to Farragi and Simon generalizes Cox regression to allow nonlinear functions in place of the usual linear combination of covariates (i.e., inputs). 4 Because a NN can be mathematically represented as a nonlinear function of covariates, they suggest a Cox-like model in which the NN output is used in place of the linear function. This method retains the proportional hazard nature of the Cox model, but provides the ability to model complexities and interactions in the input data that simple Cox models would miss. The Farragi-Simon method permits standard statistical methods for evaluating covariates; consequently, selecting the best model is straight-forward. 4 /LHVWRO$QGHUVHQ$QGHUVHQ0HWKRG For the method due to Liestol, Andersen and Andersen, survival times are grouped into time intervals during which the hazard is assumed to be constant and output nodes are established for each interval. 5 Under the conditions of no hidden layers, and identical weights from the same input node to all output nodes, the NN is trained to predict the conditional event probabilities for each person, with results being identical to a grouped version of the Cox analysis of the same data. Eliminating the constraint that all weights from a node be identical results in nonproportional hazards for the Cox regression Adding a hidden layer of nodes produces a form of Cox-NN hybrid with non-linear covariate effects; the degree of non-linearity depends on the number of hidden nodes and the choice of activation function. %XFNOH\-DPHV 0HWKRG Buckley and James developed an effective means of assigning event times for censored cases. 6 The original Buckley- James method, as applied to linear regression, accommodates censored observations by modifying the normal equations rather than the sum of squares of residuals. The method allows the dependent variable to be censored and the residual distribution to be unspecified. If censoring is evenly distributed, the estimated parameters are asymptotically consistent and efficient. 6 The Buckley-James approach has been generalized to the NN setting to determine the expected survival for all censored individuals (based on the current weight matrix) and to substitute the expected value for the censored value when determining and backpropagating the error. Specifically, NN predictions are compared to the actual values on each run and the differences (residuals) used to calculate a Kaplan- Meier-type curve. Based on the residual distribution (as reflected in the Kaplan-Meier curve) it is possible to estimate the expected survival time for any person who was censored. Minimization Algorithms. For a description of the following minimization algorithms, see the text by Bishop. 1 In their paper, Farragi and Simon recommended and utilized the Newton-Raphson algorithm to maximize the partial likelihood function and approximate both the first and second derivatives numerically. 4 In contrast, Liestol and coworkers recommended and utilized the gradient algorithm to minimize the error function via back propagation. 5 For the Newton-Raphson algorithm, the numerical approximation of the derivatives can lead to large round-up errors and can become computationally prohibitive for large networks. Although the gradient algorithm takes less time for a single iteration step, it requires many iterations and consequently more computing time. For these reasons, we evaluated a limited memory quasi-newton minimization algorithm for all four methods under study (see Ref. 1, pp. 289-290). This quasi-newton minimization algorithm builds up an approximation to the inverse Hessian matrix over a number of a steps, and uses a line-search to detect the learning rate. As a result, the quasi-newton algorithm requires less computer memory and converges much faster than the other algorithms. In preliminary evaluations, we found that the quasi- Newton algorithm performed very well for the Liestol-Andersen-Andersen and Buckley-James methods. However, we found that it was not a good approximation for non-quadratic error functions associated with the Farragi-Simon method. As a consequence, the simple gradient algorithm was used in this study for the Farragi-Simon method. NN Architecture and Parameters. For all four NN methods evaluated in this study, the architecture consisted of an input layer, one hidden layer and an

output layer. The input layer consisted of two or four inputs (covariates) plus a special bias node with value equal to 1. The hidden layer consisted of two hidden nodes and a special hidden node that played a role similar to the constant term in linear regression. For the Farragi-Simon method, this hidden node was not needed because its effect is incorporated into the baseline hazard. The output layer consisted of one output node for the Farragi-Simon and Buckley- James methods. The output layer consisted of multiple nodes for the Liestol-Andersen-Andersen method with the number of nodes equal to the number of intervals specified. In this paper, we considered three intervals with the cut points determined by the times when the survival probabilities are approximately 35% and 65%. For all NN methods, the logistic activation function was applied to the output of the hidden layer. For all but the Farragi-Simon method, the logistic function was applied to the output of the output layer. For the Farragi-Simon method, no activation function was used for the output layer. The error functions to be minimized were: the negative partial likelihood for the Farragi-Simon method; the quadratic error function for the Liestol-Andersen-Andersen method; and the quadratic error function with one output node for the Buckley-James method. The initial learning rate was set to 0.05 and updated using the line search algorithm. The criterion for convergence was that the absolute change in the error function was less than 10-6. Cox Regression. Standard Cox regression methods were used with only main effects from the covariates considered as inputs. Simulation Study. We conducted a Monte Carlo simulation study to evaluate the predictive accuracy of the four NN methods and Cox regression for handling censored data. We considered simulated data with two or four inputs (covariates), various censoring patterns, interaction between covariates, as well as proportional vs. non-proportional hazards. Let p i= 1 i j λ( t, X ) = exp{ β x + γ x x + λ x x x } ti i ijt i j ijkt i j k be the hazard at any time t given p covariates X. The following eight designs were considered. Design 1: p = 2 inputs with β 1 = 1, β 2 = 0.25 and no interaction (i.e., γ 12 = 0). The distributions of the inputs are: x 1 has a binomial distribution with probability 0.5, x 2 has a normal distribution with mean 0 and standard deviation 1, and x 1 and x 2 are i j k independent. All subjects were followed to extinction (i.e., no censoring). Design 2: Same as Design 1, except that a 20% censoring rate was applied, under the assumption that the censoring time was exponentially distributed and independent of the survival time. Design 3: Same as Design 2, except that an interaction between x 1 and x 2 was assumed, i.e., γ 12 = 0.2. Design 4: p = 4 inputs with four-way interactions: β 1 = β 2 = 2, β 3 = 0.5 and β 4 = 1.0, γ 12 = γ 13 = 1.0, all other γ s = 0.5. The distributions of the inputs are: x 1 and x 2 each have a binomial distribution with probabilities 0.25 and 0.5, respectively, and x 3 and x 4 each have a normal distributions with mean 0 and standard deviation 1, and all variables are independent. A 30% censoring rate was applied. Design 5: Same as Design 4, except the censoring rate was increased to 70%. Design 6: Modification of Design 2, so that the hazard is not proportional. The log hazard ratio = 1.0 and 0.25 for x 1 and x 2 before the time with 70% survival probability. Thereafter, the log hazard ratio was reduced to 0 for x 1 and x 2 (i.e., the survival probability was not associated with x 1 and x 2 ). Design 7: Equivalent to Design 3 with n = 200. Design 8: Equivalent to Design 5 with n = 200. For Designs 1-6, a database of 200 realizations (cases) was generated. The database was randomly split into 100 training cases and 100 testing cases. For Design 7 and 8, a database with 400 cases was generated and randomly split into 200 training cases and 200 testing cases. The predictive accuracy of each of the four NN methods and Cox regression was determined by the C index for both the training and the testing sets (see below). The simulation process was then repeated 50 times. For comparison purposes, the mean C index (± SD) was then tabulated. Repeated measures ANOVA were used to compare the performance among all methods. Tukey s studentized range test was used for pairwise comparisons between methods. The C Index of Predictive Accuracy. Motivated by rank tests based on Kendall s tau developed by Brown et al., Harrell et al. derived an index of concordance, the C index, which can be considered as a generalization of the area under the ROC curve for

censored data. 8,9 The C index is calculated by taking all possible pairings of patients. For a given pair, the predictions are said to be concordant with the outcome if the patient having a higher predicted probability of survival lived longer. If both patient s survival times are censored, or if only one died and the follow-up duration of the other was less than the survival time of the first, the pair is not counted. The C index is the proportion of predictions that are concordant out of all pairs of patients for which ordering or the survival times can be determined. Values of the C index near 0.5 indicate that the model is not predictive. Value of C index near 1 indicate the input data virtually always determine which patient has better prognosis. Software: Epilog Plus. For the simulation study, we used EPILOG PLUS (Pasadena, CA), an integrated statistical package, designed for use by epidemiologists and clinical trial statisticians. PROC NEURAL, a comprehensive set of features and options for modeling clinical data, is integrated within EPILOG PLUS so that users would have access to both NN and more conventional methods for modeling censored data. Cox regression analyses utilized SAS (Cary, NC). RESULTS The results of the simulation study for Designs 1-6 are presented in Table 1. For the training phase, two of the three NN methods significantly outperformed Cox regression: Faraggi-Simon for Design 4, and Liestol-Andersen-Andersen for Designs 2 6 (p < 0.05, ANOVA with Tukey studentized range tests for pairwise comparisons). For two designs (Designs 4 and 5), the Buckley-James performed worse than Cox regression (p < 0.05). For the testing phase, a different pattern of performance for the NN methods was observed. In the situation where there was no censoring and no interaction (Design 1), the Faraggi-Simon and Buckley-James methods performed as well as Cox regression. In contrast, the Liestol-Andersen- Andersen method performed significantly worse than Cox regression (p < 0.05). A similar pattern among the NN methods was observed in the situation where there was moderate censoring (approximately 20%) with or without an interaction between the two inputs (Designs 2 and 3). For the situation (Design 4) with four inputs, all possible interactions, and moderate censoring (approximately 30%), the Faraggi-Simon performed as well as Cox regression, and significantly outperformed the Buckley-James method, which in turn outperformed the Liestol-Andersen-Andersen method (p < 0.05). For Design 5, which is similar to Design 4, except that the rate of censoring was increased from 30% to 70%, the Faraggi-Simon performed as well as Cox regression, and significantly outperformed the other two NN methods (p < 0.05). In the situation of a non-proportional hazard model (Design 6), all NN methods performed as well as Cox regression. We then evaluated the effect of increasing the number of cases from n = 100 to n = 200 (data not shown in table). For the training phase, increasing the sample size generally decreased the performance of Cox regression and all NN methods for Design 7 (equivalent to Design 3) and for Design 8 (equivalent to Design 5). In addition, increasing the sample size reduced the variability in performance among the methods. In comparing similar designs, the Liestol- Andersen-Andersen method outperformed Cox regression and the other two NN methods for Design 3, whereas two of the NN methods (Liestol- Andersen-Andersen and Faraggi-Simon) and Cox regression performed similarly for Design 7. For Design 5, the Liestol-Andersen-Andersen method performed better and the Faraggi-Simon method performed the same as Cox regression, whereas both the Liestol-Andersen-Andersen and Faraggi-Simon methods performed the same as Cox regression for Design 8. For the testing phase, Cox regression and two of the NN methods (Faraggi-Simon and Buckley-James) performed similarly for Design 7. Cox regression and two of the NN methods (Faraggi-Simon and Liestol-Andersen-Andersen) performed similarly for Design 8. DISCUSSION The purpose of this paper is to compare three available neural network models for censored survival data and demonstrate the relative merits of these techniques. The simulations presented in this paper illustrate a representative sample of possible models and methods. Based on the mean C index, the overall finding of our Monte-Carlo study for Designs 1-8 is as follows: For the training phase, the Liestol-Andersen-Andersen method outperformed Cox regression for five of the eight designs (Designs 2-6), and the Faraggi-Simon method outperformed Cox regression for one of the eight designs (Design 4). The Buckley-James method performed worse than Cox regression for three of the eight designs (Designs 4, 5, 8).

For the testing phase of the study, the Faraggi-Simon method performed as well as Cox regression for all 8 designs. These results are not surprising in that the Faraggi-Simon method retains the proportional hazard nature of the Cox model, but provides the ability to model complexities and interactions in the input data. The Liestol-Andersen-Andersen method performed as well as Cox regression for 2 of the 8 designs. Specifically, the Liestol-Andersen-Andersen method performed as well as Cox regression in the situation of non-proportional hazards (Design 6) and in the situation of higher-order models with a high censoring rate and larger sample size (Design 8). These results are not surprising since we utilized the general form of the Liestol-Andersen-Andersen method, which permitted non-proportional hazards. Finally, the Buckley-James method performed as well as Cox regression for 5 of the 8 designs, including the design with non-proportional hazards (Design 6). However, the Buckley-James method performed worse than Cox regression in the case of higher level models with interactions (Designs 4, 5 and 8). In our study Cox regression was applied only in a simple manner. Other modeling approaches can also be developed to systematically explore potential data interactions and model complex relationships. Because a data analyst would typically test at least the two-way interactions, we reran the simulations for the Cox regression for Design 5. Considering all 2 nd order interaction in the Cox regression model improved the performance for the training set from 0.889 + 0.034 (no interactions) to 0.916 + 0.030 (2- way interactions). In contrast, performance for the testing set remained the same: 0.876 + 0.038 to 0.870 + 0.041. Another limitation of our study was that we considered only three intervals for the Liestol- Andersen-Andersen method. NN software has generated a great deal of interest partly because it effectively places advanced modeling tools in the hands of users of personal computers. One of the advantages of NNs is that they can detect complex patterns among the inputs. However, this was not seen in our study. In contrast, one disadvantage of NN models is that they can model idiosyncratic features of the training dataset. When this happens the NN has overlearned, and will appear to perform extremely well on the training dataset but further testing on new data will generally be far less successful. In fact, overlearning was apparent in our testing datasets. For example, comparing training to testing results for Design 5, Cox performance declined by only 0.013, whereas the Faraggi-Simon and Liestol-Andersen-Anderson methods declined by 0.041 and 0.049, respectively (Table 2). Overfitting can be reduced by using the m- item out validation approach, which is timeintensive. 10 Another disadvantage is that NNs are computationally intensive and may take a long time to train and converge to a solution. Using the quasi- Newton minimization algorithm 1, the Liestol- Andersen-Andersen and Buckley-James methods converged to a solution very fast. In contrast, the Faraggi-Simon method, which used the simple gradient minimization algorithm, took a long time to train. Additionally, the NN may converge not to the optimal solution, but rather to a local minimum, so that the resulting NN will perform sub-optimally. Some methods such as genetic algorithms can be used to avoid local minima. However, the genetic algorithm requires a lot of computer memory and is slow in convergence. One advantage of Cox regression, is that the regression coefficients can be interpreted as the likelihood of an outcome given some value(s) of the risk factor (e.g., odds ratios or relative risks). NN weights usually do not lend themselves to such interpretation. The NN methods presented here represent only a few of the many options in survival modeling. It may be more valuable to explore issues of methods, applications, and potential for improvement than to draw conclusions about whether one method is inherently superior to another. In summary, the results of our study shed light on the relative merits of the NN methods. In general, the results of our study also suggest that NNs can serve as effective methods for modeling right-censored data. However, the performance of the NN is somewhat variable, depending on the method that is used. REFERENCES 1. Bishop, CM. Neural Networks for Pattern Recognition. Oxford: Clarendon Press (1995). 2. Hornik, K. Multilayer feedforward networks are universal approximators, Neural Networks, 2, 359-366 (1989). 3. Warner, B and Misra, M. Understanding neural networks as statistical tools, American Statistician, 50, 184-293 (1996). 4. Faraggi, D and Simon, R. A neural network model for survival data, Statistics in Medicine 14, 73-82 (1995).

5. Liestol, K, Andersen, PK and Andersen, U. Survival analysis and neural nets, Statistics in Medicine, 13, 1189-1200 (1994). 6. Buckley, J and James, I. Linear regression with censored data, Biometrika, 66, 429-436 (1979). 7. Newman, TG and Odell, PL. The Generation of Random Variates. London: Charles Griffin and Co., p. 28 (1971). 8. Harrell, FE, Califf, RM, Pryor, DB, Lee, KL and Rosati,RA. Evaluating the yield of medical tests, Journal of the American Medical Association 247, 2543-2546 (1982). 9. Harrell, FE, Lee KL, Califf RM, Pryor DB, Rosati RA. Regression modelling strategies for improved prognostic prediction, Statistics in Medicine 3, 143-152 (1984). 10. Lachenbruch, PA and Mickey MR. Estimation of error rates in discriminant analysis, Technometrics 10, 1-11 (1968). Supported by: National Institutes of Health Grants: R03 HL55150 and NS RR33899

Table 1: Results of Simulation Study: C-index (mean + SD) (n = 100 cases, 50 replications) DESIGN 1 DESIGN 2 DESIGN 3 Method Training Testing Training Testing Training Testing COX 0.650 + 0.028a 0.647 + 0.029a 0.645 + 0.030b 0.652 + 0.035a 0.671 + 0.029b 0.652 + 0.039a FS 0.655 + 0.027a 0.642 + 0.035a 0.650 + 0.033a,b 0.648 + 0.034a 0.671 + 0.032b 0.648 + 0.041a Liestol 0.634 + 0.031b 0.600 + 0.041b 0.655 + 0.038a 0.620 + 0.051b 0.679 + 0.030a 0.635 + 0.037b Buckley 0.650 + 0.029a 0.648 + 0.039a 0.644 + 0.032b 0.646 + 0.043a 0.667 + 0.037b 0.647 + 0.042a DESIGN 4 DESIGN 5 DESIGN 6 Method Training Testing Training Testing Training Testing COX 0.814 + 0.032b 0.805 + 0.030a 0.889 + 0.034b 0.876 + 0.038a 0.604 + 0.028b 0.592 + 0.039a FS 0.825 + 0.032a 0.801 + 0.032a 0.904 + 0.032a,b 0.863 + 0.043a 0.609 + 0.029b 0.587 + 0.041a Liestol 0.823 + 0.031a 0.775 + 0.044c 0.906 + 0.034a 0.827 + 0.051b 0.623 + 0.031a 0.583 + 0.038a Buckley 0.803 + 0.033c 0.786 + 0.034b 0.861 + 0.066c 0.842 + 0.071b 0.597 + 0.043b 0.580 + 0.059a FS = Farragi-Simon; Liestol = Liestol-Andersen-Andersen method with 3 time intervals; Buckley = Buckley-James method. Design 1: γ = exp(x 1 + 0.25x 2 ), x 1 ~ bin(0.5), x 2 ~ N(0,1), no censoring Design 2: γ = exp(x 1 + 0.25x 2 ), x 1 ~ bin(0.5), x 2 ~ N(0,1), average censoring for training and testing sets = 19% each Design 3: γ = exp(x 1 + 0.25x 2 + 0.2x 1 x 2 ), x 1 ~ bin(0.5), x 2 ~ N(0,1), average censoring for training and testing sets = 20% each Design 4: γ = exp[2(x 1 + x 2 )+ 0.5(x 3 ) + 1.0x 4 + 1.0(x 1 x 2 + x 1 x 3 ) + 0.5 (x 1 x 4 + x 2 x 3 + x 2 x 4 + x 3 x 4 ) + 0.5 (x 1 x 2 x 3 + x 1 x 2 x 4 + x 1 x 3 x 4 + x 2 x 3 x 4 + x 1 x 2 x 3 x 4 )], x 1 ~ bin(0.25), x 2 ~ bin(0.5), x 3 ~ N(0,1), x 4 ~ N(0,1), average censoring for training and testing sets = 32% each Design 5: same as Design 4, average censoring for training and testing sets = 70% each Design 6: non-proportional hazard modification of Design 2. Log hazard ratio = 1.0 and 0.25 for x 1 and x 2 before 70% of survival time, respectively. Log hazard ratio = 0 for x 1 and x 2, thereafter. Values in a column that share the same letter (a, b, c) are not statistically different. Different letters indicate significant differences by Tukey s studentized range test.