The practical utility of incorporating model selection uncertainty into prognostic models for survival data

Size: px
Start display at page:

Download "The practical utility of incorporating model selection uncertainty into prognostic models for survival data"

Transcription

1 The practical utility of incorporating model selection uncertainty into prognostic models for survival data Short title: Incorporating model uncertainty into prognostic models Nicole Augustin Department of Statistics, University of Glasgow, Glasgow, G12 8QQ, UK Willi Sauerbrei Institut für Medizinische Biometrie und Medizinische Informatik, Universitätsklinikum Freiburg, Stefan-Meier-Str. 26, D Freiburg, Germany Martin Schumacher Institut für Medizinische Biometrie und Medizinische Informatik, Universitätsklinikum Freiburg, Freiburg, Germany April 8, 2004

2 Abstract Predictions of disease outcome in prognostic factor models are usually based on one selected model. However, often several models fit the data equally well, but these models might differ substantially in terms of included explanatory variables and might lead to different predictions for individual patients. For survival data we discuss two approaches for accounting for model selection uncertainty in two data examples with the main emphasis on variable selection in a proportional hazard Cox model. The main aim of our investigation is to establish in which ways either of the two approaches are useful in such prognostic models. The first approach is Bayesian model averaging (BMA) adapted for the proportional hazard model (Volinsky et al., 1997). As a new approach we propose a method which averages over a set of possible models using weights estimated from bootstrap resampling as proposed by Buckland et al. (1997), but in addition we perform an initial screening of variables based on the inclusion frequency of each variable to reduce the set of variables and corresponding models. The main objective of prognostic models is prediction, but the interpretation of single effects is also important and models should be general enough to ensure transportability to other clinical centres. In the data examples we compare predictions of the two approaches with conventional predictions from one selected model and with predictions from the full model. Confidence intervals are compared in one example. Comparisons are based on the partial predictive score and the Brier score. We conclude that the two model averaging methods yield similar results and are especially useful when there is a high number of potential prognostic factors, most likely some of them without influence in a multivariable context. The new method based on bootstrap resampling has the additional positive effects of achieving model parsimony by reducing the number of explanatory variables and dealing with correlated variables in an automatic fashion. keywords model selection uncertainty; survival analysis; model averaging; bootstrap; prognostic factor models. 1

3 1 Introduction Prognostic studies in medicine are about identifying factors influencing disease outcome. Identifying and ranking these factors involves statistical modelling, usually of a regression model of some form, e.g. in medicine this is often the logistic regression model or the Cox proportional hazards model (Cox, 1972). The resulting models are the basis for predictions of disease outcome, which are of potential value for therapeutic and other clinical decisions. For many diseases studied there is a vast number of potential factors and there is often a lot of uncertainty about the best model for prediction. For example in breast cancer the effects of more than 200 prognostic factors are currently discussed. Even if only 10 variables are investigated in one study, this can lead to a large set of possible models. Not surprisingly it is often the case that several models provide similar fits to the data, but these models differ in terms of included factors and predicted prognostic indices for individual patients. The differences are often substantial if these predictions are made at extreme values of the range of the data. Also, finding the variables with the highest prognostic value is a difficult task, mainly because selected models are often unstable (Steyerberg et al., 2001; Sauerbrei, 1999; Breiman, 1996b; Sauerbrei and Schumacher, 1992; Altman and Andersen, 1989; Chen and George, 1985). Standard statistical methodology involved in making the predictions and variance estimation is based on the assumption that the model used was pre-specified without any data-dependent modelling. However, every model presented is based on many implicit assumptions made before the actual formal model selection process starts. These assumptions include data interpretation, e.g. variable definitions, coding and dealing with missing values. Also a model class is chosen, in survival data whether a Cox model or a fully parametric model is fitted to survival data. Then a decision on the appropriate criterion for defining the best model for model selection is needed. These include combinations of measures on the discrepancy of fit (the deviance function), the prediction error, model stability or criteria like model complexity. Obviously, many other assumptions are required before starting a model selection process and decisions on these aspects can have a strong influence on the final model. See Blettner and Sauerbrei (1993) for some background on these pre-model selection decision processes for a specific example 2

4 and Chatfield (2002) for the discussion of further aspects. In this paper we will not tackle all the modelling uncertainties mentioned above. We consider two studies with survival data analysed with a Cox regression model. We restrict ourselves to only fitting linear terms in our model and assume that further assumptions of the Cox model are acceptable. We will concentrate on the issue of whether a variable is in- or excluded in a prognostic model. Despite these assumptions and pre-modelling decisions we still have thousands of possible models, e.g. 15 variables in one example results in 2 15 = candidate models, from which to select a prognostic model. Ignoring the uncertainty about the variable selection process may lead to unjustifiable confidence in parameter estimates and classification schemes derived from the selected model (Chatfield, 1995; Draper, 1995; Hoeting et al., 1999; Holländer et al., 2004). For example, confidence intervals may not yield their nominal coverage. One possible way to incorporate model selection uncertainty is to average over the set of possible models. In cases where the aim is to find the most likely model, it is useful to state the uncertainty attached to the selected model along with the results of other good approximating models. In this paper we describe some of the currently existing methods and propose a modification of one approach for handling model selection uncertainty. These are mainly Bayesian approaches (Chatfield, 1995; Draper, 1995; Kass and Raftery, 1995; Raftery et al., 1997; Hoeting et al., 1999) and methods based on bootstrap resampling (Buckland et al., 1997; Burnham and Anderson, 1998). In recent years these methods have been applied in many fields, such as mark-recapture (Stanley and Burnham, 1998; Brooks et al., 2000), ecology (MacKenzie and Kendall, 2002), econometry (Bunnin et al., 2002) and medicine (Viallefont et al., 2001). Here we are specifically interested in applications of the above approaches to survival data. The Bayesian model averaging (BMA) method was adapted to survival data by Volinsky et al. (1997); see also Hoeting et al. (1999) for an application. Buckland et al. (1997) proposed an approach using bootstrap resampling and use it in a simple example with a model space consisting of 4 models only. For more complex situations with a larger model space we propose a modification using a two step approach. First a subset of variables with no or little influence is eliminated and then model averaging is applied to models selected from the remaining variables using weights derived from subsequent bootstrap resampling investigation. We investigate whether these two methods can improve predictions for 3

5 prognostic models of cancer (glioma) and a liver disease (primary biliary cirrhosis) relative to three alternatives: (i) a single model selected using backward elimination (BE) with a predefined low nominal selection level, (ii) the full model and (iii) predictions using the Kaplan-Meier estimate, where no covariates are used. 2 Data Example 1: Glioma In a randomised clinical trial two chemotherapies were compared for patients with malignant glioma (Ulm et al., 1989; Sauerbrei and Schumacher, 1992; Sauerbrei, 1999). Besides therapy, 15 variables are considered to be potential prognostic factors (see (Table 1)). Here we use the data from 411 patients with complete information for the variables of interest. After a median follow-up time of 712 days, 274 deaths have occurred. In Sauerbrei (1999) the full model was considered and three other possible models were selected using BE with different significance levels (α =0.157, 0.05 and 0.01). Resulting models had nine, five and four variables respectively (Table 1). All models had the variables mali1, age, Karn1 and surg1 in common. Furthermore the linear predictors from the three models after selction are highly correlated with the predictor from the full model, with the highest pairwise Pearson s correlation coefficient of 0.99 with the BE(0.157) model having 9 variables and the lowest 0.94 (BE(0.01)) with 4 variables. Each of the three variables originally measured with three ordered categories were coded with two dummy variables. Eliminating one of the dummies corresponds to collapsing two categories which is sensible if two adjacent categories have a similar influence on outcome. Judged from the log-likelihood the full model has the largest value. However, the leaveone-out cross-validated log-likelihood, a measure of the predictive value of a model (Verweij and Houwelingen, 1993), shows that even the model based on BE(0.157) may include variables which do not increase the predictive value of the model. Cross-validated loglikelihood values of the models with 4 or 5 variables are very similar, the corresponding value of the full model is worse. 4

6 Table 1: Brain tumour (Glioma) example: Results and selected explanatory variables of the full model and different selected models using backward elimination (BE) with different significance levels are shown. l (1) is the cross-validated log-likelihood (Van Houwelingen and LeCessie, 1990). full model BE(0.157) BE(0.05) BE(0.01) log-likelihood l (1) 1) sex x anam: anamneses x mali1: grade of malignancy 2-3 x x x x mali2: grade of malignancy 3 x x age: age x x x x Karn1: Karnofsky index >= 70 x x x x Karn2: Karnofsky index > 80 x surg1: surgical resection (0 = biopsy, 1 = partial or total) x x x x surg2: surgical resection (total) x x conv: convulsiva x cort: cortisone x x epil: epilepsy x x x amne: amnesia x psyc: organic psychosyndrome x x apha: aphasia x 5

7 Example 2: Primary Biliary Cirrhosis Our second example are survival times of 312 patients with the fatal chronic liver disease primary biliary cirrhosis (PBC). The data are from a double-blind randomised clinical trial to compare the drug DPCA with a placebo. The clinical, biochemical, serological and histological parameters listed in Table 5 are possible prognostic factors for the disease. Fleming and Harrington (1991) showed that the drug DPCA did only have a negligible effect on prognosis, hence we do not include it. Following Hoeting et al. (1999) we exclude 8 observations due to missing values in the covariates. This leaves 304 patients with 123 events for this analysis. In addition we exclude the binary variable ascites (swelling in the abdominal cavity) due to the fact that less than 8% of cases had ascites present and this caused convergence problems in some of the bootstrap samples. In the full model the effect of ascites is negligible. Like Hoeting et al. (1999) we used the log of bilirubin, albumin, SGOT, alkaline phosphates. In contrast to Hoeting et al. (1999) we do not use the log of prothrombin time. From previous analysis (Fleming and Harrington, 1991) it is not clear whether transforming this variable improves the model fit. Also, using the log leads to a very high expected odds ratio in the full model and there are convergence problems in the bootstrap samples. Fleming and Harrington (1991) used a Cox regression model for the data and selected five of the variables (age, l.albu, l.bili, edema and l.prothr). Hoeting et al. (1999) and Raftery et al. (1996) analysed these data using BMA. 3 Survival models We will use the Cox model where the hazard λ at time t of the event is modelled as a function of the explanatory variables X = (x 1,...,x p ). The hazard function is λ i (t X) = λ 0 (t) exp(x i β), (1) where β is the vector of coefficients of the explanatory variables. x i is the covariate vector of patient i and λ 0 (t) is the baseline hazard. For prognosis the survivor function will be estimated by: Ŝ i (t) = [Ŝ0(t)] exp(x i where S 0 (t) is the baseline survivor function. 6 β) ˆ,

8 As apparent from the two data examples there is typically a high number of possible prognostic factors: often above 10. Even without fitting interactions and considering only the in- and exclusion of variables this results in 2 10 = 1024 models. In the glioma example we have 2 15 = and in the PBC example there are 2 13 = 8192 possible models. If prior information does not lead to a smaller set of candidate variables with a potential effect, the model space needs to be reduced. The two approaches we describe below achieve this (a) by directly choosing a subset of models using Occams window or (b) indirectly by choosing a smaller subset of explanatory variables after using a screening step. As variable selection method we will use backward elimination (BE) with the BIC criterion (BE BIC ). BE may be used as a substitute for all-subset procedure if the corresponding predefined selection level is chosen (Sauerbrei 1999). Because of the comparison to Bayesian model averaging we used the BIC criterion, however others with a higher selection level as AIC or a selection level of 0.05 could have been choosen. 4 Quantifying model selection uncertainty 4.1 Bayes approaches Bayesian model averaging (BMA) is one approach to incorporating model uncertainty into inference. We give a short summary of the method; more details are given in Chatfield (1995), Draper (1995), Kass and Raftery (1995) and Hoeting et al. (1999). Consider the possible set of models M 1,..., M j,..., M J with associated prior probabilities of M j being the true model. The set M 1,..., M J is often called the model space. Let θ be the parameter or quantity of interest, which may depend on the covariates, e.g. the linear predictor (prognostic index). The most general definition of θ is the whole predictive distribution. Given the data D the posterior distribution pr(θ M j, D) of the estimated quantity θ under a specific model j is averaged over the different models considered, weighted by the respective pr(m j D): pr(θ D) = J pr(θ M j, D)pr(M j D) (2) j 7

9 Both terms in equation 2 require solving integrals, for the general linear model these are available in closed form; this is not the case for the Cox model. The posterior summary statistics are evaluated by incorporating the model posterior probabilities (Hoeting et al., 1999; Draper, 1995), e.g. the posterior mean is E(θ D) = L E(θ M j, D)pr(M j D). (3) j The specification of prior model probabilities pr(m j ) assumes that one of the models considered is true, since all other models which do not belong to the set M 1,..., M J have a prior probability of zero. Defining suitable priors pr(m j ) for the different models appears to be difficult with this approach, especially with a large set of possible models. The simplest suggested solution is to give each model equal weight when no prior information is available. This choice of prior can be problematic if two explanatory variables are correlated leading to two models M 1 and M 2 containing one of the two correlated variables respectively, then the two models will be almost equivalent. Giving the two models equal weight a priori amounts to giving the single model of which M 1 and M 2 are slightly different versions twice as much actual prior weight. Another problem of the BMA approach is that large summations and integrals are necessary especially if the model space is large, and these are often hard to compute. One approach to handle the large model space is to restrict it to models with a posterior probability above a certain threshold, e.g. by using Occam s window (Madigan and Raftery, 1994). Occam s window includes all models with a posterior probability greater than a certain proportion of the model with the highest posterior. 4.2 Bayesian model averaging for survival data (approx. BMA) For carrying out BMA within the Cox model, the integrals of both components in equation 2 are not available in closed form. Volinsky et al. (1997) suggest several approximations (see also Hoeting et al. (1999)). Instead of integrating the posterior distribution of the quantity of interest (Volinsky et al., 1997) use the partial maximum likelihood approximation. If the quantity of interest is the prognostic index, this so called predictive distribution is the partial likelihood with the maximum likelihood estimator ˆθ j inserted. For deriving pr(m j D), the posterior model probabilities in equation 2, the log of the inte- 8

10 grated likelihood is approximated by the Bayesian information criterion (BIC) (Schwarz, 1978). Before averaging, models not supported by the data are excluded using Occam s window. These are all models where the posterior probability is smaller than a specified constant, e.g. 1/C of the model with the highest posterior probability. Instead of calculating all posterior model probabilities an altered version of the leaps and bounds algorithm, also used for all subset selection (Furnival and Wilson, 1974; Lawless and Singhal, 1978; Kuk, 1984), is used to narrow down the choice of models. Volinsky et al. (1997) claim that by choosing a large enough set of best models with the leaps and bounds algorithm, all models within Occam s window are found. Posterior mean estimates and variance of θ are calculated using weighted averages from the separate models with normalised weights proportional to pr(m j D), with summation over models within Occam s window. Volinsky et al. (1997) apply this approach to Cox models for assessing the risk of stroke from a cardiovascular health study. As a measure for the importance of the different variables, posterior effect probabilities pr(β j 0 D) for the coefficients associated with each variable, are given. This is the sum over all pr(m j D) of all models containing variable x i. In the following we will call their approach approx. BMA. To our knowledge it is currently not implemented in any of the mainstream statistics packages or Bayesian specialist software such as BUGS, but the authors have made some of their code publicly available under Non-Bayesian approaches Buckland et al. (1997) have suggested several more or less ad hoc approaches to accounting for model uncertainty. In their approach an information criterion (I), e.g. the Akaike information criterion, is used to estimate model weights for parameter averaging and variance estimation. These weights have the following form w j = exp( I j /2) J i=1 exp( I, j = 1,...J, i/2) where I j is the information criterion for model j. The expected value of the averaged parameter is then ˆθ = J j 9 ˆθ j w j, (4)

11 where ˆθ j is the estimated parameter or quantity from the observed data. The variance of ˆθ j is estimated as follows (Buckland et al., 1997): { J 2 var(ˆθ) = w j var(ˆθ j γ j ) + γj} 2 (5) j where the model misspecification bias γ j is estimated by ˆγ j = ˆθ j ˆθ. ˆθj and ˆvar(ˆθ j γ j ) are the conditional estimates under the selected model j and ˆθ is found using equation 4. These estimates are plugged into the above equation. Setting I equal to the Bayes information criterion, the weights in equation 4 are almost equal to the model posterior pr(m j D) as it is approximated by Volinsky et al. (1997) when non-informative priors are used (see above). The main difference is that in pr(m j D) the normalising constant is not the sum over all possible models, but a sum over a smaller subset of best models as identified by the leaps and bounds algorithm and Occam s window. As an alternative way of obtaining model weights w j Buckland et al. (1997) suggest to apply the bootstrap to model selection and give an illustration for an example with 4 models under consideration. Then the weight of model j is estimated by the proportion of bootstrap samples in which model j was selected. Then bootstrap percentile confidence intervals are given by ordering the bootstrap estimates from smallest to largest and selecting the appropriate percentiles from the list. The idea of bootstrapping the modelling process has been suggested in the past mainly for investigating model selection stability issues e.g. Efron and Gong (1983), Chen and George (1985), Dijkstra and Veldkamp (1994), Altman and Andersen (1989), Sauerbrei and Schumacher (1992) and Hjorth (1994). Not many investigations exist incorporating a term to account for model uncertainty in the estimate of the variance. In the special case of two orthogonal covariates Candolo et al. (2003) show analytically that in the linear regression model the above approach reduces the mean squared error if effects of covariates are relatively weak. They support their findings by results from a simulation study considering 8 models. Especially when working with survival data resampling patients is the preferred bootstrap approach. A method that uses model averaging is bootstrap aggregating or bagging (Breiman, 1996a). In bagging the bootstrap is applied to model selection and the parameter of 10

12 interest, e.g. the prognostic index, is an average of the B bootstrap estimates: ˆθ = 1 B B j ˆθ b j, where ˆθ b j is the j th bootstrap estimate with j = 1,..., B. 4.4 Screening of variables and bootstrap model averaging (bootstrap MA) We use the bootstrap approach of Buckland et al. (1997) but eliminate in a first step single variables thought to have a weak influence at most before we apply model averaging. This screening step helps to reduce the number of models substantially, as we have thousands of models in our examples, in contrast to only 4 models in Buckland et al. (1997). It helps to get more reliable weights for the models and it is most helpful for the use of the predictor in future applications as it does not require to measure all variables, even if they only have a small influence on the predictor. Our approach is also in contrast to approx. BMA approach where the set of models is reduced by Occam s window, however each variable may still be required because it is included in at least one of the models passing Occam s window criterion. For the first step we use parts of a strategy suggested by Sauerbrei and Schumacher (1992) based on bootstrap resampling. Model selection is carried out in the bootstrap samples. In order to make our selection procedure comparable to approx. BMA we use BE with a selection level corresponding to the BIC criterion where the penalty for each covariate is log(n), n is here the number of events in the survival data. We denote this selection algorithm by BE BIC. Then the number of times each covariate is included in the model, i.e. the inclusion frequency h(x i ), is recorded. Instead of the BIC criterion another criterion, e.g. the AIC or a nominal selection level, say 0.05, can be used. Now a subset of variables is selected only containing variables with an inclusion frequency above a minimum value. Variables with no or hardly any prognostic influence will have very low inclusion frequencies and will therefore be eliminated in this first step (Sauerbrei and Schumacher (1992)). Obviously, the selection criteria influences the interpretation of the inclusion frequencies; see Sauerbrei and Schumacher (1992) where this is discussed for the special situation of one variable. In a second step we carry out another bootstrap resampling procedure and apply BE BIC to the subset of remaining variables from the 11

13 first step. The set of possible models is substantially reduced after the first step, e.g. from 2 15 to 2 7 in the glioma example (Table 2). Now we record the number of times each model is selected and use these inclusion frequencies h(m j ) as weights in equation 4 for model averaging. The screening step also influences estimation of the variance of the predictor, as it uses the bootstrap samples of the second step. Furthermore, the second step in our procedure makes model selection in each bootstrap sample much faster and gives larger weights to models supported by the data because it considers only models in which every variable has probably some influence. This allows us to work with a relatively small number of bootstrap samples (several hundreds to about a thousand). The precision of relative model inclusion frequencies h(m j ) can be increased by a larger number of bootstrap replicates. 5 Assessing the predictive performance 5.1 Criteria We compare the predictive performance of the different approaches using the partial predictive score (PPS) and the Brier score (Brier, 1950; Graf et al., 1999). Partial predictive log-score (PPS): Volinsky et al. (1997) suggested using the partial predictive log-score (PPS). This is a possible criteria if the quantity of interest is the prognostic index. Let D build be the data used for the model building process and D test the data used for validation. The PPS is based on the logarithmic scoring rule, the summed logarithm of the observed ordinate of the predictive density distribution for the observations in D test : log pr(d M j, D build ) d D test and for BMA the PPS is then { J } pr(d M j, D build ) pr(m j D build ). d D test log j=1 Since for the Cox model the predictive density pr(d M j, D build ) is not available. Volinsky et al. (1997) suggest to use the partial likelihood (PL) instead, with the maximum 12

14 Table 2: Glioma example: Results of the bootstrap MA approach. The inclusion frequencies h(x i ) are based on the models selected in 500 bootstrap samples with BE BIC. The 27 models are the result of step two based on 7 explanatory variables with h(x i ) > 30% selected in step one. The corresponding h(m j ) resulting from different selection levels in step one are also given. Variables with h(x i ) > 30% and part of the subset of explanatory variables after step one. explan. step one: step two: 10 models variables h(x i )% with the highest h(m j )% out of 27 sex anam mali1 81 x x x x x x - x - - mali x x - x x age 100 x x x x x x x x x x Karn1 64 x x x - - x x - x x Karn surg1 99 x x x x x x x x x x surg conv cort x x - x epil 37 - x - - x - x amne psyc apha h(m j )%, h(x i ) > h(m j )%, h(x i ) > h(m j )%, h(x i ) > h(m j )%, h(x i ) > h(m j )%, h(x i ) >

15 likelihood estimator ˆβ j from D build inserted. PPS = { J } PL(ˆβ build )pr(m j D build ), d D test log j=1 then the PPS is the sum of the partial log-likelihood evaluated for each observation in D test. That is (assuming no ties): PL(ˆβ build ) = i D test exp(x iˆβ build ) l R(t i ) exp(x lˆβ build ) where x i = (x i1,..., x ip ), t i is the observed time for the i th patient. The indicator δ i is zero if t i is right-censored, unity otherwise. R(t i ) is the risk set at time t i. Since the partial likelihood is only defined for a whole data set, the derivation is not straightforward for a subset. The cross-validated leave-one-out partial log-likelihood as defined by Verweij and Houwelingen (1993) takes account of this problem. The above definition of the partial likelihood is only an approximation and is only valid for a fairly high number of observations in the test data. We use k-fold crossvalidation where the sample size of the test data is between 16 and 20 in the two examples presented. For evaluating the PPS under bootstrap MA the weight pr(m j D build ) is replaced by bootstrap inclusion frequency h(m j ) for model j. The higher the PPS the better the predictive performance. If observations in D test yielding high ordinates in the predictive density distribution as derived from the model selected and estimated from D build, the model is also plausible for D test. Brier score: If the quantity of interest is the predicted survivor function at a given time t, which is a function of the linear predictor and the baseline hazard function, a criterion evaluating the predicted survivor function directly is possibly most appropriate. The Brier score (Graf et al., 1999) is such a criterion. The prediction of the survivor function involves estimation of the prognostic index and of the integrated baseline hazard t 0 λ 0(s)ds in equation 1. The Brier score is the expected prediction error at a given time point t : BS(t ) = E[(I(T > t ) Ŝ(t X)) 2 ], where I(.) is an indicator function. It is estimated as follows BS(t ) = 1 n v i (I(t i > t ) n Ŝ(t x i )) 2, i=1 14 δ i

16 where the weights v i are equal to one if no censoring is present at all. With censoring, weights for observations with an event are equal to the inverse of the Kaplan-Meier estimate of the censoring distribution at the time of the event. Individuals without an event up to t are weighted according to the Kaplan-Meier estimate of the censoring distribution at time t. Observations censored before time t do not contribute to the Brier score, i.e. v i = 0. The lower the Brier score the better the predictive performance. With model averaging the survival probability S(t X) is a weighted average of the estimated survivor functions under the different models. In the situation where one data set D build is used for the model building process and another data set D test is used for validation, the parameters β and the baseline hazard λ 0 are estimated from D build. The Brier score from the pooled Kaplan-Meier estimate of the survivor function is a benchmark value for a prediction without taking explanatory information into account. For measuring the gain from explanatory information in the model we use the percentage change in Brier score compared to a prediction from the Kaplan-Meier estimate (Graf et al., 1999): R2 = 1 BS(t ) model / BS(t ) K M. R 2 is thus analogous to the explained variation commonly quoted for general linear models. A naive prediction with constant survival probability of 0.5 yields a BS of 0.25: useful as a worst case benchmark. 5.2 Cross-validation We use K-fold cross-validation (Davison and Hinkley, 1997) to estimate the predictive criteria. The data are split into K disjoint sets of similar size. A rule of thumb is to use either K = n or K = 10 in the case of large data sets, where choosing K = n becomes very computer intensive (Davison and Hinkley, 1997). These K sets define the test set D test for assessing the predictive performance and the build set D build used to fit the model. In turn, one of the K sets is the test set and the remaining data is the build set. This procedure ensures one prediction per observation. For survival data it is sensible to stratify the observations in the K sets by event and censored cases. Then the following cross-validation procedure is carried out: 1. Perform the model building procedure (Kaplan-Meier, full model, backwards elimination (BE BIC ), bootstrap MA or approx. BMA) on the build data to obtain: (a) under the full model and BE BIC the estimates ˆβ build and 15 ˆλ build 0. ˆβ build is a vector

17 of estimated regression coefficients for variables selected, coefficients of variables which are not selected are set to zero and ˆλ build 0 are the baseline hazard estimates. (b) under bootstrap MA the estimates referring to the 1 to J selected models ˆβ build build 1,..., ˆβ J and build build ˆλ 01,..., ˆλ 0J referring to the 1 to L selected models are found. Under approx. BMA the estimates build build build build ˆβ 1,..., ˆβ L and ˆλ 01,..., ˆλ 0L. 2. Using the parameters estimated in the build data we predict the quantity of interest for the test data. (a) full model and BE BIC : Ŝ(t X test ) or the linear predictor ˆθ = X test β build ˆ for the test data, with covariate matrix X test. (b) bootstrap MA: Ŝ1(t X test )...ŜJ(t X test ) or the linear predictor ˆθ 1...ˆθ J for the test data, with covariate matrix X test are averaged using equation 4. approx. BMA: Ŝ1(t X test )...ŜL(t X test ) or ˆθ 1...ˆθ L are averaged using equation Calculate the assessment criteria, e.g. the Brier score or the PPS. 4. repeat steps 1-3 for each of the K data splits and sum the criteria obtained from the 1 to K data splits. When D test is too small, say below 10 observations, the censoring distribution for the weights in the Brier score is calculated using D build. However in our examples, the sample sizes of D test are around 20 and hence this is not a problem. 6 Examples 6.1 Glioma We applied our bootstrap MA approach to the glioma data as described above using 500 bootstrap resamples in all steps involved. In the first step of bootstrap MA the inclusion frequencies h(x i ) range between 4% and 100% (Table 2 and Figure 1). The variables mali1, age and surg1 have the highest inclusion frequencies, all are above 75%. The variable Karn2 has the lowest inclusion frequency of 4%. In the second step of the bootstrap MA procedure only variables with h(x i ) > 30% are included and the bootstrap procedure is applied again to the remaining subset with 16

18 Glioma example bootstrap inclusion frequencies (%) age surg1 mali1 Karn1 ** * * * * * * * * p-value (log) from full model Figure 1: The bootstrap inclusion frequencies versus p-values from full model for the Glioma data. mali1, mali2, age, Karn1, surg1, epil and cort. In total there were 27 different models selected out of 2 7 possible models; the top ten are shown with model inclusion frequencies h(m j ) in Table 2. The model including variables mali1, age, Karn1 and surg1 has the highest inclusion frequency (17%). Repeating the bootstrap MA procedure with different selection levels h(x i ) > 40%, 20%, 10% and 0% in the first screening step always yields the same model with the highest h(m j ), but, of course, the distributions of the respective h(m j ) differ. The order of the next important models differ also (Table 2). For the relatively high selection level of 40% in the screening step only 7 of the 16 (2 4 ) possible models were selected and in 99% of the replicates one of the 4 top models was selected. For selection levels 20%, 10% and 0% the resulting model sets are much larger. There where 107, 184 and 226 models in the respective sets. This reduced the h(m j ) of the strongest models. When applying approx. BMA to the glioma data, using model priors pr(m j ) which give each model equal weight, we obtain fairly similar answers. As in the example of Volinsky et al. (1997) all models with a posterior probability greater 1/C with C = 20, were included. In total 41 models were selected into the set. Table 3 shows the first 10 models with the highest posterior model probabilities. The model representing the highest part 17

19 Glioma example bootstrap inclusion frequencies (%) amne anam apha Karn2 cort mali2 psyc surg2 sex epil Karn1 age surg1 mali posterior prob (%) Figure 2: The bootstrap inclusion frequencies versus the posterior effect probabilities pr(β l 0 D) for the Glioma data marked with *. Posterior effect probabilities pr(β l 0 D) derived without Occam s window model space reduction are marked with. (16%) of the total posterior model probability is the same as was selected by BE BIC and under bootstrap MA. The variables mali1, age and surg1 have the highest posterior effect probabilities pr(β j 0 D) of 100%. The variables age and surg1 also have a 100% inclusion frequency under bootstrap MA but mali1 has only an inclusion frequency of 81 %. Figure 2 shows the scatter plot of bootstrap inclusion probabilities from the bootstrap MA approach versus the posterior effect probabilities pr(β j 0 D). With BMA variables with strong effects appear more important than with bootstrap MA, i.e. obtain a higher posterior effect probability compared to the inclusion frequency. Variables with weak effects appear more important with bootstrap MA compared to BMA. Some of these differences are due to the fact that the pr(β 0 D) are derived after the model space has been reduced by Occam s window, whereas the h(x i ) are estimated before the model space reduction. The pr(β 0 D) resulting from the BMA procedure without reducing the model space yield a closer agreement with the h(m j ) for most variables (also shown in Figure 2). We compare the predictive performance of the model averaging methods with the performance of predictions using the Kaplan-Meier estimate without covariates, the full Cox 18

20 Table 3: Glioma example: Results of approx BMA. The second column shows the posterior effect probabilities pr(β l 0 D). In the right half models with the highest pr(m j D) are shown. explan. Pr(β l 0 D) 10 models with the variables (%) highest pr(m j D) out of 41 sex x - x anam mali1 100 x x x x x x x x x x mali x age 100 x x x x x x x x x x Karn1 82 x x x x - x - x x x Karn surg1 100 x x x x x x x x x x surg x conv cort x epil 32 - x x - - x amne psyc x - apha pr(m j D)(%)

21 Table 4: Glioma example 20-fold cross-validation results: Partial predictive log-score (PPS) and Brier score estimated using 20-fold cross-validation. methods used for prediction Kaplan- bootstrap approx. full criterion Meier MA BMA BE BIC model PPS n.a BS(t = 310) R 2 (%) BS(t = 730) R 2 (%) model and the single model as chosen by BE BIC. We predict the probability of survival at a relatively early time point after diagnosis, at 310 days, and after two years at 730 days. These time-points where chosen by looking at the overall Kaplan-Meier curve, at day 310 the estimated survival probability is about 0.5. Day 730 is after the median follow-up time of 712 days and the estimated survival probability is about 0.2. We estimate the Brier score using 20-fold cross-validation. The results shown in Table 4 do not indicate any major differences. There is some benefit in introducing explanatory information: the percentage reduction of the Brier score (R 2 ) is around 15% for the four model based predictions at day 310 and at day 730 it is around 28%. The full model yields the highest R 2 at day 310 and at day 730 the lowest. Differences in predictions of the prognostic indices with the four model-based approaches are also small, the estimated partial predictive scores (PPS) from the 20-fold cross-validation are all similar (Table 4), with bootstrap MA yielding the lowest, i.e. worst, and the full model the highest, i.e. best, PPS. 20

22 6.2 Primary Biliary Cirrhosis In the PBC example our pre-model selection decisions are as follows. We apply BE BIC to bootstrap samples in the bootstrap MA approach. For the 123 events this is in correspondence to the BIC with a penalty of log(123). The results given in Table 5 show that the model selection uncertainty in this example is much higher than in the glioma example. 65 different models were selected in the second step of bootstrap MA, of which inclusion frequencies are all below 12%. Table 6 shows the corresponding figures for approx. BMA, where only 32 different models were selected. The model with the highest posterior probability (18%) is the same as for the bootstrap MA method. In Figure 3 we see that the relation between the bootstrap inclusion frequencies and posterior effect probabilities is similar as in the glioma example. The variables l.albu, age, prothr, l.urcop and edema with high posterior probabilities appear to be more important under approx. BMA than under two step bootstrap MA. Variables with low posterior probabilities appear more important under two step bootstrap MA. We compare predictive performance of our averaging approaches with the predictions using the Kaplan-Meier estimate, the full Cox model, and the selected model from BE BIC (Table 7). The Brier score is estimated, at given time points t =1000 and 3000 days, approximately 3 and 8 years; with an estimated survival probability of about 0.8 and 0.6 respectively using the Kaplan-Meier method. At t =1000 the percentage reduction of the Brier score (R 2 ) is around 35% for the four model-based predictions; the full model yields the highest R 2 = 36% and there is no difference between bootstrap MA and approx. BMA. At t = 3000 the bootstrap MA procedure leads the highest R 2 = 31%, BE BIC gives only 28%. Figure 4 shows the distribution of the proportion in bootstrap percentile confidence interval width of the full model for bootstrap MA and BE. The differences between models are larger for the confidence intervals than for the Brier Score or R 2 criterion. On average, the full model yields the widest confidence interval, BE BIC the narrowest and bootstrap MA is in between. The intervals from BE BIC are shorter than intervals of the full model for all observations and the median width is 0.61 of the width of intervals from the full model. The intervals from bootstrap MA have a median width which is 0.88 of the of the width of intervals from the full model. In contrast to the full model, bootstrap MA and 21

23 pbc example bootstrap inclusion frequencies (%) platel sex spiders l.alkal hepato l.sgot histo prothr l.urcop l.sbili edema age l.albu posterior prob (%) Figure 3: The bootstrap inclusion frequencies versus the posterior effect probabilities pr(β l 0 D) for the PBC data. proportion of confidence interval width compared to full model bootma BE Figure 4: PBC example. Distribution of the proportion of the width of confidence intervals of linear predictors for all observations. Proportion of the full model are shown for bootstrap MA and BE BIC interval width. The dotted line at a proportion of one shows where the confidence intervals of the compared methods have the same width. 22

24 Table 5: PBC example: Results of the bootstrap MA approach. The inclusion frequencies h(x i ) are based on the models selected in 500 bootstrap samples with BE BIC. The 65 models are the result of step two based on 7 explanatory variables with h(x i ) > 30% selected in step one, of which 10 are shown. For these models the corresponding h(m j ) resulting from different selection levels in step one are also given. Model selected by Fleming and Harrington (1991). Model selected using backward elimination with the BIC criterion (BE BIC ). step one: step two: 10 models explan. h(x i ) with the highest h(m j )% out of 65 variables % age 86 x x x x x x x x x x sex hepato: hepatomegaly spiders edema 62 x - - x x x x - x x l.sbili : log(bilirubin) 100 x x x x x x x x x x l.albu : log(albumin) 85 x x x x x x x x x x l.urcop :log(urine copper) 60 x x x x x x l.alkal: log(alkaline phosphates) l.sgot: log(sgot) x x platel: platelets prothr : prothrombin time 66 x x x - x x - x - x histo: histologic stage x - - x x - - h(m j ) %

25 Table 6: PBC example: Results of approx BA applied. The second column shows the posterior effect probabilities pr(β l 0 D). 10 of 33 models with the highest posterior probability are shown. explan. Pr(β l 0 D) 10 models with the variables % highest pr(m j D) out of 32 age 100 x x x x x x x x x x sex hepato spiders edema 87 x x x x - x x x x x l.sbili 100 x x x x x x x x x x l.albu 100 x x x x x x x x x x l.urcop 73 x x x x x - x - x - l.alkal l.sgot x x x x platel prothr 75 x x - - x x x x x x histo 40 - x x x x - pr(m j D)(%)

26 Table 7: PBC example: Partial predictive log-score (PPS) and Brier score at day 1000 and 3000 estimated using 20-fold cross-validation. methods used for prediction Kaplan- bootstrap approx. full criterion Meier MA BMA BE BIC model PPS n.a BS(t = 1000) R 2 (%) BS(t = 3000) R 2 (%)

27 BE BIC do not require information from all variables. 7 Discussion In statistical analysis it has been common practice to base inference on a conditional model, a model given a priori. This ignores uncertainty of model predictions, estimates of effects and variance caused by model selection. Breiman (1992) calls this a quiet scandal. A possible solution to the problem may be to use a weighted average of the models when making predictions and similarly to base variance estimates on several models, rather than on one conditional model. Most of this methodology has been developed in a Bayesian framework and was put into practice in the last decade (Chatfield, 1995; Draper, 1995; Hoeting et al., 1999). In this paper we use Bayesian model averaging (approx. BMA) and suggest a bootstrap approach (bootstrap MA) for model averaging which has its roots in the papers by Sauerbrei and Schumacher (1992) and Buckland et al. (1997). We present survival analyses from two studies where we concentrate on variable selection in the Cox model. We assume that the effects are linear, constant over time, interactions between covariates do not exist and that all assumptions implied by the Cox model hold. Obviously, this is still a severe restriction of the parameter space considered. However, we see no chance to escape from this problem of many simplifying assumptions. In a prototype linear regression analysis tool more model selection aspects have been incorporated to consider the cost of data analysis Faraway (1992), but we are not aware of any substantial improvement from this approach. In our bootstrap MA procedure we start with a screening step aiming to eliminate variables which have no or at most a weak effect. The cost may be to loose a variable which may improve a predictor slightly, however this step reduces the computational burden substantially, more importantly it gives a clearer impression concerning the models most supported by the data. This helps with the interpretation and may thereby increase the practical usefulness of the results, e.g. by making the transportation of the model to other medical centres easier. The screening step of our approach can reduce the number of variables used in the predictor substantially, e.g. in the Glioma example the predictor is based on only 7 out of 15 variables. For a more parsimonious model even less variables 26

28 could be retained by increasing the cut-off value for the variable selection frequency h(x i ) above which variables are included in the subset after the first screening step. Based on the cross-validated log-likelihood we may conclude in the Glioma example that the full model and the model with 9 variables selected by BE (0.157) may be overparameterised. Hence they are less useful for describing data of a similar study collected under slightly different conditions, e.g. at a different time point, or data from different clinics but under the same protocol. In addition, it is questionable whether all factors from the full model will ever be collected in a similar study. Therefore, a model requiring all variables lacks practical usefulness. The Occam s window used in Bayes model averaging excludes models not supported by the data, but the remaining models may still require to measure all the available variables, although some of them may have a close to zero weight. Hence if model parsimony is important in aiding interpretability, reproducibility and transportability of a model (Sauerbrei, 1999; Altman and Royston, 2000; Burnham and Anderson, 2001) the bootstrap MA approach has an advantage over the BMA method. In both model averaging approaches, more or less arbitrary decisions on settings for the model space reduction are made. For bootstrap MA, this is the cut-off value of h(x i ) above which variables are included in the subset after the first screening step. In approx. BMA it is the setting of C in the Occam s window. Only models with a posterior probability greater than 1/C are included. In the glioma example we have shown how the number of selected models changes depending on the cut-off value of bootstrap MA. In bootstrap MA, results may also be sensitive to the choice of the selection level applied in the backward elimination procedure. We have used the BIC criterion to make our procedure comparable to the approx. BMA procedure. To what extent results are sensitive to these settings is currently under investigation in a large simulation study. In both of our examples there is a potential gain from using the proposed techniques. For the Glioma example predictions from both model averaging methods were very similar and do not differ substantially from predictions of conventional methods (BE BIC, full model). In studies with a larger number of potential predictors it is more likely that many of them have no influence in a multivariable context and differences between the approaches are more likely. In the Glioma example the similarity of results is not surprising as the actual reduction of the Brier score of all model-based predictions is relatively small with a maximum of R 2 of 17% at day 310 and 30% at day 730. The sample size in relation 27

Variable Selection in Cox s Proportional Hazards Model Using a Parallel Genetic Algorithm

Variable Selection in Cox s Proportional Hazards Model Using a Parallel Genetic Algorithm Variable Selection in Cox s Proportional Hazards Model Using a Parallel Genetic Algorithm Mu Zhu and Guangzhe Fan Department of Statistics and Actuarial Science University of Waterloo Waterloo, Ontario

More information

Fractional Polynomials and Model Averaging

Fractional Polynomials and Model Averaging Fractional Polynomials and Model Averaging Paul C Lambert Center for Biostatistics and Genetic Epidemiology University of Leicester UK paul.lambert@le.ac.uk Nordic and Baltic Stata Users Group Meeting,

More information

The influence of categorising survival time on parameter estimates in a Cox model

The influence of categorising survival time on parameter estimates in a Cox model The influence of categorising survival time on parameter estimates in a Cox model Anika Buchholz 1,2, Willi Sauerbrei 2, Patrick Royston 3 1 Freiburger Zentrum für Datenanalyse und Modellbildung, Albert-Ludwigs-Universität

More information

A new strategy for meta-analysis of continuous covariates in observational studies with IPD. Willi Sauerbrei & Patrick Royston

A new strategy for meta-analysis of continuous covariates in observational studies with IPD. Willi Sauerbrei & Patrick Royston A new strategy for meta-analysis of continuous covariates in observational studies with IPD Willi Sauerbrei & Patrick Royston Overview Motivation Continuous variables functional form Fractional polynomials

More information

Towards stratified medicine instead of dichotomization, estimate a treatment effect function for a continuous covariate

Towards stratified medicine instead of dichotomization, estimate a treatment effect function for a continuous covariate Towards stratified medicine instead of dichotomization, estimate a treatment effect function for a continuous covariate Willi Sauerbrei 1, Patrick Royston 2 1 IMBI, University Medical Center Freiburg 2

More information

Quantitative Methods for Stratified Medicine

Quantitative Methods for Stratified Medicine Quantitative Methods for Stratified Medicine The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Yong, Florence Hiu-Ling.

More information

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j Standard Errors & Confidence Intervals β β asy N(0, I( β) 1 ), where I( β) = [ 2 l(β, φ; y) ] β i β β= β j We can obtain asymptotic 100(1 α)% confidence intervals for β j using: β j ± Z 1 α/2 se( β j )

More information

Residuals and model diagnostics

Residuals and model diagnostics Residuals and model diagnostics Patrick Breheny November 10 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/42 Introduction Residuals Many assumptions go into regression models, and the Cox proportional

More information

MAS3301 / MAS8311 Biostatistics Part II: Survival

MAS3301 / MAS8311 Biostatistics Part II: Survival MAS3301 / MAS8311 Biostatistics Part II: Survival M. Farrow School of Mathematics and Statistics Newcastle University Semester 2, 2009-10 1 13 The Cox proportional hazards model 13.1 Introduction In the

More information

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010 Model-Averaged l 1 Regularization using Markov Chain Monte Carlo Model Composition Technical Report No. 541 Department of Statistics, University of Washington Chris Fraley and Daniel Percival August 22,

More information

Introduction to Statistical Analysis

Introduction to Statistical Analysis Introduction to Statistical Analysis Changyu Shen Richard A. and Susan F. Smith Center for Outcomes Research in Cardiology Beth Israel Deaconess Medical Center Harvard Medical School Objectives Descriptive

More information

On the covariate-adjusted estimation for an overall treatment difference with data from a randomized comparative clinical trial

On the covariate-adjusted estimation for an overall treatment difference with data from a randomized comparative clinical trial Biostatistics Advance Access published January 30, 2012 Biostatistics (2012), xx, xx, pp. 1 18 doi:10.1093/biostatistics/kxr050 On the covariate-adjusted estimation for an overall treatment difference

More information

Lecture 7 Time-dependent Covariates in Cox Regression

Lecture 7 Time-dependent Covariates in Cox Regression Lecture 7 Time-dependent Covariates in Cox Regression So far, we ve been considering the following Cox PH model: λ(t Z) = λ 0 (t) exp(β Z) = λ 0 (t) exp( β j Z j ) where β j is the parameter for the the

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

3 The Multivariable Fractional Polynomial Approach, with Thoughts about Opportunities and Challenges in Big Data

3 The Multivariable Fractional Polynomial Approach, with Thoughts about Opportunities and Challenges in Big Data Sauerbrei W and Royston P (2017): The Multivariable Fractional Polynomial Approach, with Thoughts about Opportunities and Challenges in Big Data. In: Hans-Joachim Mucha (Ed.)Big data clustering: Data preprocessing,

More information

Extensions of Cox Model for Non-Proportional Hazards Purpose

Extensions of Cox Model for Non-Proportional Hazards Purpose PhUSE Annual Conference 2013 Paper SP07 Extensions of Cox Model for Non-Proportional Hazards Purpose Author: Jadwiga Borucka PAREXEL, Warsaw, Poland Brussels 13 th - 16 th October 2013 Presentation Plan

More information

Penalized Loss functions for Bayesian Model Choice

Penalized Loss functions for Bayesian Model Choice Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented

More information

Individualized Treatment Effects with Censored Data via Nonparametric Accelerated Failure Time Models

Individualized Treatment Effects with Censored Data via Nonparametric Accelerated Failure Time Models Individualized Treatment Effects with Censored Data via Nonparametric Accelerated Failure Time Models Nicholas C. Henderson Thomas A. Louis Gary Rosner Ravi Varadhan Johns Hopkins University July 31, 2018

More information

CHOOSING AMONG GENERALIZED LINEAR MODELS APPLIED TO MEDICAL DATA

CHOOSING AMONG GENERALIZED LINEAR MODELS APPLIED TO MEDICAL DATA STATISTICS IN MEDICINE, VOL. 17, 59 68 (1998) CHOOSING AMONG GENERALIZED LINEAR MODELS APPLIED TO MEDICAL DATA J. K. LINDSEY AND B. JONES* Department of Medical Statistics, School of Computing Sciences,

More information

Reduced-rank hazard regression

Reduced-rank hazard regression Chapter 2 Reduced-rank hazard regression Abstract The Cox proportional hazards model is the most common method to analyze survival data. However, the proportional hazards assumption might not hold. The

More information

Introduction to Statistical modeling: handout for Math 489/583

Introduction to Statistical modeling: handout for Math 489/583 Introduction to Statistical modeling: handout for Math 489/583 Statistical modeling occurs when we are trying to model some data using statistical tools. From the start, we recognize that no model is perfect

More information

Statistical aspects of prediction models with high-dimensional data

Statistical aspects of prediction models with high-dimensional data Statistical aspects of prediction models with high-dimensional data Anne Laure Boulesteix Institut für Medizinische Informationsverarbeitung, Biometrie und Epidemiologie February 15th, 2017 Typeset by

More information

Multivariable model-building with continuous covariates: 2. Comparison between splines and fractional polynomials

Multivariable model-building with continuous covariates: 2. Comparison between splines and fractional polynomials Multivariable model-building with continuous covariates: 2. Comparison between splines and fractional polynomials Harald Binder, Willi Sauerbrei & Patrick Royston Universität Freiburg i. Br. Nr. 106 July

More information

Ph.D. course: Regression models. Introduction. 19 April 2012

Ph.D. course: Regression models. Introduction. 19 April 2012 Ph.D. course: Regression models Introduction PKA & LTS Sect. 1.1, 1.2, 1.4 19 April 2012 www.biostat.ku.dk/~pka/regrmodels12 Per Kragh Andersen 1 Regression models The distribution of one outcome variable

More information

Survival Analysis Math 434 Fall 2011

Survival Analysis Math 434 Fall 2011 Survival Analysis Math 434 Fall 2011 Part IV: Chap. 8,9.2,9.3,11: Semiparametric Proportional Hazards Regression Jimin Ding Math Dept. www.math.wustl.edu/ jmding/math434/fall09/index.html Basic Model Setup

More information

MODELING MISSING COVARIATE DATA AND TEMPORAL FEATURES OF TIME-DEPENDENT COVARIATES IN TREE-STRUCTURED SURVIVAL ANALYSIS

MODELING MISSING COVARIATE DATA AND TEMPORAL FEATURES OF TIME-DEPENDENT COVARIATES IN TREE-STRUCTURED SURVIVAL ANALYSIS MODELING MISSING COVARIATE DATA AND TEMPORAL FEATURES OF TIME-DEPENDENT COVARIATES IN TREE-STRUCTURED SURVIVAL ANALYSIS by Meredith JoAnne Lotz B.A., St. Olaf College, 2004 Submitted to the Graduate Faculty

More information

PhD course: Statistical evaluation of diagnostic and predictive models

PhD course: Statistical evaluation of diagnostic and predictive models PhD course: Statistical evaluation of diagnostic and predictive models Tianxi Cai (Harvard University, Boston) Paul Blanche (University of Copenhagen) Thomas Alexander Gerds (University of Copenhagen)

More information

Obnoxious lateness humor

Obnoxious lateness humor Obnoxious lateness humor 1 Using Bayesian Model Averaging For Addressing Model Uncertainty in Environmental Risk Assessment Louise Ryan and Melissa Whitney Department of Biostatistics Harvard School of

More information

Model comparison. Patrick Breheny. March 28. Introduction Measures of predictive power Model selection

Model comparison. Patrick Breheny. March 28. Introduction Measures of predictive power Model selection Model comparison Patrick Breheny March 28 Patrick Breheny BST 760: Advanced Regression 1/25 Wells in Bangladesh In this lecture and the next, we will consider a data set involving modeling the decisions

More information

Ph.D. course: Regression models. Regression models. Explanatory variables. Example 1.1: Body mass index and vitamin D status

Ph.D. course: Regression models. Regression models. Explanatory variables. Example 1.1: Body mass index and vitamin D status Ph.D. course: Regression models Introduction PKA & LTS Sect. 1.1, 1.2, 1.4 25 April 2013 www.biostat.ku.dk/~pka/regrmodels13 Per Kragh Andersen Regression models The distribution of one outcome variable

More information

Model Selection in GLMs. (should be able to implement frequentist GLM analyses!) Today: standard frequentist methods for model selection

Model Selection in GLMs. (should be able to implement frequentist GLM analyses!) Today: standard frequentist methods for model selection Model Selection in GLMs Last class: estimability/identifiability, analysis of deviance, standard errors & confidence intervals (should be able to implement frequentist GLM analyses!) Today: standard frequentist

More information

BIAS OF MAXIMUM-LIKELIHOOD ESTIMATES IN LOGISTIC AND COX REGRESSION MODELS: A COMPARATIVE SIMULATION STUDY

BIAS OF MAXIMUM-LIKELIHOOD ESTIMATES IN LOGISTIC AND COX REGRESSION MODELS: A COMPARATIVE SIMULATION STUDY BIAS OF MAXIMUM-LIKELIHOOD ESTIMATES IN LOGISTIC AND COX REGRESSION MODELS: A COMPARATIVE SIMULATION STUDY Ingo Langner 1, Ralf Bender 2, Rebecca Lenz-Tönjes 1, Helmut Küchenhoff 2, Maria Blettner 2 1

More information

STAT331. Cox s Proportional Hazards Model

STAT331. Cox s Proportional Hazards Model STAT331 Cox s Proportional Hazards Model In this unit we introduce Cox s proportional hazards (Cox s PH) model, give a heuristic development of the partial likelihood function, and discuss adaptations

More information

Group Sequential Tests for Delayed Responses. Christopher Jennison. Lisa Hampson. Workshop on Special Topics on Sequential Methodology

Group Sequential Tests for Delayed Responses. Christopher Jennison. Lisa Hampson. Workshop on Special Topics on Sequential Methodology Group Sequential Tests for Delayed Responses Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Lisa Hampson Department of Mathematics and Statistics,

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Adapting Prediction Error Estimates for Biased Complexity Selection in High-Dimensional Bootstrap Samples. Harald Binder & Martin Schumacher

Adapting Prediction Error Estimates for Biased Complexity Selection in High-Dimensional Bootstrap Samples. Harald Binder & Martin Schumacher Adapting Prediction Error Estimates for Biased Complexity Selection in High-Dimensional Bootstrap Samples Harald Binder & Martin Schumacher Universität Freiburg i. Br. Nr. 100 December 2007 Zentrum für

More information

Analysis of Time-to-Event Data: Chapter 4 - Parametric regression models

Analysis of Time-to-Event Data: Chapter 4 - Parametric regression models Analysis of Time-to-Event Data: Chapter 4 - Parametric regression models Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/25 Right censored

More information

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation Biost 58 Applied Biostatistics II Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture 5: Review Purpose of Statistics Statistics is about science (Science in the broadest

More information

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Glenn Heller and Jing Qin Department of Epidemiology and Biostatistics Memorial

More information

ECLT 5810 Linear Regression and Logistic Regression for Classification. Prof. Wai Lam

ECLT 5810 Linear Regression and Logistic Regression for Classification. Prof. Wai Lam ECLT 5810 Linear Regression and Logistic Regression for Classification Prof. Wai Lam Linear Regression Models Least Squares Input vectors is an attribute / feature / predictor (independent variable) The

More information

Building a Prognostic Biomarker

Building a Prognostic Biomarker Building a Prognostic Biomarker Noah Simon and Richard Simon July 2016 1 / 44 Prognostic Biomarker for a Continuous Measure On each of n patients measure y i - single continuous outcome (eg. blood pressure,

More information

Experimental Design and Data Analysis for Biologists

Experimental Design and Data Analysis for Biologists Experimental Design and Data Analysis for Biologists Gerry P. Quinn Monash University Michael J. Keough University of Melbourne CAMBRIDGE UNIVERSITY PRESS Contents Preface page xv I I Introduction 1 1.1

More information

Package CoxRidge. February 27, 2015

Package CoxRidge. February 27, 2015 Type Package Title Cox Models with Dynamic Ridge Penalties Version 0.9.2 Date 2015-02-12 Package CoxRidge February 27, 2015 Author Aris Perperoglou Maintainer Aris Perperoglou

More information

Part III Measures of Classification Accuracy for the Prediction of Survival Times

Part III Measures of Classification Accuracy for the Prediction of Survival Times Part III Measures of Classification Accuracy for the Prediction of Survival Times Patrick J Heagerty PhD Department of Biostatistics University of Washington 102 ISCB 2010 Session Three Outline Examples

More information

Dose-response modeling with bivariate binary data under model uncertainty

Dose-response modeling with bivariate binary data under model uncertainty Dose-response modeling with bivariate binary data under model uncertainty Bernhard Klingenberg 1 1 Department of Mathematics and Statistics, Williams College, Williamstown, MA, 01267 and Institute of Statistics,

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Regression II: Regularization and Shrinkage Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Model Combining in Factorial Data Analysis

Model Combining in Factorial Data Analysis Model Combining in Factorial Data Analysis Lihua Chen Department of Mathematics, The University of Toledo Panayotis Giannakouros Department of Economics, University of Missouri Kansas City Yuhong Yang

More information

Proportional hazards regression

Proportional hazards regression Proportional hazards regression Patrick Breheny October 8 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/28 Introduction The model Solving for the MLE Inference Today we will begin discussing regression

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

Bayesian Nonparametric Accelerated Failure Time Models for Analyzing Heterogeneous Treatment Effects

Bayesian Nonparametric Accelerated Failure Time Models for Analyzing Heterogeneous Treatment Effects Bayesian Nonparametric Accelerated Failure Time Models for Analyzing Heterogeneous Treatment Effects Nicholas C. Henderson Thomas A. Louis Gary Rosner Ravi Varadhan Johns Hopkins University September 28,

More information

Algorithm-Independent Learning Issues

Algorithm-Independent Learning Issues Algorithm-Independent Learning Issues Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2007 c 2007, Selim Aksoy Introduction We have seen many learning

More information

Multivariable Fractional Polynomials

Multivariable Fractional Polynomials Multivariable Fractional Polynomials Axel Benner September 7, 2015 Contents 1 Introduction 1 2 Inventory of functions 1 3 Usage in R 2 3.1 Model selection........................................ 3 4 Example

More information

Multi-state Models: An Overview

Multi-state Models: An Overview Multi-state Models: An Overview Andrew Titman Lancaster University 14 April 2016 Overview Introduction to multi-state modelling Examples of applications Continuously observed processes Intermittently observed

More information

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials Two-stage Adaptive Randomization for Delayed Response in Clinical Trials Guosheng Yin Department of Statistics and Actuarial Science The University of Hong Kong Joint work with J. Xu PSI and RSS Journal

More information

ST440/540: Applied Bayesian Statistics. (9) Model selection and goodness-of-fit checks

ST440/540: Applied Bayesian Statistics. (9) Model selection and goodness-of-fit checks (9) Model selection and goodness-of-fit checks Objectives In this module we will study methods for model comparisons and checking for model adequacy For model comparisons there are a finite number of candidate

More information

Biostatistics-Lecture 16 Model Selection. Ruibin Xi Peking University School of Mathematical Sciences

Biostatistics-Lecture 16 Model Selection. Ruibin Xi Peking University School of Mathematical Sciences Biostatistics-Lecture 16 Model Selection Ruibin Xi Peking University School of Mathematical Sciences Motivating example1 Interested in factors related to the life expectancy (50 US states,1969-71 ) Per

More information

Multivariable Fractional Polynomials

Multivariable Fractional Polynomials Multivariable Fractional Polynomials Axel Benner May 17, 2007 Contents 1 Introduction 1 2 Inventory of functions 1 3 Usage in R 2 3.1 Model selection........................................ 3 4 Example

More information

Chapter 4. Parametric Approach. 4.1 Introduction

Chapter 4. Parametric Approach. 4.1 Introduction Chapter 4 Parametric Approach 4.1 Introduction The missing data problem is already a classical problem that has not been yet solved satisfactorily. This problem includes those situations where the dependent

More information

Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Mela. P.

Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Mela. P. Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Melanie M. Wall, Bradley P. Carlin November 24, 2014 Outlines of the talk

More information

Group Sequential Designs: Theory, Computation and Optimisation

Group Sequential Designs: Theory, Computation and Optimisation Group Sequential Designs: Theory, Computation and Optimisation Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj 8th International Conference

More information

Finite Population Correction Methods

Finite Population Correction Methods Finite Population Correction Methods Moses Obiri May 5, 2017 Contents 1 Introduction 1 2 Normal-based Confidence Interval 2 3 Bootstrap Confidence Interval 3 4 Finite Population Bootstrap Sampling 5 4.1

More information

Model Selection. Frank Wood. December 10, 2009

Model Selection. Frank Wood. December 10, 2009 Model Selection Frank Wood December 10, 2009 Standard Linear Regression Recipe Identify the explanatory variables Decide the functional forms in which the explanatory variables can enter the model Decide

More information

Statistics 262: Intermediate Biostatistics Model selection

Statistics 262: Intermediate Biostatistics Model selection Statistics 262: Intermediate Biostatistics Model selection Jonathan Taylor & Kristin Cobb Statistics 262: Intermediate Biostatistics p.1/?? Today s class Model selection. Strategies for model selection.

More information

Practice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes:

Practice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes: Practice Exam 1 1. Losses for an insurance coverage have the following cumulative distribution function: F(0) = 0 F(1,000) = 0.2 F(5,000) = 0.4 F(10,000) = 0.9 F(100,000) = 1 with linear interpolation

More information

Extensions of Cox Model for Non-Proportional Hazards Purpose

Extensions of Cox Model for Non-Proportional Hazards Purpose PhUSE 2013 Paper SP07 Extensions of Cox Model for Non-Proportional Hazards Purpose Jadwiga Borucka, PAREXEL, Warsaw, Poland ABSTRACT Cox proportional hazard model is one of the most common methods used

More information

Univariate shrinkage in the Cox model for high dimensional data

Univariate shrinkage in the Cox model for high dimensional data Univariate shrinkage in the Cox model for high dimensional data Robert Tibshirani January 6, 2009 Abstract We propose a method for prediction in Cox s proportional model, when the number of features (regressors)

More information

How the mean changes depends on the other variable. Plots can show what s happening...

How the mean changes depends on the other variable. Plots can show what s happening... Chapter 8 (continued) Section 8.2: Interaction models An interaction model includes one or several cross-product terms. Example: two predictors Y i = β 0 + β 1 x i1 + β 2 x i2 + β 12 x i1 x i2 + ɛ i. How

More information

Tied survival times; estimation of survival probabilities

Tied survival times; estimation of survival probabilities Tied survival times; estimation of survival probabilities Patrick Breheny November 5 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/22 Introduction Tied survival times Introduction Breslow approximation

More information

MS-C1620 Statistical inference

MS-C1620 Statistical inference MS-C1620 Statistical inference 10 Linear regression III Joni Virta Department of Mathematics and Systems Analysis School of Science Aalto University Academic year 2018 2019 Period III - IV 1 / 32 Contents

More information

A Bayesian Nonparametric Approach to Causal Inference for Semi-competing risks

A Bayesian Nonparametric Approach to Causal Inference for Semi-competing risks A Bayesian Nonparametric Approach to Causal Inference for Semi-competing risks Y. Xu, D. Scharfstein, P. Mueller, M. Daniels Johns Hopkins, Johns Hopkins, UT-Austin, UF JSM 2018, Vancouver 1 What are semi-competing

More information

Definitions and examples Simple estimation and testing Regression models Goodness of fit for the Cox model. Recap of Part 1. Per Kragh Andersen

Definitions and examples Simple estimation and testing Regression models Goodness of fit for the Cox model. Recap of Part 1. Per Kragh Andersen Recap of Part 1 Per Kragh Andersen Section of Biostatistics, University of Copenhagen DSBS Course Survival Analysis in Clinical Trials January 2018 1 / 65 Overview Definitions and examples Simple estimation

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Lecture 3. Hypothesis testing. Goodness of Fit. Model diagnostics GLM (Spring, 2018) Lecture 3 1 / 34 Models Let M(X r ) be a model with design matrix X r (with r columns) r n

More information

Model Selection in Cox regression

Model Selection in Cox regression Model Selection in Cox regression Suppose we have a possibly censored survival outcome that we want to model as a function of a (possibly large) set of covariates. How do we decide which covariates to

More information

Lecture 6 PREDICTING SURVIVAL UNDER THE PH MODEL

Lecture 6 PREDICTING SURVIVAL UNDER THE PH MODEL Lecture 6 PREDICTING SURVIVAL UNDER THE PH MODEL The Cox PH model: λ(t Z) = λ 0 (t) exp(β Z). How do we estimate the survival probability, S z (t) = S(t Z) = P (T > t Z), for an individual with covariates

More information

Analysis of Fast Input Selection: Application in Time Series Prediction

Analysis of Fast Input Selection: Application in Time Series Prediction Analysis of Fast Input Selection: Application in Time Series Prediction Jarkko Tikka, Amaury Lendasse, and Jaakko Hollmén Helsinki University of Technology, Laboratory of Computer and Information Science,

More information

Supporting Information for Estimating restricted mean. treatment effects with stacked survival models

Supporting Information for Estimating restricted mean. treatment effects with stacked survival models Supporting Information for Estimating restricted mean treatment effects with stacked survival models Andrew Wey, David Vock, John Connett, and Kyle Rudser Section 1 presents several extensions to the simulation

More information

Today. HW 1: due February 4, pm. Aspects of Design CD Chapter 2. Continue with Chapter 2 of ELM. In the News:

Today. HW 1: due February 4, pm. Aspects of Design CD Chapter 2. Continue with Chapter 2 of ELM. In the News: Today HW 1: due February 4, 11.59 pm. Aspects of Design CD Chapter 2 Continue with Chapter 2 of ELM In the News: STA 2201: Applied Statistics II January 14, 2015 1/35 Recap: data on proportions data: y

More information

Constrained estimation for binary and survival data

Constrained estimation for binary and survival data Constrained estimation for binary and survival data Jeremy M. G. Taylor Yong Seok Park John D. Kalbfleisch Biostatistics, University of Michigan May, 2010 () Constrained estimation May, 2010 1 / 43 Outline

More information

Müller: Goodness-of-fit criteria for survival data

Müller: Goodness-of-fit criteria for survival data Müller: Goodness-of-fit criteria for survival data Sonderforschungsbereich 386, Paper 382 (2004) Online unter: http://epub.ub.uni-muenchen.de/ Projektpartner Goodness of fit criteria for survival data

More information

Statistical Methods for Alzheimer s Disease Studies

Statistical Methods for Alzheimer s Disease Studies Statistical Methods for Alzheimer s Disease Studies Rebecca A. Betensky, Ph.D. Department of Biostatistics, Harvard T.H. Chan School of Public Health July 19, 2016 1/37 OUTLINE 1 Statistical collaborations

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Matrix Data: Prediction Instructor: Yizhou Sun yzsun@ccs.neu.edu September 14, 2014 Today s Schedule Course Project Introduction Linear Regression Model Decision Tree 2 Methods

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

Random Forests for Ordinal Response Data: Prediction and Variable Selection

Random Forests for Ordinal Response Data: Prediction and Variable Selection Silke Janitza, Gerhard Tutz, Anne-Laure Boulesteix Random Forests for Ordinal Response Data: Prediction and Variable Selection Technical Report Number 174, 2014 Department of Statistics University of Munich

More information

TECHNICAL REPORT # 59 MAY Interim sample size recalculation for linear and logistic regression models: a comprehensive Monte-Carlo study

TECHNICAL REPORT # 59 MAY Interim sample size recalculation for linear and logistic regression models: a comprehensive Monte-Carlo study TECHNICAL REPORT # 59 MAY 2013 Interim sample size recalculation for linear and logistic regression models: a comprehensive Monte-Carlo study Sergey Tarima, Peng He, Tao Wang, Aniko Szabo Division of Biostatistics,

More information

Lecture 10. Diagnostics. Statistics Survival Analysis. Presented March 1, 2016

Lecture 10. Diagnostics. Statistics Survival Analysis. Presented March 1, 2016 Statistics 255 - Survival Analysis Presented March 1, 2016 Dan Gillen Department of Statistics University of California, Irvine 10.1 Are model assumptions correct? Is the proportional hazards assumption

More information

Analysing data: regression and correlation S6 and S7

Analysing data: regression and correlation S6 and S7 Basic medical statistics for clinical and experimental research Analysing data: regression and correlation S6 and S7 K. Jozwiak k.jozwiak@nki.nl 2 / 49 Correlation So far we have looked at the association

More information

Estimation in the l 1 -Regularized Accelerated Failure Time Model

Estimation in the l 1 -Regularized Accelerated Failure Time Model Estimation in the l 1 -Regularized Accelerated Failure Time Model by Brent Johnson, PhD Technical Report 08-01 May 2008 Department of Biostatistics Rollins School of Public Health 1518 Clifton Road, N.E.

More information

Chapter 4 Regression Models

Chapter 4 Regression Models 23.August 2010 Chapter 4 Regression Models The target variable T denotes failure time We let x = (x (1),..., x (m) ) represent a vector of available covariates. Also called regression variables, regressors,

More information

β j = coefficient of x j in the model; β = ( β1, β2,

β j = coefficient of x j in the model; β = ( β1, β2, Regression Modeling of Survival Time Data Why regression models? Groups similar except for the treatment under study use the nonparametric methods discussed earlier. Groups differ in variables (covariates)

More information

Grouped variable selection in high dimensional partially linear additive Cox model

Grouped variable selection in high dimensional partially linear additive Cox model University of Iowa Iowa Research Online Theses and Dissertations Fall 2010 Grouped variable selection in high dimensional partially linear additive Cox model Li Liu University of Iowa Copyright 2010 Li

More information

Robustifying Trial-Derived Treatment Rules to a Target Population

Robustifying Trial-Derived Treatment Rules to a Target Population 1/ 39 Robustifying Trial-Derived Treatment Rules to a Target Population Yingqi Zhao Public Health Sciences Division Fred Hutchinson Cancer Research Center Workshop on Perspectives and Analysis for Personalized

More information

Harvard University. Harvard University Biostatistics Working Paper Series

Harvard University. Harvard University Biostatistics Working Paper Series Harvard University Harvard University Biostatistics Working Paper Series Year 2008 Paper 94 The Highest Confidence Density Region and Its Usage for Inferences about the Survival Function with Censored

More information

Model selection for good estimation or prediction over a user-specified covariate distribution

Model selection for good estimation or prediction over a user-specified covariate distribution Graduate Theses and Dissertations Iowa State University Capstones, Theses and Dissertations 2010 Model selection for good estimation or prediction over a user-specified covariate distribution Adam Lee

More information

Comparison of Predictive Accuracy of Neural Network Methods and Cox Regression for Censored Survival Data

Comparison of Predictive Accuracy of Neural Network Methods and Cox Regression for Censored Survival Data Comparison of Predictive Accuracy of Neural Network Methods and Cox Regression for Censored Survival Data Stanley Azen Ph.D. 1, Annie Xiang Ph.D. 1, Pablo Lapuerta, M.D. 1, Alex Ryutov MS 2, Jonathan Buckley

More information

Multiple imputation in Cox regression when there are time-varying effects of covariates

Multiple imputation in Cox regression when there are time-varying effects of covariates Received: 7 January 26 Revised: 4 May 28 Accepted: 7 May 28 DOI:.2/sim.7842 RESEARCH ARTICLE Multiple imputation in Cox regression when there are time-varying effects of covariates Ruth H. Keogh Tim P.

More information

Problems in model averaging with dummy variables

Problems in model averaging with dummy variables Problems in model averaging with dummy variables David F. Hendry and J. James Reade Economics Department, Oxford University Model Evaluation in Macroeconomics Workshop, University of Oslo 6th May 2005

More information

Prerequisite: STATS 7 or STATS 8 or AP90 or (STATS 120A and STATS 120B and STATS 120C). AP90 with a minimum score of 3

Prerequisite: STATS 7 or STATS 8 or AP90 or (STATS 120A and STATS 120B and STATS 120C). AP90 with a minimum score of 3 University of California, Irvine 2017-2018 1 Statistics (STATS) Courses STATS 5. Seminar in Data Science. 1 Unit. An introduction to the field of Data Science; intended for entering freshman and transfers.

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

A Regression Model For Recurrent Events With Distribution Free Correlation Structure

A Regression Model For Recurrent Events With Distribution Free Correlation Structure A Regression Model For Recurrent Events With Distribution Free Correlation Structure J. Pénichoux(1), A. Latouche(2), T. Moreau(1) (1) INSERM U780 (2) Université de Versailles, EA2506 ISCB - 2009 - Prague

More information

Faculty of Health Sciences. Cox regression. Torben Martinussen. Department of Biostatistics University of Copenhagen. 20. september 2012 Slide 1/51

Faculty of Health Sciences. Cox regression. Torben Martinussen. Department of Biostatistics University of Copenhagen. 20. september 2012 Slide 1/51 Faculty of Health Sciences Cox regression Torben Martinussen Department of Biostatistics University of Copenhagen 2. september 212 Slide 1/51 Survival analysis Standard setup for right-censored survival

More information