Bayesian semiparametric inference for the accelerated failure time model using hierarchical mixture modeling with N-IG priors

Size: px
Start display at page:

Download "Bayesian semiparametric inference for the accelerated failure time model using hierarchical mixture modeling with N-IG priors"

Transcription

1 Bayesian semiparametric inference for the accelerated failure time model using hierarchical mixture modeling with N-IG priors Raffaele Argiento 1, Alessandra Guglielmi 2, Antonio Pievatolo 1, Fabrizio Ruggeri 1 CNR-IMATI, Milano 1 Politecnico di Milano, Milano 2 Abstract We will pursue a Bayesian semiparametric approach for an Accelerated Failure Time regression model, usually considered in survival analysis, when the error distribution is a mixture of parametric densities with a nonparametric mixing measure. The Dirichlet process is a popular choice for the mixing measure, yielding a Dirichlet process mixture model for the error; the paper considers the same model, but here, as an alternative to the Dirichlet process, the mixing measure is equal to a normalized inverse-gaussian prior, built from normalized inverse-gaussian finite dimensional distributions, as recently proposed in the literature. A comparison between the two models will be carried out. Markov chain Monte Carlo techniques will be used to estimate the predictive distribution of the survival time, along with the posterior distribution of the regression parameters. The efficiency of computational methods will also be compared, using both real and simulated data. Keywords: AFT regression models, Bayesian semiparametrics, Mixture models, MCMC algorithms. 1 Introduction In the survival literature, the Accelerated Failure Time AFT) model is usually meant as the multiplicative effect of a fixed p-vector of covariates x = x 1,..., x p ) on the failure time T, i.e. log T = x β + W, where β = β 1,..., β p ) is the vector of regression parameters and W denotes the error. Recently this model has received much attention in the Bayesian community, in particular in papers where the error W or expw)) has been represented hierarchically as a mixture of parametric densities with a Dirichlet process as mixing measure, i.e., the well-known Dirichlet process mixture DPM) models, introduced by Lo 1984). In Kuo and Mallick 1997), the authors model expw) as a Dirichlet process location mixture of normals with variance fixed and small). Kottas and Gelfand 2001) and Gelfand and Kottas 2003) propose a flexible semiparametric class of zero median distributions for the error, which essentially consists of a Dirichlet process scale mixture of split 0-mean normals with skewness handled parametrically). In Ghosh and Goshal 2006) the distribution of expw) is given as a scale mixture of Weibull distributions with a Dirichlet process as mixing measure, and the consistency of the posterior distribution of the regression parameters is established. Pointing out that the marginal prior for expw) in Kuo and Mallick 1997) gives positive probability to the negative reals, Hanson 2006) proposes a DPM model of gamma densities, mixing over both the scale and the shape of the gammas, for the distribution of expw). In the paper we will consider the model T = e x β V, where V := e W is a mixture of some parametric family of densities on the positive real line, mixed by a random distribution function G on R s s is a positive integer). The paper mainly focuses on the performances of two hierarchical mixture models comparing Bayesian inferences on the regression parameters and the survival times. On one hand we will assume that G has a Dirichlet process prior, yielding DPM models for V ; on the other hand, G will have the normalized inverse-gaussian prior N-IG prior), as introduced in Lijoi et al. 2005), thus defining what we call N-IG mixtures for short. The matching between the two priors is achieved centering G at the same distribution function G 0, and assuming equal prior means for the number of components in the mixture. N-IG mixtures of normals have been studied in Lijoi et al. 2006), but, at the best of our knowledge, no papers are available on N-IG mixture with kernel having support on R +, including a regression component. The N-IG prior preserves the same tractability as the Dirichlet process prior, and, furthermore, it is characterized by a more elaborate clustering property: the weight given to a single observation depends on the whole number of ties in the sample. Moreover, the prior distribution of the number of components for the N-IG mixture is more spread out compared to the same prior under the DPM model. Following Hanson 2006), we will consider hierarchical mixtures of gamma densities, mixed on both the scale and the shape parameters. The centering distribution G 0 we assume yields an infinite mean marginal prior for V ; however prior information can be expressed by means of the median of V. Posterior inference on β and PT > t) is carried out via Gibbs sampling, incorporating censoring when necessary. Of course, density estimation can be performed within this model ignoring the regression aspect i.e. simply assuming a null vector of covariates). The two competing models will be tested on real and simulated data. We will use the same notation to denote distribution functions and the corresponding probability measures. 2 The model Let T 1,...,T n be the survival times of n subjects, and let x i = x i1,..., x ip ) be the covariate vector associated with observed or censored) t i, i = 1,...,n. The model we consider can be hierarchically expressed as follows

2 G q, T i = e x i β V i, i = 1,...,n, V i θ i ind k ; θ i ), θ i G iid G, 1) G 0 A) := E q GA)), A BΘ) β G, β πβ), where k ; θ i ) is a family of densities on R +, depending on a vector of parameters θ i belonging to a Borel subset Θ of R s, and q is the prior distribution on the random d.f. G; G 0 is a d.f. on Θ, expressing the mean of G. Here we will assume that G is a Dirichlet process, or an N-IG process; in both cases, the prior is completely known when an additional positive parameter is given. Indeed, we will write G DPaG 0 )) to denote the Dirichlet prior with parameter ag 0 ), and G N-IGM, G 0 )) to denote the N-IG prior; see Lijoi et al. 2005) for notation and definition. Recall that, for both the Dirichlet process and N-IG priors, EGA)) = G 0 A), whereas VarGA)) = G 0 A)1 G 0 A))M 2 e M Γ 2, M) for the N-IG prior Γ, ) denotes the incomplete gamma function), and, on the other hand, VarGA)) = G 0 A)1 G 0 A)))/a + 1) under the Dirichlet prior. The Bayesian model specification is completed assuming that G 0 depends on s hyperparameters γ 1,..., γ s possibly random and distributed according to πγ 1,..., γ s )). Observe that, if G is an N-IG process with diffuse mean G 0, then the predictive distributions of θ n+1, given θ 1,...,θ n, are Pθ n+1 A θ 1,..., θ n ) = 2) k w 0,n G 0 A) + w 1,n n j 1 2 )δ θj A), for any A BΘ), where θ 1,...,θ k n 1,...,n k, j n j = n) denote the k distinct observations multiplicities) within θ 1,..., θ n and w 0,n = w 1,n = P n r=0 n r) M 2 ) r+1 Γk+1+2r 2n;M) 2n P n 1 r=0 n 1 r ) M 2 ) r Γk+2+2r 2n;M) P n r=0 n r) M 2 ) r+1 Γk+2r 2n;M) n P n 1 r=0 n 1 r ) M 2 ) r Γk+2+2r 2n;M). Moreover, the distribution of the number of distinct observations k in a sample of size n is p N-IG k n) = 2n k 1 n 1 r=0 n 1 r 3) ) e M M 2 ) n 1 n 1 2 2n k 1 Γk) 4) ) Γk+2+2r 2n;M) M 2 ). r It should be pointed out that, for n fixed, the corresponding value of the mean has a lower bound, given by the limit as M goes to 0. As mentioned in the Introduction, we are going to consider hierarchical mixtures of gamma densities k ; θ), θ = ϑ 1, ϑ 2 ), with mean ϑ 1 /ϑ 2, and to choose the centering distribution G 0 on R + R + as the product of two exponential distributions, i.e. ϑ 1 and ϑ 2, under G 0, are independent, exponentially distributed with parameters γ 1 and, respectively. Assuming independent exponential distributions for G 0 is computationally convenient, since, in this case, a closedform expression for the marginal prior of V is available, which is needed in the implementation of the Gibbs sampling procedure a conjugate hierarchical mixture model is obtained). Indeed, for model 1), under both Dirichlet process and N-IG priors, the marginal prior density of V is, for v > 0, f V v) = Θ kv; θ)g 0 dθ) 5) γ 1 = vv + )γ 1 + log v+γ2 v )), 2 with distribution function F V v) = γ 1 / γ 1 + log1 + γ2 v ), ) v > 0. This distribution has infinite mean, but information about the hyperparameters γ 1 and can be derived fixing the median m of V, thus resulting in a semiparametric m-median regression model, analogously as in Kottas and Gelfand 2001). Therefore, in the paper, we will assume γ 1 = log1 + /m). On the other hand, the hyperparameter controls the dispersion of V : the interquartile range of the marginal prior, as a function of m and, is 1 + γ2 /m) 1/3 1 ) γ2 /m) 3 1 ) 1 ), which, for fixed m, increases with increasing. Observe that the 90% prior probability interval for V is [ 1 + ) 19, m ) 1/19 ]. m 1 However, we point out that this choice for G 0 is not extremely flexible, since f V is decreasing with an asymptote at 0 for any γ 1,, in such a way that it always puts an appreciable amount of mass near zero even if γ 1 log1 + /m)). A flat prior distribution is often imposed upon the β regression parameters Kuo and Mallick 1997), Hanson 2006)), regardless of the form of the kernel density k ; θ). We introduce instead the reparametrization α j = e βj, j = 1,...,p, and assign independent gamma priors to the α j s. In this way, with gamma kernel densities, the full conditional posterior distributions of the α j s associated to binary covariates are still gamma. 3 The algorithm Bayesian inference is carried out through MCMC simulation. The algorithm for the AFT model is built upon one for the hierarchical mixture model, by adding a module for the updating of the regression parameters. The desired output of the algorithm for the hierarchical mixture model is an estimate of the predictive density of a new observation V n+1, given the data v 1,...,v n. This can be obtained as the ergodic average of the conditional densities of V n+1, given θ j) = θ j) 1,..., θj) n ) and the data, where {θ j) } j 1 is the output of a Markov chain having fθ data) as its equilibrium distribution: ˆf Vn+1 v data) = 1 J J f Vn+1 v data, θ j) ). 6)

3 The density f Vn+1 v data, θ) can be found by conditioning on G first, and then integrating it out with respect to its distribution given θ and the data, which leaves us with the following closed form expressions: f Vn+1 v data, θ) = 1 [ f V v) + a + n for the DPM model, and n i=1 ] kv; θ i ) f Vn+1 v data, θ) = 8) k w 0,n f V v) + w 1,n n j 1 ) kv; θj 2 ) for the N-IG mixture, f V v) in 7) and 8) as given in 5). To produce the Markov sequence {θ j) } j 1 with fθ data) as its stationary distribution, we use a Polya urn Gibbs sampler scheme such as Algorithm 2 in Neal 2000). The algorithm updates in turn and componentwise a vector of cluster indexes c 1,...,c n ), where c i = c j if θ i = θ j, and the vector of distinct values θ = θ1,..., θ k ). The cluster indexes are linked to the distinct values of the parameters by letting c i = j if θ i = θj. The algorithm for the N-IG mixture model and for the DPM model are identical, except for the weights used for choosing the urn. For the moment, let us denote w i,n by w i,n k), in order to stress the dependence on k. Let us also denote by n i) j the multiplicity of θj after removing θ i from θ, which produces a θ i), by c i) the corresponding vector of indexes and by k i) the dimension of θ i). Then the full conditional of c i is Pc i = j c i), θ i), v i ) ) w 1,n 1 k i) ) n i) j 1 2 kv i θ i) j ) =, w 0,n 1 k i) )a i + w 1,n 1 k i) )b i if j k i), and Pc i = k i) + 1 c i), θ, v i ) = where a i := f V v i ) and b i := k i) w 0,n 1 k i) )a i w 0,n 1 k i) )a i + w 1,n 1 k i) )b i, n i) j 1 ) kv i ; θ i) j ). 2 When c i takes on a new value k i) + 1, a corresponding new value θ i must be sampled from the posterior distribution proportional to kv i ; θ i )g 0 θ i ), with the same method given by Hanson 2006) as Result 2, thus enlarging θ. The full conditional of θj is simply proportional to g 0 θ j ) i:c i=j kv i ; θ j ), 7) where g 0 is the density function associated to G 0. This distribution is easy to sample from when n j = 1 by the abovementioned Result 2, whereas it requires a Metropolis step otherwise. The extension of the Gibbs sampler to the AFT model is obtained by letting p v i = e x i β t i = α xij j t i with the reparametrization introduced in the previous Section) and by updating the indexes and θ as just described. The α j parameters corresponding to binary covariates have a full conditional gamma distribution, whereas a Metropolis step is required for the remaining regression parameters. The actual state of the Markov chain is thus c, θ, α). In the survival analysis context, the predictive distribution of an individual with covariate value x is usually presented as the predictive survival function. Therefore, with data t 1,...,t n, instead of 6) one calculates Ŝ Tn+1 t x, data) = 1 J J S Vn+1 e x β j) t data, θ j) ). The survival function S Vn+1 data, θ) can be easily found by integrating 7) or 8). When censored survival times are present, one can proceed by augmenting the unknown parameter space with the unobserved survival times, and enlarging θ and c consequently. The unobserved survival times are sampled one at a time from their full conditional distributions truncated at their respective censoring points. 4 Data illustrations 4.1 Simulated data for density estimation We consider simulated data from a mixture of 3 gamma densities to perform density estimation. We generated a random sample of size 100 from the density 0.2 gamma40, 20)+0.6 gamma6, 1)+0.2 gamma200, 20), the mean is 6 and the variance is 10.12) and computed the posterior density estimates from the Dirichlet and the N-IG mixtures of gammas the no-covariate AFT model). We assumed M = 0.01 and M = 5.39, which correspond to prior means of the number of components under the N-IG prior equal to and 30, respectively. The matching with the Dirichlet process prior is achieved when a = 3.10 and a = 14.16, respectively. We set the median m of V equal to 5.67 i.e. the true median), 56.7 and 0.57, and the hyperparameter in G 0 equal to 0.01, 1 and 10, corresponding to the values of the IQRs and the 90% prior probability intervals listed in Table 1. For each choice of the hyperparameters we ran several MCMC chains and observed that they stabilize after around 2000 iterations. Then we ran the chains for 50000

4 a) b) c) d) Figure 1: Histogram from the simulated data and density estimates under the Dirichlet process prior dotted red) and the N-IG prior dashed blue) when = Panel a): a = 3.10, M = 0.01, m = 5.67, Panel b): a = 14.16, M = 5.39, m = 5.67, Panel c): a = 3.10, M = 0.01, m = 56.7, Panel d): a = 14.16, M = 5.39, m = In each graph the solid green) line denotes the true density. Table 1: IQR and 90% probability interval for the marginal prior of V m [0.03,10.86] [0.29,107.82] [2.98, ] [4 10 9,18.20] [0.05,116.48] [2.54, ] [ ,60.08] [4 10 8,181.95] [0.48, ] Table 2: Errors in the uniform metric for the simulated dataset of size 100 between the true and estimated distribution functions. m M a iterations keeping the values every 50 iterations to reduce autocorrelation. As a measure of the performances of the models, we computed the error, in the uniform metric EUM), between the true distribution function and the predictive distributions under both priors). Figure 1 displays the true density, the histogram from the simulated data, and the density estimates under the Dirichlet and N-IG process priors for some values of the hyperparameters; in each graph the two estimates are indistinguishable. Table 2 presents the errors for the values of the hyperparameters as in Figure 1, confirming the impression gained from the graphs. To quantify the difference in estimating the true density for repeated samples and at a smaller sample size, we computed the mean error between the exact d.f. and the estimates. If ˆFT T 1,..., T n ) denotes the Bayesian estimate of the true F T ), based on a data sample T 1,..., T n ), and EUMT 1,..., T n ) := sup t ˆF T t T 1,..., T n ) F T t), we computed E T1,...,T n EUMT 1,...,T n )) generating 200 samples of size n = 30. The mean errors and the corresponding standard deviations) of the estimates for some choices of the hyperparameters are presented in Table 3. The values obtained seem to confirm a substantial equivalence of the two nonparametric priors in density estimation. However, according to Figure 2, the posterior distribution of the number of the clusters under the N-IG model corresponding to the prior 4)) is more robust to the choice of the hyperparameters. Its modal values are indeed closer than those under the DPM model to the number of clusters in the true density, and more so when the prior expected number of clusters is farther from the true value panels b) and d)).

5 Table 3: Mean errors and corresponding standard deviations in brackets) between the estimates and the true distribution for samples of size 30 from Example 1. The error is the distance, in the uniform metric, between distribution functions. EK n ) = 6.2 EK n ) = 14.5 γ N-IG DIR N-IG DIR ) ) ) ) m = ) ) ) ) ) ) ) ) ) ) ) ) m = ) ) ) ) ) ) ) ) ) ) ) ) m = ) ) ) ) ) ) ) ) Ek n ) = m = a) Ek n ) = 30 m = b) Ek n ) = m = c) d) Ek n ) = 30 m = Figure 2: Posterior distributions of the number of the clusters under the N-IG mixtures solid blue) and the DPM model dotted red) when = 0.01 for the simulated dataset of size 100.

6 patient 1,36) EK n) = 12.4 patient 0,36) EK n) = patient 1,36) EK n) = 41.9 patient 0,36) EK n) = Figure 3: Estimated survival functions under the Dirichlet process prior dotted red) and the N-IG prior dashed blue) for 2 patients covariate 1, 36) in the left column and covariate 0, 36) in the right column) from Example 2 when m = and = 1. We set a = 3.271, M = in the first row and a = , M = 10 in the second row. 4.2 Dataset involving censoring As a second example, we consider survival times in thousands of days of small-cell lung cancer data patients with right censoring from Ying, Jung and Wei 1995), and studied also in Walker and Mallick 1999), Yang 1999), Kottas and Gelfand 2001), Hanson 2006). The standard therapy is to use a combination of etoposide E) and cisplatin P); however the optimal sequencing and administration schedule was not defined. The data consist of n = 121 survival times in days of patients with limited-stage small-cell lung cancer who were randomly assigned to two different regimens Treatment A: P followed by E, administered to 62 patients, and Treatment B: E followed by P, administered to 59 patients); moreover 23 patients were administratively right-censored. In this case the covariates are x i = x i1, x i,2 ), with x i1 = 0 if patient i was assigned to Treatment A, and x i,2 denoting the patient s entry age. In Figure 3 the estimated survival distributions, under the two semiparametric Bayesian models considered here, for 2 patients, are shown, for some values of the hyperparameters. Observe that, for high values of M and a i.e. strong confidence in the prior), both estimates of the survival function show a bump near the origin. This is imputable to the choice of the marginal prior of V, presenting a heavy asymptote in zero for each value of the hyperparameters. The Bayesian estimates of α 1 and α 2, together with the 90% credible intervals - for m = and different values of - are presented in Table 4; the two models considered provide very similar results, that actually agree with Bayesian estimates provided by the literature. We ran the MCMC chains for iterations, with a thinning of 100, and a burn-in period of To establish a predictive fit measure for the model considered, we adopted a cross-validation approach via the predictive densities of each non-censored survival time, given all the others; see Gelfand, Dey, Chang 1992). If ỹ j, j = 1,..., n, is the estimated from the Gibbs sampling procedure) median of the predictive distribution of T j T j), and s j is the estimated predictive median of T j ỹ j, given T j), then t j ỹ j represents a standardized residual. In this case, a predictive fit index can be defined s j as I := t j ỹ j, s j j S where S denotes the index set of non-censored data in the sample. For m = 2.436, = 1 and EK n ) = 12.41, the indexes of predictive fit are I N-IG = and I DIR = 99.99, indicating a slightly) better fit for the N-IG mixture; see Figure 4. 5 Comments From a predictive point of view, the illustrative examples show that there is not a substantial difference between the DPM and N-IG process priors in mixture density estimation. However,

7 Table 4: Estimates of α 1 and α 2, with 90% probability credible intervals, for the small-cell lung cancer dataset under the N-IG mixture and DPM priors, when m = N-IG M=0.001 EK n ) = 12.41) M=1 EK n ) = 19.31) M=10 EK n ) = 41.93) ˆα 1 = 1.149; 1.181; 1.819) ˆα 1 = 1.421; 1.124; 1.762) ˆα 1 = 1.409; 1.115; 1.782) ˆα 2 = 1.011; 1.000; 1.023) ˆα 2 = 1.017; 1.006; 1.028) ˆα 2 = 1.018; 1.007; 1.028) ˆα 1 = 1.533; 1.203; 1.903) ˆα 1 = 1.529; 1.201; 1.895) ˆα 1 = 1.522; 1.181; 1.926) ˆα 2 = 1.015; 1.005; 1.025) ˆα 2 = 1.015; 1.005; 1.025) ˆα 2 = 1.015; 1.006; 1.024) ˆα 1 = 1.521; 1.022; 1.042) ˆα 1 = 1.531; 1.168; 1.941) ˆα 1 = 1.516; 1.158; 1.967) ˆα 2 = 1.032; 1.022; 1.042) ˆα 2 = 1.031; 1.021; 1.041) ˆα 2 = 1.029; 1.021; 1.039) DIR a=3.271 EK n ) = 12.41) a= EK n ) = 19.31) a= EK n ) = 41.93) ˆα 1 = 1.437; 1.149; 1.772) ˆα 1 = 1.409; 1.123; 1.753) ˆα 1 = 1.387; 1.096; 1.761) ˆα 2 = 1.014; 1.003; 1.025) ˆα 2 = 1.016; 1.005; 1.028) ˆα 2 = 1.021; 1.013; 1.029) ˆα 1 = 1.515; 1.174; 1.887) ˆα 1 = 1.512; 1.192; 1.869) ˆα 1 = 1.530; 1.156; 1.984) ˆα 2 = 1.016; 1.007; 1.025) ˆα 2 = 1.018; 1.009; 1.027) ˆα 2 = 1.020; 1.012; 1.028) ˆα 1 = 1.518; 1.150; 1.947) ˆα 1 = 1.521; 1.128; 1.943) ˆα 1 = 1.537; 1.123; 2.049) ˆα 2 = 1.032; 1.022; 1.043) ˆα 2 = 1.032; 1.023; 1.041) ˆα 2 = 1.033; 1.025; 1.042) I NIG = I DIR = RESIDUALS x 2 Figure 4: Standardized residuals against the continuous covariate for the small-cell lung cancer dataset under the N-IG mixture blue) and DPM red) priors, when m =

8 the N-IG process prior seems to be more effective in the estimation of the number of components in the mixture. In the regression example, the model with the N-IG process prior error fits the data a little better. However, a similar experiment on the Feigl and Zelen 1965) leukemia dataset see Argiento, 2007) gives the opposite result I N-IG = and I DIR = ). In both cases we notice that the difference in the predictive fit indexes is not dramatic, and it is difficult to choose between the two process priors based only on this. This uncertainty remains even if we consider that the result on the Feigl and Zelen dataset is influenced by a few abnormal observations, whose removal produces essentially equal predictive fit indexes. Although the N-IG process prior seems to be better than the DPM prior in finding clustering structures in the data, the computation of the weights 3) requires multiple precision arithmetics, because of the presence of the sum of several incomplete gamma functions. Therefore we did all the computations and the MCMC simulations using R R Development Core Team, 2006), but we used Maple for setting up a table with the necessary weights which do not change during simulations). An alternative to Maple is the PARI C library The PARI Group, 2006), which can be used both at initialization and at run time, because C subroutines can be loaded into R. Given the availability of these multiple precision computational tools, the calculation of the weights is not a serious concern. Kottas, A. and Gelfand, A. E. 2001), Bayesian semiparametric median regression modeling, Journal of the American Statistical Association, 96, Kuo, L. and Mallick, B. 1997), Bayesian semiparametric inference for the accelerated failure-time model, The Canadian Journal of Statistics, 25, Lijoi, A., Mena, R. H. and Pruenster, I. 2005), Hierarchical mixture modeling with normalized inverse-gaussian priors, Journal of the American Statistical Association, 100, Lijoi, A., Mena, R. H. and Pruenster, I. 2006), Controlling the reinforcement in Bayesian nonparametric mixture models, under review. Lo, A.Y. 1984), On a class of Bayesian nonparametric estimates: I. Density estimates, The Annals of Statistics, 12, Neal, R. M. 2000), Markov Chain Sampling Methods for Dirichlet Process Mixture Models, Journal of Computational and Graphical Statistics, 100, R Development Core Team 2006), R: A Language and Environment for Statistical Computing, R Foundation for Statistical Computing, Vienna, Austria The PARI Group 2006), PARI/GP, version 2.3.1, Bordeaux Yang, S. 1999). Censored median regression using weighted empirical survival and hazard functions, Journal of the American Statistical Association, 94, Ying, Z., Jung, S. H. and Wei, L. J. 1995). Survival analysis with median regression models, Journal of the American Statistical Association, 90, Walker, S. and Mallick, B.K. 1999). A Bayesian semiparametric accelerated failure time model, Biometrics, 55, Acknowledgments The authors are indebted to Ramses Mena, who provided some code and precious suggestions. Moreover, the authors would like to thank Antonio Lijoi and Igor Pruenster for helpful discussions. References Argiento, A. 2007), Bayesian semiparametric inference for the accelerated failure time model, PhD Thesis, Bocconi University. Cook, R.D. and Weisberg S. 1982), Residuals and influence in regression, Chapman & Hall, London. Feigl, P. and Zelen, M. 1965), Estimation of exponential survival probabilities with concomitant information, Biometrics, 21, Gelfand, A. E., Dey, D. K. and Chang, H. 1992), Model determination using predictive distributions, with implementation via sampling-based methods Disc: p ), Bayesian Statistics 4. Proceedings of the Fourth Valencia International Meeting, Gelfand, A. E. and Kottas, A. 2003), Bayesian semiparametric regression for median residual life, Scand. J. Statist., 30, Ghosh, S.K. and Ghosal, S. 2006), Semiparametric accelerated failure time models for censored data, Bayesian Statistics and its Applications S. K. Upadhyay et al., eds.), Anamaya Publishers, New Delhi, Hanson, T.E. 2006), Modeling censored lifetime data using a mixture of gammas baseline, Bayesian Analysis, 1, Kalbfleisch, J. D. and Prentice, R.L. 2002), The statistical analysis of failure time data, Second Edition, Wiley, Hoboken NJ). Kottas, A. 2006), Nonparametric Bayesian survival analysis using mixtures of Weibull distributions, J. Statist. Plann. Inference, 136,

Ronald Christensen. University of New Mexico. Albuquerque, New Mexico. Wesley Johnson. University of California, Irvine. Irvine, California

Ronald Christensen. University of New Mexico. Albuquerque, New Mexico. Wesley Johnson. University of California, Irvine. Irvine, California Texts in Statistical Science Bayesian Ideas and Data Analysis An Introduction for Scientists and Statisticians Ronald Christensen University of New Mexico Albuquerque, New Mexico Wesley Johnson University

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Jeffrey N. Rouder Francis Tuerlinckx Paul L. Speckman Jun Lu & Pablo Gomez May 4 008 1 The Weibull regression model

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features Yangxin Huang Department of Epidemiology and Biostatistics, COPH, USF, Tampa, FL yhuang@health.usf.edu January

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Chapter 2. Data Analysis

Chapter 2. Data Analysis Chapter 2 Data Analysis 2.1. Density Estimation and Survival Analysis The most straightforward application of BNP priors for statistical inference is in density estimation problems. Consider the generic

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

STAT Advanced Bayesian Inference

STAT Advanced Bayesian Inference 1 / 32 STAT 625 - Advanced Bayesian Inference Meng Li Department of Statistics Jan 23, 218 The Dirichlet distribution 2 / 32 θ Dirichlet(a 1,...,a k ) with density p(θ 1,θ 2,...,θ k ) = k j=1 Γ(a j) Γ(

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

Marginal Specifications and a Gaussian Copula Estimation

Marginal Specifications and a Gaussian Copula Estimation Marginal Specifications and a Gaussian Copula Estimation Kazim Azam Abstract Multivariate analysis involving random variables of different type like count, continuous or mixture of both is frequently required

More information

Multistate Modeling and Applications

Multistate Modeling and Applications Multistate Modeling and Applications Yang Yang Department of Statistics University of Michigan, Ann Arbor IBM Research Graduate Student Workshop: Statistics for a Smarter Planet Yang Yang (UM, Ann Arbor)

More information

David B. Dahl. Department of Statistics, and Department of Biostatistics & Medical Informatics University of Wisconsin Madison

David B. Dahl. Department of Statistics, and Department of Biostatistics & Medical Informatics University of Wisconsin Madison AN IMPROVED MERGE-SPLIT SAMPLER FOR CONJUGATE DIRICHLET PROCESS MIXTURE MODELS David B. Dahl dbdahl@stat.wisc.edu Department of Statistics, and Department of Biostatistics & Medical Informatics University

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

eqr094: Hierarchical MCMC for Bayesian System Reliability

eqr094: Hierarchical MCMC for Bayesian System Reliability eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our

More information

Practical Bayesian Quantile Regression. Keming Yu University of Plymouth, UK

Practical Bayesian Quantile Regression. Keming Yu University of Plymouth, UK Practical Bayesian Quantile Regression Keming Yu University of Plymouth, UK (kyu@plymouth.ac.uk) A brief summary of some recent work of us (Keming Yu, Rana Moyeed and Julian Stander). Summary We develops

More information

Nonparametric Bayesian Survival Analysis using Mixtures of Weibull Distributions

Nonparametric Bayesian Survival Analysis using Mixtures of Weibull Distributions Nonparametric Bayesian Survival Analysis using Mixtures of Weibull Distributions ATHANASIOS KOTTAS University of California, Santa Cruz ABSTRACT. Bayesian nonparametric methods have been applied to survival

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Incorporating unobserved heterogeneity in Weibull survival models: A Bayesian approach

Incorporating unobserved heterogeneity in Weibull survival models: A Bayesian approach Incorporating unobserved heterogeneity in Weibull survival models: A Bayesian approach Catalina A. Vallejos 1 Mark F.J. Steel 2 1 MRC Biostatistics Unit, EMBL-European Bioinformatics Institute. 2 Dept.

More information

Nonparametric Bayesian Methods - Lecture I

Nonparametric Bayesian Methods - Lecture I Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Rank Regression with Normal Residuals using the Gibbs Sampler

Rank Regression with Normal Residuals using the Gibbs Sampler Rank Regression with Normal Residuals using the Gibbs Sampler Stephen P Smith email: hucklebird@aol.com, 2018 Abstract Yu (2000) described the use of the Gibbs sampler to estimate regression parameters

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Quantile Regression for Residual Life and Empirical Likelihood

Quantile Regression for Residual Life and Empirical Likelihood Quantile Regression for Residual Life and Empirical Likelihood Mai Zhou email: mai@ms.uky.edu Department of Statistics, University of Kentucky, Lexington, KY 40506-0027, USA Jong-Hyeon Jeong email: jeong@nsabp.pitt.edu

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate

More information

Time-Sensitive Dirichlet Process Mixture Models

Time-Sensitive Dirichlet Process Mixture Models Time-Sensitive Dirichlet Process Mixture Models Xiaojin Zhu Zoubin Ghahramani John Lafferty May 25 CMU-CALD-5-4 School of Computer Science Carnegie Mellon University Pittsburgh, PA 523 Abstract We introduce

More information

A DDP Model for Survival Regression

A DDP Model for Survival Regression A DDP Model for Survival Regression Maria De Iorio, 1, Wesley O. Johnson, 2 Peter Müller 3 and Gary L. Rosner 3 1 Department of Epidemiology and Public Health, Imperial College London W2 1PG, U.K. 2 Department

More information

Nonparametric Bayesian Inference for Mean Residual. Life Functions in Survival Analysis

Nonparametric Bayesian Inference for Mean Residual. Life Functions in Survival Analysis Nonparametric Bayesian Inference for Mean Residual Life Functions in Survival Analysis VALERIE POYNOR 1 and ATHANASIOS KOTTAS 2 1 California State University, Fullerton 2 University of California, Santa

More information

Markov Chain Monte Carlo in Practice

Markov Chain Monte Carlo in Practice Markov Chain Monte Carlo in Practice Edited by W.R. Gilks Medical Research Council Biostatistics Unit Cambridge UK S. Richardson French National Institute for Health and Medical Research Vilejuif France

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 4 Problem: Density Estimation We have observed data, y 1,..., y n, drawn independently from some unknown

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Statistical Inference for Stochastic Epidemic Models

Statistical Inference for Stochastic Epidemic Models Statistical Inference for Stochastic Epidemic Models George Streftaris 1 and Gavin J. Gibson 1 1 Department of Actuarial Mathematics & Statistics, Heriot-Watt University, Riccarton, Edinburgh EH14 4AS,

More information

Bayesian semiparametric modeling and inference with mixtures of symmetric distributions

Bayesian semiparametric modeling and inference with mixtures of symmetric distributions Bayesian semiparametric modeling and inference with mixtures of symmetric distributions Athanasios Kottas 1 and Gilbert W. Fellingham 2 1 Department of Applied Mathematics and Statistics, University of

More information

Flexible Regression Modeling using Bayesian Nonparametric Mixtures

Flexible Regression Modeling using Bayesian Nonparametric Mixtures Flexible Regression Modeling using Bayesian Nonparametric Mixtures Athanasios Kottas Department of Applied Mathematics and Statistics University of California, Santa Cruz Department of Statistics Brigham

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units

Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units Sahar Z Zangeneh Robert W. Keener Roderick J.A. Little Abstract In Probability proportional

More information

An Alternative Infinite Mixture Of Gaussian Process Experts

An Alternative Infinite Mixture Of Gaussian Process Experts An Alternative Infinite Mixture Of Gaussian Process Experts Edward Meeds and Simon Osindero Department of Computer Science University of Toronto Toronto, M5S 3G4 {ewm,osindero}@cs.toronto.edu Abstract

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

Bayesian Additive Regression Tree (BART) with application to controlled trail data analysis

Bayesian Additive Regression Tree (BART) with application to controlled trail data analysis Bayesian Additive Regression Tree (BART) with application to controlled trail data analysis Weilan Yang wyang@stat.wisc.edu May. 2015 1 / 20 Background CATE i = E(Y i (Z 1 ) Y i (Z 0 ) X i ) 2 / 20 Background

More information

Slice Sampling Mixture Models

Slice Sampling Mixture Models Slice Sampling Mixture Models Maria Kalli, Jim E. Griffin & Stephen G. Walker Centre for Health Services Studies, University of Kent Institute of Mathematics, Statistics & Actuarial Science, University

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

Petr Volf. Model for Difference of Two Series of Poisson-like Count Data

Petr Volf. Model for Difference of Two Series of Poisson-like Count Data Petr Volf Institute of Information Theory and Automation Academy of Sciences of the Czech Republic Pod vodárenskou věží 4, 182 8 Praha 8 e-mail: volf@utia.cas.cz Model for Difference of Two Series of Poisson-like

More information

Prerequisite: STATS 7 or STATS 8 or AP90 or (STATS 120A and STATS 120B and STATS 120C). AP90 with a minimum score of 3

Prerequisite: STATS 7 or STATS 8 or AP90 or (STATS 120A and STATS 120B and STATS 120C). AP90 with a minimum score of 3 University of California, Irvine 2017-2018 1 Statistics (STATS) Courses STATS 5. Seminar in Data Science. 1 Unit. An introduction to the field of Data Science; intended for entering freshman and transfers.

More information

Bayesian Nonparametric Accelerated Failure Time Models for Analyzing Heterogeneous Treatment Effects

Bayesian Nonparametric Accelerated Failure Time Models for Analyzing Heterogeneous Treatment Effects Bayesian Nonparametric Accelerated Failure Time Models for Analyzing Heterogeneous Treatment Effects Nicholas C. Henderson Thomas A. Louis Gary Rosner Ravi Varadhan Johns Hopkins University September 28,

More information

Bagging During Markov Chain Monte Carlo for Smoother Predictions

Bagging During Markov Chain Monte Carlo for Smoother Predictions Bagging During Markov Chain Monte Carlo for Smoother Predictions Herbert K. H. Lee University of California, Santa Cruz Abstract: Making good predictions from noisy data is a challenging problem. Methods

More information

The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for Bayesian Estimation in a Finite Gaussian Mixture Model

The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for Bayesian Estimation in a Finite Gaussian Mixture Model Thai Journal of Mathematics : 45 58 Special Issue: Annual Meeting in Mathematics 207 http://thaijmath.in.cmu.ac.th ISSN 686-0209 The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

Bayesian semiparametric modeling for stochastic precedence, with applications in epidemiology and survival analysis

Bayesian semiparametric modeling for stochastic precedence, with applications in epidemiology and survival analysis Bayesian semiparametric modeling for stochastic precedence, with applications in epidemiology and survival analysis ATHANASIOS KOTTAS Department of Applied Mathematics and Statistics, University of California,

More information

Bayesian Nonparametric Autoregressive Models via Latent Variable Representation

Bayesian Nonparametric Autoregressive Models via Latent Variable Representation Bayesian Nonparametric Autoregressive Models via Latent Variable Representation Maria De Iorio Yale-NUS College Dept of Statistical Science, University College London Collaborators: Lifeng Ye (UCL, London,

More information

Integrated Non-Factorized Variational Inference

Integrated Non-Factorized Variational Inference Integrated Non-Factorized Variational Inference Shaobo Han, Xuejun Liao and Lawrence Carin Duke University February 27, 2014 S. Han et al. Integrated Non-Factorized Variational Inference February 27, 2014

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Mela. P.

Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Mela. P. Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Melanie M. Wall, Bradley P. Carlin November 24, 2014 Outlines of the talk

More information

Bayesian Nonparametrics

Bayesian Nonparametrics Bayesian Nonparametrics Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent

More information

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline

More information

Abstract INTRODUCTION

Abstract INTRODUCTION Nonparametric empirical Bayes for the Dirichlet process mixture model Jon D. McAuliffe David M. Blei Michael I. Jordan October 4, 2004 Abstract The Dirichlet process prior allows flexible nonparametric

More information

Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material

Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material Quentin Frederik Gronau 1, Monique Duizer 1, Marjan Bakker

More information

A Bayesian Approach to Phylogenetics

A Bayesian Approach to Phylogenetics A Bayesian Approach to Phylogenetics Niklas Wahlberg Based largely on slides by Paul Lewis (www.eeb.uconn.edu) An Introduction to Bayesian Phylogenetics Bayesian inference in general Markov chain Monte

More information

A Nonparametric Model-based Approach to Inference for Quantile Regression

A Nonparametric Model-based Approach to Inference for Quantile Regression A Nonparametric Model-based Approach to Inference for Quantile Regression Matthew Taddy and Athanasios Kottas Department of Applied Mathematics and Statistics, University of California, Santa Cruz, CA

More information

Lecturer: David Blei Lecture #3 Scribes: Jordan Boyd-Graber and Francisco Pereira October 1, 2007

Lecturer: David Blei Lecture #3 Scribes: Jordan Boyd-Graber and Francisco Pereira October 1, 2007 COS 597C: Bayesian Nonparametrics Lecturer: David Blei Lecture # Scribes: Jordan Boyd-Graber and Francisco Pereira October, 7 Gibbs Sampling with a DP First, let s recapitulate the model that we re using.

More information

Bayesian non-parametric model to longitudinally predict churn

Bayesian non-parametric model to longitudinally predict churn Bayesian non-parametric model to longitudinally predict churn Bruno Scarpa Università di Padova Conference of European Statistics Stakeholders Methodologists, Producers and Users of European Statistics

More information

A NOTE ON ROBUST ESTIMATION IN LOGISTIC REGRESSION MODEL

A NOTE ON ROBUST ESTIMATION IN LOGISTIC REGRESSION MODEL Discussiones Mathematicae Probability and Statistics 36 206 43 5 doi:0.75/dmps.80 A NOTE ON ROBUST ESTIMATION IN LOGISTIC REGRESSION MODEL Tadeusz Bednarski Wroclaw University e-mail: t.bednarski@prawo.uni.wroc.pl

More information

FAV i R This paper is produced mechanically as part of FAViR. See for more information.

FAV i R This paper is produced mechanically as part of FAViR. See  for more information. Bayesian Claim Severity Part 2 Mixed Exponentials with Trend, Censoring, and Truncation By Benedict Escoto FAV i R This paper is produced mechanically as part of FAViR. See http://www.favir.net for more

More information

A Fully Nonparametric Modeling Approach to. BNP Binary Regression

A Fully Nonparametric Modeling Approach to. BNP Binary Regression A Fully Nonparametric Modeling Approach to Binary Regression Maria Department of Applied Mathematics and Statistics University of California, Santa Cruz SBIES, April 27-28, 2012 Outline 1 2 3 Simulation

More information

Modelling geoadditive survival data

Modelling geoadditive survival data Modelling geoadditive survival data Thomas Kneib & Ludwig Fahrmeir Department of Statistics, Ludwig-Maximilians-University Munich 1. Leukemia survival data 2. Structured hazard regression 3. Mixed model

More information

Semiparametric Regression

Semiparametric Regression Semiparametric Regression Patrick Breheny October 22 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/23 Introduction Over the past few weeks, we ve introduced a variety of regression models under

More information

Multivariate spatial modeling

Multivariate spatial modeling Multivariate spatial modeling Point-referenced spatial data often come as multivariate measurements at each location Chapter 7: Multivariate Spatial Modeling p. 1/21 Multivariate spatial modeling Point-referenced

More information

Bayesian nonparametrics

Bayesian nonparametrics Bayesian nonparametrics 1 Some preliminaries 1.1 de Finetti s theorem We will start our discussion with this foundational theorem. We will assume throughout all variables are defined on the probability

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Dependent mixture models: clustering and borrowing information

Dependent mixture models: clustering and borrowing information ISSN 2279-9362 Dependent mixture models: clustering and borrowing information Antonio Lijoi Bernardo Nipoti Igor Pruenster No. 32 June 213 www.carloalberto.org/research/working-papers 213 by Antonio Lijoi,

More information

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined

More information

CTDL-Positive Stable Frailty Model

CTDL-Positive Stable Frailty Model CTDL-Positive Stable Frailty Model M. Blagojevic 1, G. MacKenzie 2 1 Department of Mathematics, Keele University, Staffordshire ST5 5BG,UK and 2 Centre of Biostatistics, University of Limerick, Ireland

More information

A Nonparametric Approach Using Dirichlet Process for Hierarchical Generalized Linear Mixed Models

A Nonparametric Approach Using Dirichlet Process for Hierarchical Generalized Linear Mixed Models Journal of Data Science 8(2010), 43-59 A Nonparametric Approach Using Dirichlet Process for Hierarchical Generalized Linear Mixed Models Jing Wang Louisiana State University Abstract: In this paper, we

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

STAT 425: Introduction to Bayesian Analysis

STAT 425: Introduction to Bayesian Analysis STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 2) Fall 2017 1 / 19 Part 2: Markov chain Monte

More information

A general mixed model approach for spatio-temporal regression data

A general mixed model approach for spatio-temporal regression data A general mixed model approach for spatio-temporal regression data Thomas Kneib, Ludwig Fahrmeir & Stefan Lang Department of Statistics, Ludwig-Maximilians-University Munich 1. Spatio-temporal regression

More information

Maximum Likelihood and Bayes Estimations under Generalized Order Statistics from Generalized Exponential Distribution

Maximum Likelihood and Bayes Estimations under Generalized Order Statistics from Generalized Exponential Distribution Applied Mathematical Sciences, Vol. 6, 2012, no. 49, 2431-2444 Maximum Likelihood and Bayes Estimations under Generalized Order Statistics from Generalized Exponential Distribution Saieed F. Ateya Mathematics

More information

Estimation of Quantiles

Estimation of Quantiles 9 Estimation of Quantiles The notion of quantiles was introduced in Section 3.2: recall that a quantile x α for an r.v. X is a constant such that P(X x α )=1 α. (9.1) In this chapter we examine quantiles

More information

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal

More information

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Markov chain Monte Carlo

Markov chain Monte Carlo 1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Bayesian nonparametric models for bipartite graphs

Bayesian nonparametric models for bipartite graphs Bayesian nonparametric models for bipartite graphs François Caron Department of Statistics, Oxford Statistics Colloquium, Harvard University November 11, 2013 F. Caron 1 / 27 Bipartite networks Readers/Customers

More information

AFT Models and Empirical Likelihood

AFT Models and Empirical Likelihood AFT Models and Empirical Likelihood Mai Zhou Department of Statistics, University of Kentucky Collaborators: Gang Li (UCLA); A. Bathke; M. Kim (Kentucky) Accelerated Failure Time (AFT) models: Y = log(t

More information

Infinite Mixtures of Gaussian Process Experts

Infinite Mixtures of Gaussian Process Experts in Advances in Neural Information Processing Systems 14, MIT Press (22). Infinite Mixtures of Gaussian Process Experts Carl Edward Rasmussen and Zoubin Ghahramani Gatsby Computational Neuroscience Unit

More information

Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures

Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures 17th Europ. Conf. on Machine Learning, Berlin, Germany, 2006. Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures Shipeng Yu 1,2, Kai Yu 2, Volker Tresp 2, and Hans-Peter

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture

More information

Bayesian data analysis in practice: Three simple examples

Bayesian data analysis in practice: Three simple examples Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to

More information

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008 Ages of stellar populations from color-magnitude diagrams Paul Baines Department of Statistics Harvard University September 30, 2008 Context & Example Welcome! Today we will look at using hierarchical

More information

A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness

A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A. Linero and M. Daniels UF, UT-Austin SRC 2014, Galveston, TX 1 Background 2 Working model

More information

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition. Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface

More information

Luke B Smith and Brian J Reich North Carolina State University May 21, 2013

Luke B Smith and Brian J Reich North Carolina State University May 21, 2013 BSquare: An R package for Bayesian simultaneous quantile regression Luke B Smith and Brian J Reich North Carolina State University May 21, 2013 BSquare in an R package to conduct Bayesian quantile regression

More information

Introduction to BGPhazard

Introduction to BGPhazard Introduction to BGPhazard José A. García-Bueno and Luis E. Nieto-Barajas February 11, 2016 Abstract We present BGPhazard, an R package which computes hazard rates from a Bayesian nonparametric view. This

More information