Bayesian Variable Selection Regression Of Multivariate Responses For Group Data
|
|
- Justin Weaver
- 5 years ago
- Views:
Transcription
1 Bayesian Variable Selection Regression Of Multivariate Responses For Group Data B. Liquet 1,2 and K. Mengersen 2 and A. N. Pettitt 2 and M. Sutton 2 1 LMAP, Université de Pau et des Pays de L Adour 2 ACEMS, QUT, Australia
2 Contents Motivation: Integrative Analysis for group data Application to pleiotropy investigation Sparse Model: Lasso penalty, Group penalty, Sparse Group Lasso Bayesian framework Simulation Studies Illustration: genomics data R package: MSGBSS
3 Integrative Analysis Goal of Integrative Analysis: Wikipedia. Data integration involves combining data residing in different sources and providing users with a unified view of these data. This process becomes significant in a variety of situations, which include both commercial and scientific. System Biology. Integrative Analysis: Analysis of heterogeneous types of data from inter-platform technologies. Goal. Combine multiple types of data: Contribute to a better understanding of biological mechanism. Have the potential to improve the diagnosis and treatments of complex diseases.
4 Example: Data definition p q n X - n observations - p variables Y - n observations - q variables n
5 Example: Data definition p q n X - n observations - p variables Y - n observations - q variables n Omics. Y matrix: gene expression, X matrix: SNP (single nucleotide polymorphism). Many others such as proteomic, metabolomic data. neuroimaging. Y matrix: behavioral variables, X matrix: brain activity (e.g., EEG, fmri, NIRS) neuroimaging genetics. Y matrix: fmri (Fusion of functional magnetic resonance imaging), X matrix: SNP Ecology/Environment. Y matrix: Water quality variables, X matrix: Landscape variables
6 Data: Constraints and Aims Main constraint: situation with p > n
7 Group structures within the data Natural example: Categorical variables which is a group of dummies variables in a regression setting.
8 Group structures within the data Natural example: Categorical variables which is a group of dummies variables in a regression setting. Genomics: genes within the same pathway have similar functions and act together in regulating a biological system. These genes can add up to have a larger effect can be detected as a group (i.e., at a pathway or gene set/module level).
9 Group structures within the data Natural example: Categorical variables which is a group of dummies variables in a regression setting. Genomics: genes within the same pathway have similar functions and act together in regulating a biological system. These genes can add up to have a larger effect can be detected as a group (i.e., at a pathway or gene set/module level). We consider variables are divided into groups: Example p: SNPs grouped into K genes X = [SNP 1,... + SNP } {{ } k SNP k+1, SNP k+2,..., SNP h... SNP } {{ } l+1,..., SNP p ] } {{ } gene 1 gene 2 gene K Example p: genes grouped into K pathways/modules (X j = gene j ) X = [X 1, X 2,..., X } {{ } k X k+1, X k+2,..., X h... X } {{ } l+1, X l+2,..., X p ] } {{ } M 1 M 2 M K
10 Aims in regression setting: Select group variables taking into account the data structures; all the variables within a group are selected otherwise none of them are selected Combine both sparsity of groups and within each group; only relevant variables within a group are selected
11 Some frequentist Appoaches Lasso models: regression model use L 1 and L 2 penalties for performing variable selection
12 Lasso models Univariate model : Y = Xβ + ε, Y R n, β R p Lasso regression: sparse model with L 1 penalty β lasso = argmin β R p n p (Y i X T i β) 2 + λ β j i=1 = argmin Y Xβ 2 2 +λ β 1 β R p } {{ }}{{} Penalty Loss Group Lasso regression: sparse model using group structure with L 2 penalty on group (X g with β g R p g ) min β R p j=1 G G Y X g β g λ pg β g 2 g=1 -If p g = 1 = Lasso method ( β g 2 = g=1 β 2 g = β g )
13 Lasso models Univariate model : Y = G g=1 X gβ g + ε, Y R n, β R p Sparse Group Lasso regression: sparse group model using group structure combining L 1 and L 2 penalties min β R p G G Y X g β g λ 1 pg β g 2 + λ 2 β 1 g=1 g=1
14 Lasso model for Multivariate Y where Y (n q), B (p q) Y = XB + E Sparse model: select predictors associated to the multiple phenotype Y (n q): p min B Y XB 2 F + λ β l 2 where B = β 1 1 β β q 1 β 1 2 β β q 2 l=1, β l = (β 1 l...., β2 l,..., βq l ) β 1 p β 2 p... β q p Implemented in glmnet() Sparse Group model: see e.g., Li et al. (2015), Wand et al. (2012).
15 Bayesian Framework Univariate regression model: Y G g=1 X gβ g N n ( 0, σ 2 I n ) Xu and Ghosh (2015): Bayesian group lasso using spike and slab priors for group variable selection. Rockova and Lesafre (2014): (EM) algorithm for a hierarchical model incorporating grouping information. Stingo et al. (2011): PLS approach for pathway and gene selection using variable selection priors and Markov chain Monte Carlo (MCMC)
16 Bayesian Framework Multivariate regression model: G ( Y X g B g MN n q 0p q, I n, ) Σ g=1 Zhu et al. (2014): Bayesian generalized low rank regression model. Greenlaw et al. (2016): Bayesian hierarchical modeling for imaging genomics data. Limited to Σ = σ 2 I q.
17 Bayesian group lasso model with spike and slab priors Xu and Ghosh (2015) approaches: spike and slab priors providing variable selection at the group level. hierarchical spike and slab prior structure to select variables both at the group level and within each group. Limited to univariate case. Doesn t take into account group size in the penalisation part.
18 Multivariate Bayesian Group Lasso with Spike and Slab prior Y X, B, Σ MN n q (XB, Σ, I n ), (1) Vec(B T g Σ, τ g, π 0 ) ind (1 π 0 )N mg q(0, I mg τ 2 g Σ) + π 0δ 0 (Vec(B T g )), (2) ( ) τ 2 ind mg + 1 g Gamma, λ2, g = 1,..., G, (3) 2 2 Σ IW(d, Q) (4) π 0 Beta(a, b) (5) where δ 0 (Vec(B T g )) denotes a point mass at 0 Rm gq, B g is the m g q regression coefficient matrix for the group g
19 Calibration of λ λ is related to the coefficient of shrinkage. tuned using an empirical Bayes approach. a Monte Carlo EM algorithm is used. the kth EM update for λ is: λ (k) p + G = G g=1 E [ λ τ 2 (k 1) g Y ], in which the posterior expectation of τ 2 g is replaced by the Monte Carlo sample average of τ 2 g generated in the Gibbs sample based on λ (k 1).
20 Median Thresholding Estimator In the case of a block orthogonal design matrix X (i.e., X T i X j = 0 for i j), we have for 1 g G Vec( B T g ) = Vec ( ((X T g X g ) 1 X T g Y) T) N mg q ( Vec(B T g ), (X T g X g ) 1 Σ ). We showed assuming π 0 > c 1+c, then there exists t(π 0) > 0, such that the marginal posterior median of β g satisfies ij Med(β g ij B g ) = 0 for any 1 i m g and 1 j q when Vec( B T g ) 2 < t. The marginal posterior median estimator of the gth group of regression coefficients is zero when the norm of the corresponding block least square estimator is less than a certain threshold.
21 Posterior median as a soft thresholding estimator Assuming an orthogonal design matrix X, i.e., X T X = ni p and consider our model defined with fixed τ 2 g (1 g G). Posterior distribution of B g is a spike and slab distribution, ( Vec(B T g ) X, Y (1 l g)n mgq (1 D g )Vec(B T LS,g ), 1 D ) g I mg Σ + l g δ 0 (Vec(B T g n )), where B LS,g is the least squares estimator of B g, D g = l g = π 0 1, and l 1+nτ 2 g = p(b g = 0 rest) g π 0 + (1 π 0 )(τ 2 mg (q 1) g ) 2 (1 + nτ 2 mg { g ) 2 exp (1 D g )ntr[σ 1 B T LS,g B } LS,g] The marginal posterior distribution of β g ij is also a spike and slab distribution, ( β g ij X, Y (1 l g)n (1 D g ) β g LS,ij, 1 D ) g Σ jj + l g δ 0 (β g n ij ), where Σ jj is the j-th diagonal element of Σ.
22 Posterior median as a soft thresholding estimator The resulting median is a soft thresholding estimator defined by β Med,g ij = Med(β g ij X, Y) = sgn ( β g LS,ij ) (1 D g) β g Σjj LS,ij Q g 1 D g, n + where z + denotes the positive part of z and Q g = φ 1 ( 1 ). 2(1 min( 1 2,l g)) For a univariate response (q = 1) the matrix Σ reduces to the scalar σ 2, and our result matches the previous work of Xu et al (2015). In the multivariate frequentist setting, Li et al. (2015) proposed an iterative algorithm with similar soft thresholding function to incorporate group structure in estimating the regression estimates.
23 Oracle property Let B 0, B 0 g, β 0,g ij denote the true values of B, B g, β g ij, respectively. The index vector of the true model as A = (I( Vec(B g ) 2 0), g = 1,..., G), The index vector of the model selected by certain thresolding estimator B g as A n = (I( Vec( B g ) 2 0), g = 1,..., G). Model selection consistency is attained if and only if lim n P(A n = A) = 1.
24 Oracle property Under orthogonal design the median thresholding estimator has oracle property. Theorem: Assume orthogonal design matrix, i.e., X T X = ni p. Suppose nτ 2 g,n and log(τ 2 g,n )/n 0 as n, for g = 1,..., G, then the median thresholding estimator has oracle property, that is, variable selection consistency, and asymptotic normality, lim n P(AMed n = A) = 1 ( n Vec( B Med A ) Vec(B0 A )) d N(0, Σ I).
25 Gibbs Sampler: full posterior distribution where p(b, τ 2, Σ, π 0 Y, X) p(y B, τ 2, Σ, π 0 ) p(b τ 2, Σ, π 0 ) p(y B, τ 2, Σ, π 0 ) Σ n/2 exp p(b τ 2, Σ, π 0 ) = p(τ 2 ) p(σ) p(π 0 ), { 1 2 Tr [ (Y XB)Σ 1 (Y XB) T ]}, G p(b g τ 2 g, Σ, π 0 ), g=1 p(b g τ 2 g, Σ, π 0 ) (1 π 0 )(2π) qmg 2 (τ 2 qmg g ) 2 Σ mg 1 2 exp [ ] 2τ 2 Tr B g Σ 1 B T g I[B g 0] g +π 0 δ 0 (Vec(B T g )), G p(τ 1,..., τ g ) (λ 2 qmg +1 ) 2 (τ 2 qmg +1 g ) 2 1 exp λ2 m g g=1 p(π 0 ) π a 1 0 (1 π 0 ) b 1, p(σ) Σ d+q+1 2 exp { 1 2 Tr(QΣ 1 ) } 2 τ2 g,
26 Conditional posterior distribution Let B (g) denote the B matrix without the gth group, and X (g) denote the covariate matrix corresponding to B (g), that is, X (g) = (X 1,..., X g 1, X g+1,..., X G ) where X g is the design matrix corresponding to B g. Conditional posterior distribution of B g : spike and slab distribution Conditional posterior distribution of α 2 g = 1 τ 2 g Inverse Gamma if B g = 0 Inverse Gaussian if B g 0 Conditional posterior distribution of Σ: Inverse Wishart Conditional posterior distribution of π 0 : Beta distribution
27 Multivariate Sparse Group Selection with Spike and Slab Prior (MBSGS-SS) Reparametrize the coefficients matrices to tackle the two kinds of sparsity separately: B g = V 1 2 g B g, g = 1,..., G; j = 1,..., m g, with V 1 2 g = diag{τ g1,..., τ gmg }, τ gj 0 and where B g, when nonzero, follow Vec( B T g ) N m g q(0, I mg Σ). Diagonal element of V 1 2 g control the magnitude of the elements of B g.
28 Multivariate Sparse Group Selection with Spike and Slab Prior (MBSGS-SS) To select variables at the group level, we assume the multivariate spike and slab prior for each Vec( B T g ): Vec( B T g Σ, τ g, π 0 ) ind (1 π 0 )N mg q(0, I mg Σ) + π 0 δ 0 (Vec( B T g )) Note that when τ gj = 0, the j-th row of B g is essentially dropped out of the model even when B j g 0. So in order to choose variables within each relevant group, we assume the following spike and slab prior for each τ gj : τ gj ind (1 π 1 )N + (0, s 2 ) +π 1 δ 0 (τ gj ), g = 1,..., G; j = 1,..., m g, where N + (0, s 2 ) denotes a normal N(0, s 2 ) distribution truncated below at 0.
29 Prior Specification We assume an Inverse Wishart prior for Σ IW(d, Q) We assume conjugate beta hyper-priors for π 0 and π 1. We use a conjugate inverse gamma prior for s 2 Inverse Gamma(1, t), We stimate t with the Monte Carlao EM algorithm. For the k-th EM update, t (k) 1 = [ 1 Y ], s 2 E t (k 1) where the posterior expectation of Gibbs samples based on t (k 1). 1 s 2 is estimated from the
30 Gibbs Sampler: full posterior distribution Joint posterior p( B, τ 2, Σ, π 0, π 1, s 2 Y, X) Σ n/2 exp 1 T G G 2 Tr Y X g V 1 2 g B g Σ 1 Y X g V 1 2 g B g g=1 G (1 π 0 )(2π) qmg 2 Σ mg 2 exp g=1 g=1 { 1 [ B 2 Tr g Σ 1 B ] } T g I[ B g 0] +π 0 δ 0 (Vec( B T g )) G m g (1 π 1)2(2πs 2 ) 1 2 exp τ2 gj 2s 2 I[τ gj > 0] + π 1 δ 0 (τ gj ) g=1 j=1 { Σ d+q+1 2 exp 1 } 2 Tr(QΣ 1 ) π a (1 π 0 ) a 2 1 π c 1 1 (1 π 1 1 ) c 2 1 { t(s 2 ) 2 exp t } s 2
31 Simulation Studies Comparison to Lasso, group Lasso, sparse goup Lasso for univariate setting R package glmnet and SGL Comparison to Lasso for Mulivariate setting R package glmnet Bayseian approaches BGL-SS (Bayesian Group Lasso with Spike and Slab prior) BSGS-SS (Bayesian Sparse Group selection with spike and slab priors) MBGL-SS for multivariate setting MBSGL-SS for multivariate setting running iterations in which the first are burn-ins.
32 Simulation results: summary Simulations results suggest that the multivariate Bayesian group lasso with spike and slab prior is strongly influenced by a combination of different group size structures and high correlation between predictors. The multivariate Bayesian sparse group selection with spike and slab prior does not suffer in this situation. Most powerful method for variable selection and prediction performance in the presence of group structure data All numerical results from our article could be reproduced using our R package MBSGS available on CRAN.
33 Computation The current version of our package runs for example: a MBSGS-SS model in around 2 minutes a MBGL-SS model in around 1 minute for a model with 20 groups of 5 variables for a sample size (n = 900) with iterations including for the burnin. Further improvements of the code such as parallelization over the group structure are in progress to speed up the computational time for tacking Big Data sets In this context of genetics studies, some further extensions of our model are under investigation such as integrating different group penalties given a biological prior of the pathways or different distribution priors for each group.
34 Application: Aim and Data Identify a parsimonious set of predictors that explains the joint variability of gene expression in four tissues (adrenal gland, fat, heart, and kidney). 770 SNPs in 29 inbred rats as a predictor matrix (n = 29, p = 770) 29 measured expression levels in the 4 tissues as the outcome (q = 4).
35 Application chromosome information defines the group structure of the predictor matrix ran our MBGL-SS and MBSGS-SS models using this group structure The multivariate lasso selects 69 SNPs which come from the 20 chromosomes. The MBGL-SS selects only the two first groups corresponding to the SNPs from chromosomes 1 and 2.
36 Application: MBSGS-SS results, 32 SNPs from 8 chromosomes
37 Application: results SNP D14Mit (from chromosome 10) has been previously identified Highest estimate (0.334) for the heart tissue the posterior standard deviation of the regression parameter for each selected non-zero median estimate was in the range 0.11 to SNP D14Mit3 estimate was with posterior standard deviation Importance of chromosomes could be investigated using an estimate of the probability of inclusion
38 Concluding remarks: summary Bayesian methods for group-sparse modeling in the context of a multivariate correlated response variable. Our models are based on spike and slab type priors which facilitate variable selection. Importance to include the group size information in the shrinkage part of our model. We have shown that the posterior median estimator could both select and estimate the regression coefficients. Simulation results suggest very good performance of the posterior median estimator for variable selection and prediction error. This estimator obtains similar results as the highest probability model in terms of true and false positive rates.
39 References - Li, Y., Nan, B., and Zhu, J. (2015). *Multivariate Sparse Group Lasso for the Multivariate Multiple Linear Regression with an Arbitrary Group Structure.* Biometrics, 71(2): Liquet, B., Lafaye de Micheaux, P., Hejblum, B., and Thiebaut, R. (2016). *Group and Sparse Group Partial Least Square Approaches Applied in Genomics Context.* Bioinformatics, 3(1): Liquet, B. and Sutton, M. (2016). *MBSGS: Multivariate Bayesian Sparse Group Selection with Spike and Slab*. R package version URL - Liquet, B., Mengersen, K., Pettitt, A. N. and Sutton, M. (2016). *Bayesian Variable Selection Regression Of Multivariate Responses For Group Data* Bayesian Analysis. Volume 12, Number 4 (2017), Received the Lindley award from ISBA Prize Committee, delivered during ISBA World Meeting in Edinburgh Xu, X. and Ghosh, M. (2015). *Bayesian Variable Selection and Estimation for Group Lasso.* Bayesian Analysis, 10(4):
40 Extension to Pleotropic mapping for genome-wide association studies using group variable selection Pleiotropy: genetic variants which affect multiple different complex diseases Example: genetic variants which affect both Breast and Thyroid cancer. Results from GWAS suggest that complex diseases are often affected by many variants with small effects (known as polygenicity) Aims: statistical method to leverage pleiotropic effects incoporate prior pathway knowledge to increase statistical power and identify important risk variants.
41 Model for multiple GWAS studies Suppose we have data from K independent GWAS datasets, D = D 1 D K, where D k = ({y 1, x 1 },..., {y nk, x nk }) y ik {0, 1} denotes the phenotype of the kth study x ik R p is the vector with corresponding p SNPs. Logistic regression model y ik Bernoulli ( g 1 (ν ik ) ) ν ik = x T ik β k for k = 1,..., K, where g( ) is the logistic link function β k R p the regression coefficients for the kth GWAS. Let β j R K, j = 1,..., p, the vector of K regression coefficients corresponding to the jth SNP over the K GWAS.
42 Group Structure SNPs can be partitioned into G groups (genes) Let π g, g = 1,..., G the set of SNPs contained in the gth group with p g = π g. Matrix of all regression coefficients as B = (β 1,..., β K ) = (β 1,..., β p ) T.
43 Frequentist Approach The log likelihood for the combined datasets: p(d B) = K n k ( ) L y ik β T k x ik k=1 i=1 The penalised likelihood estimate B = argmin B R p K where L(x) = ln(1 + e x ) K n k ( ) L y ik β T k x ik + λ1 B + λ G2,1 2 B l2,1 (6) k=1 i=1 G 2,1 -norm penalty B G2,1 = G l 2,1 -norm penalty B l2,1 = p i=1 g=1 i π g K j=1 β2 ik K k=1 β2 ik respectively. The G 2,1 -norm fixes the group structure across studies and encourages sparsity at a gene level. The l 2,1 -norm which allows sparsity within a group.
44 Inference Inference using the alternating direction method of multipliers algorithm (ADMM). Novel approach for identifying pleiotropic effects as it accounts for gene specific and SNP specific effects using a variable selection approach. The method is only capable of producing a point estimate of B and acurrate estimation of the variance for these parameters is not easily given.
45 Bayesian Logistic regression with multivariate spike and slab prior: LogitMBGL-SS Let γ = (γ 1,..., γ p ) T indicate the association status for SNPs where γ j = 1 indicates that the jth SNP is associated to all K traits. Spike and slab prior for the jth SNP β j R K, β j (1 γ j )N K (0, τ 2 j V) + γ jδ 0 (β j ) ( K + 1 τ 2 j Gamma 2, λ ), 2 V IW(d, Q), γ j Bernolli(α 0 ) α 0 Beta(a, b) for j = 1,..., p, where δ 0 (β j ) denotes a point mass at 0 R K. Here, V R K K is a covariance matrix modeling the covariance of the SNP effect on the traits.
46 Extension Should perform well when the SNPs are independent. GWAS datasets: strong correlations that can occur between SNPs within the same gene. Solution: reparameterise the coefficeints to handle the sparsity at a gene grouping level and individual feature level separately. τ R p to model individual sparsity b (g) R pgk with b (g) = (b (g)t,..., b (g)t 1 p g ) where b (g) j group sparsity. R K for β j = τ j b (g) j, where τ j 0, for all j π g.
47 Bayesian Logistic regression using multivariate sparse group selection with spike and slab priors β j = τ j b (g) j, where τ j 0, for all j π g. We assume the following multivariate spike and slab b (g) (1 α 0 )N pg K (0, I pg V) + α 0 δ 0 (b (g) ) τ j (1 α 1 )N ( + 0, s 2) + α 1 δ 0 (τ j ), α 0 Beta(a 1, a 2 ) α 1 Beta(c 1, c 2 ) s 2 InvGamma(1, t) for j π g and g = 1,..., G
48 Signal recovery: (i) gmt: Grouped multi-task penalised logistic regression (λ 1 > 0, λ 2 = 0) using G 2,1 -norm (ii) smt: Sparse multi-task penalid penalised logistic regression (λ 1 = 0, λ 2 > 0 ) using l 2,1 -norm (iii) sgmt: Sparse group multi-task penalised logistic regression (λ 1 > 0, λ 2 > 0 ) (iv) logitmbgl: Bayesian logistic regression using multivariate group lasso with spike and slab prior (v) logitmbsgs: Bayesian logistic regression using multivariate sparse group selection with spike and slab prior
49 Results: Frequentist group gmt Study 1 gmt Study value of coefficient sgmt Study 1 sgmt Study 2 smt Study 1 smt Study 2 Study 1 Study 2 index of coefficient
50 Results: Bayesian group MBGL Study 1 MBGL Study 2 value of coefficient MBSGS Study 1 MBSGS Study 2 Study 1 Study 2 index of coefficient
51 Main Conclusion on the simulation studies The penalised approaches perform reasonably well in variable selection but the reconstructed signal is underestimated. In general, the penalised likelihood methods suffer in terms of false negatives, selecting more variables to be nonzero than the Bayesian methods. The Bayesian methods perform the best in terms of signal recovery measured by the l 1 error and variable selection performance metrics. The penalised likelihood approaches are computationally efficient using alternating direction method of multipliers algorithm Simulation results suggest that when computationally possible the Bayesian estimators should be used. The multivariate Bayesian sparse group selection with spike and slab prior performed the best in terms of signal recovery. The Bayesian method provides a natural method for quantifying the variability of the estimated coefficients.
52 What Next? Application on real data: case/control studies Breast Cancer and Thyroide Cancer Thyroide Cancer (482 case, 463 control) Breast Cancer (1172 case, 1125 control) 6677 SNPs from 618 genes from 10 non-overlapping gene pathways.
53 What Next? Application on real data: case/control studies Breast Cancer and Thyroide Cancer Thyroide Cancer (482 case, 463 control) Breast Cancer (1172 case, 1125 control) 6677 SNPs from 618 genes from 10 non-overlapping gene pathways. Looking for one postdoc position to fill for 2 years granted by la ligue contre le Cancer post-doctorat.html
Bayesian Variable Selection Regression of Multivariate Responses for Group Data
Bayesian Analysis 2017) 12, Number 4, pp. 1039 1067 Bayesian Variable Selection Regression of Multivariate Responses for Group Data B. Liquet,, K. Mengersen, A. N. Pettitt, and M. Sutton, Abstract. We
More informationRegularization Parameter Selection for a Bayesian Multi-Level Group Lasso Regression Model with Application to Imaging Genomics
Regularization Parameter Selection for a Bayesian Multi-Level Group Lasso Regression Model with Application to Imaging Genomics arxiv:1603.08163v1 [stat.ml] 7 Mar 016 Farouk S. Nathoo, Keelin Greenlaw,
More informationOr How to select variables Using Bayesian LASSO
Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO On Bayesian Variable Selection
More information25 : Graphical induced structured input/output models
10-708: Probabilistic Graphical Models 10-708, Spring 2013 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Meghana Kshirsagar (mkshirsa), Yiwen Chen (yiwenche) 1 Graph
More informationEstimating Sparse High Dimensional Linear Models using Global-Local Shrinkage
Estimating Sparse High Dimensional Linear Models using Global-Local Shrinkage Daniel F. Schmidt Centre for Biostatistics and Epidemiology The University of Melbourne Monash University May 11, 2017 Outline
More informationAn Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models
Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS023) p.3938 An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Vitara Pungpapong
More informationA New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables
A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables Qi Tang (Joint work with Kam-Wah Tsui and Sijian Wang) Department of Statistics University of Wisconsin-Madison Feb. 8,
More informationCS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS
CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS * Some contents are adapted from Dr. Hung Huang and Dr. Chengkai Li at UT Arlington Mingon Kang, Ph.D. Computer Science, Kennesaw State University Problems
More informationIntegrated Anlaysis of Genomics Data
Integrated Anlaysis of Genomics Data Elizabeth Jennings July 3, 01 Abstract In this project, we integrate data from several genomic platforms in a model that incorporates the biological relationships between
More informationScalable MCMC for the horseshoe prior
Scalable MCMC for the horseshoe prior Anirban Bhattacharya Department of Statistics, Texas A&M University Joint work with James Johndrow and Paolo Orenstein September 7, 2018 Cornell Day of Statistics
More informationSparse Linear Models (10/7/13)
STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine
More informationExpression Data Exploration: Association, Patterns, Factors & Regression Modelling
Expression Data Exploration: Association, Patterns, Factors & Regression Modelling Exploring gene expression data Scale factors, median chip correlation on gene subsets for crude data quality investigation
More informationBayesian construction of perceptrons to predict phenotypes from 584K SNP data.
Bayesian construction of perceptrons to predict phenotypes from 584K SNP data. Luc Janss, Bert Kappen Radboud University Nijmegen Medical Centre Donders Institute for Neuroscience Introduction Genetic
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationProbabilistic machine learning group, Aalto University Bayesian theory and methods, approximative integration, model
Aki Vehtari, Aalto University, Finland Probabilistic machine learning group, Aalto University http://research.cs.aalto.fi/pml/ Bayesian theory and methods, approximative integration, model assessment and
More informationShrinkage Methods: Ridge and Lasso
Shrinkage Methods: Ridge and Lasso Jonathan Hersh 1 Chapman University, Argyros School of Business hersh@chapman.edu February 27, 2019 J.Hersh (Chapman) Ridge & Lasso February 27, 2019 1 / 43 1 Intro and
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationBayesian Inference of Multiple Gaussian Graphical Models
Bayesian Inference of Multiple Gaussian Graphical Models Christine Peterson,, Francesco Stingo, and Marina Vannucci February 18, 2014 Abstract In this paper, we propose a Bayesian approach to inference
More informationVariable Selection in Structured High-dimensional Covariate Spaces
Variable Selection in Structured High-dimensional Covariate Spaces Fan Li 1 Nancy Zhang 2 1 Department of Health Care Policy Harvard University 2 Department of Statistics Stanford University May 14 2007
More information25 : Graphical induced structured input/output models
10-708: Probabilistic Graphical Models 10-708, Spring 2016 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Raied Aljadaany, Shi Zong, Chenchen Zhu Disclaimer: A large
More informationBayesian Grouped Horseshoe Regression with Application to Additive Models
Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu 1,2, Daniel F. Schmidt 1, Enes Makalic 1, Guoqi Qian 2, John L. Hopper 1 1 Centre for Epidemiology and Biostatistics,
More informationConsistent high-dimensional Bayesian variable selection via penalized credible regions
Consistent high-dimensional Bayesian variable selection via penalized credible regions Howard Bondell bondell@stat.ncsu.edu Joint work with Brian Reich Howard Bondell p. 1 Outline High-Dimensional Variable
More informationMultidimensional data analysis in biomedicine and epidemiology
in biomedicine and epidemiology Katja Ickstadt and Leo N. Geppert Faculty of Statistics, TU Dortmund, Germany Stakeholder Workshop 12 13 December 2017, PTB Berlin Supported by Deutsche Forschungsgemeinschaft
More informationKatsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract
Bayesian analysis of a vector autoregressive model with multiple structural breaks Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus Abstract This paper develops a Bayesian approach
More informationESL Chap3. Some extensions of lasso
ESL Chap3 Some extensions of lasso 1 Outline Consistency of lasso for model selection Adaptive lasso Elastic net Group lasso 2 Consistency of lasso for model selection A number of authors have studied
More informationPackage LBLGXE. R topics documented: July 20, Type Package
Type Package Package LBLGXE July 20, 2015 Title Bayesian Lasso for detecting Rare (or Common) Haplotype Association and their interactions with Environmental Covariates Version 1.2 Date 2015-07-09 Author
More informationPre-Selection in Cluster Lasso Methods for Correlated Variable Selection in High-Dimensional Linear Models
Pre-Selection in Cluster Lasso Methods for Correlated Variable Selection in High-Dimensional Linear Models Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider variable
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationNonparametric Bayes tensor factorizations for big data
Nonparametric Bayes tensor factorizations for big data David Dunson Department of Statistical Science, Duke University Funded from NIH R01-ES017240, R01-ES017436 & DARPA N66001-09-C-2082 Motivation Conditional
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationStatistical Inference
Statistical Inference Liu Yang Florida State University October 27, 2016 Liu Yang, Libo Wang (Florida State University) Statistical Inference October 27, 2016 1 / 27 Outline The Bayesian Lasso Trevor Park
More informationModel-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate
Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Lucas Janson, Stanford Department of Statistics WADAPT Workshop, NIPS, December 2016 Collaborators: Emmanuel
More informationBayesian Linear Regression
Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective
More informationFeature Selection via Block-Regularized Regression
Feature Selection via Block-Regularized Regression Seyoung Kim School of Computer Science Carnegie Mellon University Pittsburgh, PA 3 Eric Xing School of Computer Science Carnegie Mellon University Pittsburgh,
More informationAlternative implementations of Monte Carlo EM algorithms for likelihood inferences
Genet. Sel. Evol. 33 001) 443 45 443 INRA, EDP Sciences, 001 Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Louis Alberto GARCÍA-CORTÉS a, Daniel SORENSEN b, Note a
More informationHypes and Other Important Developments in Statistics
Hypes and Other Important Developments in Statistics Aad van der Vaart Vrije Universiteit Amsterdam May 2009 The Hype Sparsity For decades we taught students that to estimate p parameters one needs n p
More informationLarge-scale Ordinal Collaborative Filtering
Large-scale Ordinal Collaborative Filtering Ulrich Paquet, Blaise Thomson, and Ole Winther Microsoft Research Cambridge, University of Cambridge, Technical University of Denmark ulripa@microsoft.com,brmt2@cam.ac.uk,owi@imm.dtu.dk
More informationSparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference
Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference Shunsuke Horii Waseda University s.horii@aoni.waseda.jp Abstract In this paper, we present a hierarchical model which
More informationCoupled Hidden Markov Models: Computational Challenges
.. Coupled Hidden Markov Models: Computational Challenges Louis J. M. Aslett and Chris C. Holmes i-like Research Group University of Oxford Warwick Algorithms Seminar 7 th March 2014 ... Hidden Markov
More informationBayesian linear regression
Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding
More informationGibbs Sampling in Linear Models #2
Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling
More informationA Bayesian Nonparametric Model for Predicting Disease Status Using Longitudinal Profiles
A Bayesian Nonparametric Model for Predicting Disease Status Using Longitudinal Profiles Jeremy Gaskins Department of Bioinformatics & Biostatistics University of Louisville Joint work with Claudio Fuentes
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationPart 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior
Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior Tom Heskes joint work with Marcel van Gerven
More informationChris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010
Model-Averaged l 1 Regularization using Markov Chain Monte Carlo Model Composition Technical Report No. 541 Department of Statistics, University of Washington Chris Fraley and Daniel Percival August 22,
More informationOn Bayesian Computation
On Bayesian Computation Michael I. Jordan with Elaine Angelino, Maxim Rabinovich, Martin Wainwright and Yun Yang Previous Work: Information Constraints on Inference Minimize the minimax risk under constraints
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationGeneralized Elastic Net Regression
Abstract Generalized Elastic Net Regression Geoffroy MOURET Jean-Jules BRAULT Vahid PARTOVINIA This work presents a variation of the elastic net penalization method. We propose applying a combined l 1
More informationA new Hierarchical Bayes approach to ensemble-variational data assimilation
A new Hierarchical Bayes approach to ensemble-variational data assimilation Michael Tsyrulnikov and Alexander Rakitko HydroMetCenter of Russia College Park, 20 Oct 2014 Michael Tsyrulnikov and Alexander
More informationSTA 216, GLM, Lecture 16. October 29, 2007
STA 216, GLM, Lecture 16 October 29, 2007 Efficient Posterior Computation in Factor Models Underlying Normal Models Generalized Latent Trait Models Formulation Genetic Epidemiology Illustration Structural
More informationMotivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University
Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined
More informationBayesian methods in economics and finance
1/26 Bayesian methods in economics and finance Linear regression: Bayesian model selection and sparsity priors Linear Regression 2/26 Linear regression Model for relationship between (several) independent
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationGWAS IV: Bayesian linear (variance component) models
GWAS IV: Bayesian linear (variance component) models Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS IV: Bayesian
More informationNovember 2002 STA Random Effects Selection in Linear Mixed Models
November 2002 STA216 1 Random Effects Selection in Linear Mixed Models November 2002 STA216 2 Introduction It is common practice in many applications to collect multiple measurements on a subject. Linear
More informationBAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage
BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage Lingrui Gan, Naveen N. Narisetty, Feng Liang Department of Statistics University of Illinois at Urbana-Champaign Problem Statement
More informationVector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.
Vector Autoregressive Model Vector Autoregressions II Empirical Macroeconomics - Lect 2 Dr. Ana Beatriz Galvao Queen Mary University of London January 2012 A VAR(p) model of the m 1 vector of time series
More informationAspects of Feature Selection in Mass Spec Proteomic Functional Data
Aspects of Feature Selection in Mass Spec Proteomic Functional Data Phil Brown University of Kent Newton Institute 11th December 2006 Based on work with Jeff Morris, and others at MD Anderson, Houston
More informationBayesian Inference of Interactions and Associations
Bayesian Inference of Interactions and Associations Jun Liu Department of Statistics Harvard University http://www.fas.harvard.edu/~junliu Based on collaborations with Yu Zhang, Jing Zhang, Yuan Yuan,
More informationPart 6: Multivariate Normal and Linear Models
Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of
More informationChapter 17: Undirected Graphical Models
Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)
More informationSelection-adjusted estimation of effect sizes
Selection-adjusted estimation of effect sizes with an application in eqtl studies Snigdha Panigrahi 19 October, 2017 Stanford University Selective inference - introduction Selective inference Statistical
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public
More informationBayes methods for categorical data. April 25, 2017
Bayes methods for categorical data April 25, 2017 Motivation for joint probability models Increasing interest in high-dimensional data in broad applications Focus may be on prediction, variable selection,
More informationGraphical Models for Collaborative Filtering
Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,
More informationNearest Neighbor Gaussian Processes for Large Spatial Data
Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns
More informationStat 542: Item Response Theory Modeling Using The Extended Rank Likelihood
Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal
More informationAn algorithm for the multivariate group lasso with covariance estimation
An algorithm for the multivariate group lasso with covariance estimation arxiv:1512.05153v1 [stat.co] 16 Dec 2015 Ines Wilms and Christophe Croux Leuven Statistics Research Centre, KU Leuven, Belgium Abstract
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationCross-sectional space-time modeling using ARNN(p, n) processes
Cross-sectional space-time modeling using ARNN(p, n) processes W. Polasek K. Kakamu September, 006 Abstract We suggest a new class of cross-sectional space-time models based on local AR models and nearest
More informationMultivariate Survival Analysis
Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in
More information27: Case study with popular GM III. 1 Introduction: Gene association mapping for complex diseases 1
10-708: Probabilistic Graphical Models, Spring 2015 27: Case study with popular GM III Lecturer: Eric P. Xing Scribes: Hyun Ah Song & Elizabeth Silver 1 Introduction: Gene association mapping for complex
More informationOn Gaussian Process Models for High-Dimensional Geostatistical Datasets
On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May
More informationScale Mixture Modeling of Priors for Sparse Signal Recovery
Scale Mixture Modeling of Priors for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Jason Palmer, Zhilin Zhang and Ritwik Giri Outline Outline Sparse
More informationProteomics and Variable Selection
Proteomics and Variable Selection p. 1/55 Proteomics and Variable Selection Alex Lewin With thanks to Paul Kirk for some graphs Department of Epidemiology and Biostatistics, School of Public Health, Imperial
More informationLearning Gaussian Graphical Models with Unknown Group Sparsity
Learning Gaussian Graphical Models with Unknown Group Sparsity Kevin Murphy Ben Marlin Depts. of Statistics & Computer Science Univ. British Columbia Canada Connections Graphical models Density estimation
More informationBayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson
Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n
More informationStrong Lens Modeling (II): Statistical Methods
Strong Lens Modeling (II): Statistical Methods Chuck Keeton Rutgers, the State University of New Jersey Probability theory multiple random variables, a and b joint distribution p(a, b) conditional distribution
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationPackage IGG. R topics documented: April 9, 2018
Package IGG April 9, 2018 Type Package Title Inverse Gamma-Gamma Version 1.0 Date 2018-04-04 Author Ray Bai, Malay Ghosh Maintainer Ray Bai Description Implements Bayesian linear regression,
More informationBi-level feature selection with applications to genetic association
Bi-level feature selection with applications to genetic association studies October 15, 2008 Motivation In many applications, biological features possess a grouping structure Categorical variables may
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationMonte Carlo in Bayesian Statistics
Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview
More informationPartial factor modeling: predictor-dependent shrinkage for linear regression
modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework
More informationMULTILEVEL IMPUTATION 1
MULTILEVEL IMPUTATION 1 Supplement B: MCMC Sampling Steps and Distributions for Two-Level Imputation This document gives technical details of the full conditional distributions used to draw regression
More informationAssociation studies and regression
Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration
More informationThe STS Surgeon Composite Technical Appendix
The STS Surgeon Composite Technical Appendix Overview Surgeon-specific risk-adjusted operative operative mortality and major complication rates were estimated using a bivariate random-effects logistic
More informationPackage horseshoe. November 8, 2016
Title Implementation of the Horseshoe Prior Version 0.1.0 Package horseshoe November 8, 2016 Description Contains functions for applying the horseshoe prior to highdimensional linear regression, yielding
More informationwissen leben WWU Münster
MÜNSTER Sparsity Constraints in Bayesian Inversion Inverse Days conference in Jyväskylä, Finland. Felix Lucka 19.12.2012 MÜNSTER 2 /41 Sparsity Constraints in Inverse Problems Current trend in high dimensional
More informationIntegrated Non-Factorized Variational Inference
Integrated Non-Factorized Variational Inference Shaobo Han, Xuejun Liao and Lawrence Carin Duke University February 27, 2014 S. Han et al. Integrated Non-Factorized Variational Inference February 27, 2014
More informationWavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis
Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Jeffrey S. Morris University of Texas, MD Anderson Cancer Center Joint wor with Marina Vannucci, Philip J. Brown,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationConfidence Intervals for Low-dimensional Parameters with High-dimensional Data
Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology
More informationBayesian Grouped Horseshoe Regression with Application to Additive Models
Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu, Daniel F. Schmidt, Enes Makalic, Guoqi Qian, and John L. Hopper Centre for Epidemiology and Biostatistics, Melbourne
More informationApproximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA)
Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Willem van den Boom Department of Statistics and Applied Probability National University
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More information