Bayesian Variable Selection Regression Of Multivariate Responses For Group Data

Size: px
Start display at page:

Download "Bayesian Variable Selection Regression Of Multivariate Responses For Group Data"

Transcription

1 Bayesian Variable Selection Regression Of Multivariate Responses For Group Data B. Liquet 1,2 and K. Mengersen 2 and A. N. Pettitt 2 and M. Sutton 2 1 LMAP, Université de Pau et des Pays de L Adour 2 ACEMS, QUT, Australia

2 Contents Motivation: Integrative Analysis for group data Application to pleiotropy investigation Sparse Model: Lasso penalty, Group penalty, Sparse Group Lasso Bayesian framework Simulation Studies Illustration: genomics data R package: MSGBSS

3 Integrative Analysis Goal of Integrative Analysis: Wikipedia. Data integration involves combining data residing in different sources and providing users with a unified view of these data. This process becomes significant in a variety of situations, which include both commercial and scientific. System Biology. Integrative Analysis: Analysis of heterogeneous types of data from inter-platform technologies. Goal. Combine multiple types of data: Contribute to a better understanding of biological mechanism. Have the potential to improve the diagnosis and treatments of complex diseases.

4 Example: Data definition p q n X - n observations - p variables Y - n observations - q variables n

5 Example: Data definition p q n X - n observations - p variables Y - n observations - q variables n Omics. Y matrix: gene expression, X matrix: SNP (single nucleotide polymorphism). Many others such as proteomic, metabolomic data. neuroimaging. Y matrix: behavioral variables, X matrix: brain activity (e.g., EEG, fmri, NIRS) neuroimaging genetics. Y matrix: fmri (Fusion of functional magnetic resonance imaging), X matrix: SNP Ecology/Environment. Y matrix: Water quality variables, X matrix: Landscape variables

6 Data: Constraints and Aims Main constraint: situation with p > n

7 Group structures within the data Natural example: Categorical variables which is a group of dummies variables in a regression setting.

8 Group structures within the data Natural example: Categorical variables which is a group of dummies variables in a regression setting. Genomics: genes within the same pathway have similar functions and act together in regulating a biological system. These genes can add up to have a larger effect can be detected as a group (i.e., at a pathway or gene set/module level).

9 Group structures within the data Natural example: Categorical variables which is a group of dummies variables in a regression setting. Genomics: genes within the same pathway have similar functions and act together in regulating a biological system. These genes can add up to have a larger effect can be detected as a group (i.e., at a pathway or gene set/module level). We consider variables are divided into groups: Example p: SNPs grouped into K genes X = [SNP 1,... + SNP } {{ } k SNP k+1, SNP k+2,..., SNP h... SNP } {{ } l+1,..., SNP p ] } {{ } gene 1 gene 2 gene K Example p: genes grouped into K pathways/modules (X j = gene j ) X = [X 1, X 2,..., X } {{ } k X k+1, X k+2,..., X h... X } {{ } l+1, X l+2,..., X p ] } {{ } M 1 M 2 M K

10 Aims in regression setting: Select group variables taking into account the data structures; all the variables within a group are selected otherwise none of them are selected Combine both sparsity of groups and within each group; only relevant variables within a group are selected

11 Some frequentist Appoaches Lasso models: regression model use L 1 and L 2 penalties for performing variable selection

12 Lasso models Univariate model : Y = Xβ + ε, Y R n, β R p Lasso regression: sparse model with L 1 penalty β lasso = argmin β R p n p (Y i X T i β) 2 + λ β j i=1 = argmin Y Xβ 2 2 +λ β 1 β R p } {{ }}{{} Penalty Loss Group Lasso regression: sparse model using group structure with L 2 penalty on group (X g with β g R p g ) min β R p j=1 G G Y X g β g λ pg β g 2 g=1 -If p g = 1 = Lasso method ( β g 2 = g=1 β 2 g = β g )

13 Lasso models Univariate model : Y = G g=1 X gβ g + ε, Y R n, β R p Sparse Group Lasso regression: sparse group model using group structure combining L 1 and L 2 penalties min β R p G G Y X g β g λ 1 pg β g 2 + λ 2 β 1 g=1 g=1

14 Lasso model for Multivariate Y where Y (n q), B (p q) Y = XB + E Sparse model: select predictors associated to the multiple phenotype Y (n q): p min B Y XB 2 F + λ β l 2 where B = β 1 1 β β q 1 β 1 2 β β q 2 l=1, β l = (β 1 l...., β2 l,..., βq l ) β 1 p β 2 p... β q p Implemented in glmnet() Sparse Group model: see e.g., Li et al. (2015), Wand et al. (2012).

15 Bayesian Framework Univariate regression model: Y G g=1 X gβ g N n ( 0, σ 2 I n ) Xu and Ghosh (2015): Bayesian group lasso using spike and slab priors for group variable selection. Rockova and Lesafre (2014): (EM) algorithm for a hierarchical model incorporating grouping information. Stingo et al. (2011): PLS approach for pathway and gene selection using variable selection priors and Markov chain Monte Carlo (MCMC)

16 Bayesian Framework Multivariate regression model: G ( Y X g B g MN n q 0p q, I n, ) Σ g=1 Zhu et al. (2014): Bayesian generalized low rank regression model. Greenlaw et al. (2016): Bayesian hierarchical modeling for imaging genomics data. Limited to Σ = σ 2 I q.

17 Bayesian group lasso model with spike and slab priors Xu and Ghosh (2015) approaches: spike and slab priors providing variable selection at the group level. hierarchical spike and slab prior structure to select variables both at the group level and within each group. Limited to univariate case. Doesn t take into account group size in the penalisation part.

18 Multivariate Bayesian Group Lasso with Spike and Slab prior Y X, B, Σ MN n q (XB, Σ, I n ), (1) Vec(B T g Σ, τ g, π 0 ) ind (1 π 0 )N mg q(0, I mg τ 2 g Σ) + π 0δ 0 (Vec(B T g )), (2) ( ) τ 2 ind mg + 1 g Gamma, λ2, g = 1,..., G, (3) 2 2 Σ IW(d, Q) (4) π 0 Beta(a, b) (5) where δ 0 (Vec(B T g )) denotes a point mass at 0 Rm gq, B g is the m g q regression coefficient matrix for the group g

19 Calibration of λ λ is related to the coefficient of shrinkage. tuned using an empirical Bayes approach. a Monte Carlo EM algorithm is used. the kth EM update for λ is: λ (k) p + G = G g=1 E [ λ τ 2 (k 1) g Y ], in which the posterior expectation of τ 2 g is replaced by the Monte Carlo sample average of τ 2 g generated in the Gibbs sample based on λ (k 1).

20 Median Thresholding Estimator In the case of a block orthogonal design matrix X (i.e., X T i X j = 0 for i j), we have for 1 g G Vec( B T g ) = Vec ( ((X T g X g ) 1 X T g Y) T) N mg q ( Vec(B T g ), (X T g X g ) 1 Σ ). We showed assuming π 0 > c 1+c, then there exists t(π 0) > 0, such that the marginal posterior median of β g satisfies ij Med(β g ij B g ) = 0 for any 1 i m g and 1 j q when Vec( B T g ) 2 < t. The marginal posterior median estimator of the gth group of regression coefficients is zero when the norm of the corresponding block least square estimator is less than a certain threshold.

21 Posterior median as a soft thresholding estimator Assuming an orthogonal design matrix X, i.e., X T X = ni p and consider our model defined with fixed τ 2 g (1 g G). Posterior distribution of B g is a spike and slab distribution, ( Vec(B T g ) X, Y (1 l g)n mgq (1 D g )Vec(B T LS,g ), 1 D ) g I mg Σ + l g δ 0 (Vec(B T g n )), where B LS,g is the least squares estimator of B g, D g = l g = π 0 1, and l 1+nτ 2 g = p(b g = 0 rest) g π 0 + (1 π 0 )(τ 2 mg (q 1) g ) 2 (1 + nτ 2 mg { g ) 2 exp (1 D g )ntr[σ 1 B T LS,g B } LS,g] The marginal posterior distribution of β g ij is also a spike and slab distribution, ( β g ij X, Y (1 l g)n (1 D g ) β g LS,ij, 1 D ) g Σ jj + l g δ 0 (β g n ij ), where Σ jj is the j-th diagonal element of Σ.

22 Posterior median as a soft thresholding estimator The resulting median is a soft thresholding estimator defined by β Med,g ij = Med(β g ij X, Y) = sgn ( β g LS,ij ) (1 D g) β g Σjj LS,ij Q g 1 D g, n + where z + denotes the positive part of z and Q g = φ 1 ( 1 ). 2(1 min( 1 2,l g)) For a univariate response (q = 1) the matrix Σ reduces to the scalar σ 2, and our result matches the previous work of Xu et al (2015). In the multivariate frequentist setting, Li et al. (2015) proposed an iterative algorithm with similar soft thresholding function to incorporate group structure in estimating the regression estimates.

23 Oracle property Let B 0, B 0 g, β 0,g ij denote the true values of B, B g, β g ij, respectively. The index vector of the true model as A = (I( Vec(B g ) 2 0), g = 1,..., G), The index vector of the model selected by certain thresolding estimator B g as A n = (I( Vec( B g ) 2 0), g = 1,..., G). Model selection consistency is attained if and only if lim n P(A n = A) = 1.

24 Oracle property Under orthogonal design the median thresholding estimator has oracle property. Theorem: Assume orthogonal design matrix, i.e., X T X = ni p. Suppose nτ 2 g,n and log(τ 2 g,n )/n 0 as n, for g = 1,..., G, then the median thresholding estimator has oracle property, that is, variable selection consistency, and asymptotic normality, lim n P(AMed n = A) = 1 ( n Vec( B Med A ) Vec(B0 A )) d N(0, Σ I).

25 Gibbs Sampler: full posterior distribution where p(b, τ 2, Σ, π 0 Y, X) p(y B, τ 2, Σ, π 0 ) p(b τ 2, Σ, π 0 ) p(y B, τ 2, Σ, π 0 ) Σ n/2 exp p(b τ 2, Σ, π 0 ) = p(τ 2 ) p(σ) p(π 0 ), { 1 2 Tr [ (Y XB)Σ 1 (Y XB) T ]}, G p(b g τ 2 g, Σ, π 0 ), g=1 p(b g τ 2 g, Σ, π 0 ) (1 π 0 )(2π) qmg 2 (τ 2 qmg g ) 2 Σ mg 1 2 exp [ ] 2τ 2 Tr B g Σ 1 B T g I[B g 0] g +π 0 δ 0 (Vec(B T g )), G p(τ 1,..., τ g ) (λ 2 qmg +1 ) 2 (τ 2 qmg +1 g ) 2 1 exp λ2 m g g=1 p(π 0 ) π a 1 0 (1 π 0 ) b 1, p(σ) Σ d+q+1 2 exp { 1 2 Tr(QΣ 1 ) } 2 τ2 g,

26 Conditional posterior distribution Let B (g) denote the B matrix without the gth group, and X (g) denote the covariate matrix corresponding to B (g), that is, X (g) = (X 1,..., X g 1, X g+1,..., X G ) where X g is the design matrix corresponding to B g. Conditional posterior distribution of B g : spike and slab distribution Conditional posterior distribution of α 2 g = 1 τ 2 g Inverse Gamma if B g = 0 Inverse Gaussian if B g 0 Conditional posterior distribution of Σ: Inverse Wishart Conditional posterior distribution of π 0 : Beta distribution

27 Multivariate Sparse Group Selection with Spike and Slab Prior (MBSGS-SS) Reparametrize the coefficients matrices to tackle the two kinds of sparsity separately: B g = V 1 2 g B g, g = 1,..., G; j = 1,..., m g, with V 1 2 g = diag{τ g1,..., τ gmg }, τ gj 0 and where B g, when nonzero, follow Vec( B T g ) N m g q(0, I mg Σ). Diagonal element of V 1 2 g control the magnitude of the elements of B g.

28 Multivariate Sparse Group Selection with Spike and Slab Prior (MBSGS-SS) To select variables at the group level, we assume the multivariate spike and slab prior for each Vec( B T g ): Vec( B T g Σ, τ g, π 0 ) ind (1 π 0 )N mg q(0, I mg Σ) + π 0 δ 0 (Vec( B T g )) Note that when τ gj = 0, the j-th row of B g is essentially dropped out of the model even when B j g 0. So in order to choose variables within each relevant group, we assume the following spike and slab prior for each τ gj : τ gj ind (1 π 1 )N + (0, s 2 ) +π 1 δ 0 (τ gj ), g = 1,..., G; j = 1,..., m g, where N + (0, s 2 ) denotes a normal N(0, s 2 ) distribution truncated below at 0.

29 Prior Specification We assume an Inverse Wishart prior for Σ IW(d, Q) We assume conjugate beta hyper-priors for π 0 and π 1. We use a conjugate inverse gamma prior for s 2 Inverse Gamma(1, t), We stimate t with the Monte Carlao EM algorithm. For the k-th EM update, t (k) 1 = [ 1 Y ], s 2 E t (k 1) where the posterior expectation of Gibbs samples based on t (k 1). 1 s 2 is estimated from the

30 Gibbs Sampler: full posterior distribution Joint posterior p( B, τ 2, Σ, π 0, π 1, s 2 Y, X) Σ n/2 exp 1 T G G 2 Tr Y X g V 1 2 g B g Σ 1 Y X g V 1 2 g B g g=1 G (1 π 0 )(2π) qmg 2 Σ mg 2 exp g=1 g=1 { 1 [ B 2 Tr g Σ 1 B ] } T g I[ B g 0] +π 0 δ 0 (Vec( B T g )) G m g (1 π 1)2(2πs 2 ) 1 2 exp τ2 gj 2s 2 I[τ gj > 0] + π 1 δ 0 (τ gj ) g=1 j=1 { Σ d+q+1 2 exp 1 } 2 Tr(QΣ 1 ) π a (1 π 0 ) a 2 1 π c 1 1 (1 π 1 1 ) c 2 1 { t(s 2 ) 2 exp t } s 2

31 Simulation Studies Comparison to Lasso, group Lasso, sparse goup Lasso for univariate setting R package glmnet and SGL Comparison to Lasso for Mulivariate setting R package glmnet Bayseian approaches BGL-SS (Bayesian Group Lasso with Spike and Slab prior) BSGS-SS (Bayesian Sparse Group selection with spike and slab priors) MBGL-SS for multivariate setting MBSGL-SS for multivariate setting running iterations in which the first are burn-ins.

32 Simulation results: summary Simulations results suggest that the multivariate Bayesian group lasso with spike and slab prior is strongly influenced by a combination of different group size structures and high correlation between predictors. The multivariate Bayesian sparse group selection with spike and slab prior does not suffer in this situation. Most powerful method for variable selection and prediction performance in the presence of group structure data All numerical results from our article could be reproduced using our R package MBSGS available on CRAN.

33 Computation The current version of our package runs for example: a MBSGS-SS model in around 2 minutes a MBGL-SS model in around 1 minute for a model with 20 groups of 5 variables for a sample size (n = 900) with iterations including for the burnin. Further improvements of the code such as parallelization over the group structure are in progress to speed up the computational time for tacking Big Data sets In this context of genetics studies, some further extensions of our model are under investigation such as integrating different group penalties given a biological prior of the pathways or different distribution priors for each group.

34 Application: Aim and Data Identify a parsimonious set of predictors that explains the joint variability of gene expression in four tissues (adrenal gland, fat, heart, and kidney). 770 SNPs in 29 inbred rats as a predictor matrix (n = 29, p = 770) 29 measured expression levels in the 4 tissues as the outcome (q = 4).

35 Application chromosome information defines the group structure of the predictor matrix ran our MBGL-SS and MBSGS-SS models using this group structure The multivariate lasso selects 69 SNPs which come from the 20 chromosomes. The MBGL-SS selects only the two first groups corresponding to the SNPs from chromosomes 1 and 2.

36 Application: MBSGS-SS results, 32 SNPs from 8 chromosomes

37 Application: results SNP D14Mit (from chromosome 10) has been previously identified Highest estimate (0.334) for the heart tissue the posterior standard deviation of the regression parameter for each selected non-zero median estimate was in the range 0.11 to SNP D14Mit3 estimate was with posterior standard deviation Importance of chromosomes could be investigated using an estimate of the probability of inclusion

38 Concluding remarks: summary Bayesian methods for group-sparse modeling in the context of a multivariate correlated response variable. Our models are based on spike and slab type priors which facilitate variable selection. Importance to include the group size information in the shrinkage part of our model. We have shown that the posterior median estimator could both select and estimate the regression coefficients. Simulation results suggest very good performance of the posterior median estimator for variable selection and prediction error. This estimator obtains similar results as the highest probability model in terms of true and false positive rates.

39 References - Li, Y., Nan, B., and Zhu, J. (2015). *Multivariate Sparse Group Lasso for the Multivariate Multiple Linear Regression with an Arbitrary Group Structure.* Biometrics, 71(2): Liquet, B., Lafaye de Micheaux, P., Hejblum, B., and Thiebaut, R. (2016). *Group and Sparse Group Partial Least Square Approaches Applied in Genomics Context.* Bioinformatics, 3(1): Liquet, B. and Sutton, M. (2016). *MBSGS: Multivariate Bayesian Sparse Group Selection with Spike and Slab*. R package version URL - Liquet, B., Mengersen, K., Pettitt, A. N. and Sutton, M. (2016). *Bayesian Variable Selection Regression Of Multivariate Responses For Group Data* Bayesian Analysis. Volume 12, Number 4 (2017), Received the Lindley award from ISBA Prize Committee, delivered during ISBA World Meeting in Edinburgh Xu, X. and Ghosh, M. (2015). *Bayesian Variable Selection and Estimation for Group Lasso.* Bayesian Analysis, 10(4):

40 Extension to Pleotropic mapping for genome-wide association studies using group variable selection Pleiotropy: genetic variants which affect multiple different complex diseases Example: genetic variants which affect both Breast and Thyroid cancer. Results from GWAS suggest that complex diseases are often affected by many variants with small effects (known as polygenicity) Aims: statistical method to leverage pleiotropic effects incoporate prior pathway knowledge to increase statistical power and identify important risk variants.

41 Model for multiple GWAS studies Suppose we have data from K independent GWAS datasets, D = D 1 D K, where D k = ({y 1, x 1 },..., {y nk, x nk }) y ik {0, 1} denotes the phenotype of the kth study x ik R p is the vector with corresponding p SNPs. Logistic regression model y ik Bernoulli ( g 1 (ν ik ) ) ν ik = x T ik β k for k = 1,..., K, where g( ) is the logistic link function β k R p the regression coefficients for the kth GWAS. Let β j R K, j = 1,..., p, the vector of K regression coefficients corresponding to the jth SNP over the K GWAS.

42 Group Structure SNPs can be partitioned into G groups (genes) Let π g, g = 1,..., G the set of SNPs contained in the gth group with p g = π g. Matrix of all regression coefficients as B = (β 1,..., β K ) = (β 1,..., β p ) T.

43 Frequentist Approach The log likelihood for the combined datasets: p(d B) = K n k ( ) L y ik β T k x ik k=1 i=1 The penalised likelihood estimate B = argmin B R p K where L(x) = ln(1 + e x ) K n k ( ) L y ik β T k x ik + λ1 B + λ G2,1 2 B l2,1 (6) k=1 i=1 G 2,1 -norm penalty B G2,1 = G l 2,1 -norm penalty B l2,1 = p i=1 g=1 i π g K j=1 β2 ik K k=1 β2 ik respectively. The G 2,1 -norm fixes the group structure across studies and encourages sparsity at a gene level. The l 2,1 -norm which allows sparsity within a group.

44 Inference Inference using the alternating direction method of multipliers algorithm (ADMM). Novel approach for identifying pleiotropic effects as it accounts for gene specific and SNP specific effects using a variable selection approach. The method is only capable of producing a point estimate of B and acurrate estimation of the variance for these parameters is not easily given.

45 Bayesian Logistic regression with multivariate spike and slab prior: LogitMBGL-SS Let γ = (γ 1,..., γ p ) T indicate the association status for SNPs where γ j = 1 indicates that the jth SNP is associated to all K traits. Spike and slab prior for the jth SNP β j R K, β j (1 γ j )N K (0, τ 2 j V) + γ jδ 0 (β j ) ( K + 1 τ 2 j Gamma 2, λ ), 2 V IW(d, Q), γ j Bernolli(α 0 ) α 0 Beta(a, b) for j = 1,..., p, where δ 0 (β j ) denotes a point mass at 0 R K. Here, V R K K is a covariance matrix modeling the covariance of the SNP effect on the traits.

46 Extension Should perform well when the SNPs are independent. GWAS datasets: strong correlations that can occur between SNPs within the same gene. Solution: reparameterise the coefficeints to handle the sparsity at a gene grouping level and individual feature level separately. τ R p to model individual sparsity b (g) R pgk with b (g) = (b (g)t,..., b (g)t 1 p g ) where b (g) j group sparsity. R K for β j = τ j b (g) j, where τ j 0, for all j π g.

47 Bayesian Logistic regression using multivariate sparse group selection with spike and slab priors β j = τ j b (g) j, where τ j 0, for all j π g. We assume the following multivariate spike and slab b (g) (1 α 0 )N pg K (0, I pg V) + α 0 δ 0 (b (g) ) τ j (1 α 1 )N ( + 0, s 2) + α 1 δ 0 (τ j ), α 0 Beta(a 1, a 2 ) α 1 Beta(c 1, c 2 ) s 2 InvGamma(1, t) for j π g and g = 1,..., G

48 Signal recovery: (i) gmt: Grouped multi-task penalised logistic regression (λ 1 > 0, λ 2 = 0) using G 2,1 -norm (ii) smt: Sparse multi-task penalid penalised logistic regression (λ 1 = 0, λ 2 > 0 ) using l 2,1 -norm (iii) sgmt: Sparse group multi-task penalised logistic regression (λ 1 > 0, λ 2 > 0 ) (iv) logitmbgl: Bayesian logistic regression using multivariate group lasso with spike and slab prior (v) logitmbsgs: Bayesian logistic regression using multivariate sparse group selection with spike and slab prior

49 Results: Frequentist group gmt Study 1 gmt Study value of coefficient sgmt Study 1 sgmt Study 2 smt Study 1 smt Study 2 Study 1 Study 2 index of coefficient

50 Results: Bayesian group MBGL Study 1 MBGL Study 2 value of coefficient MBSGS Study 1 MBSGS Study 2 Study 1 Study 2 index of coefficient

51 Main Conclusion on the simulation studies The penalised approaches perform reasonably well in variable selection but the reconstructed signal is underestimated. In general, the penalised likelihood methods suffer in terms of false negatives, selecting more variables to be nonzero than the Bayesian methods. The Bayesian methods perform the best in terms of signal recovery measured by the l 1 error and variable selection performance metrics. The penalised likelihood approaches are computationally efficient using alternating direction method of multipliers algorithm Simulation results suggest that when computationally possible the Bayesian estimators should be used. The multivariate Bayesian sparse group selection with spike and slab prior performed the best in terms of signal recovery. The Bayesian method provides a natural method for quantifying the variability of the estimated coefficients.

52 What Next? Application on real data: case/control studies Breast Cancer and Thyroide Cancer Thyroide Cancer (482 case, 463 control) Breast Cancer (1172 case, 1125 control) 6677 SNPs from 618 genes from 10 non-overlapping gene pathways.

53 What Next? Application on real data: case/control studies Breast Cancer and Thyroide Cancer Thyroide Cancer (482 case, 463 control) Breast Cancer (1172 case, 1125 control) 6677 SNPs from 618 genes from 10 non-overlapping gene pathways. Looking for one postdoc position to fill for 2 years granted by la ligue contre le Cancer post-doctorat.html

Bayesian Variable Selection Regression of Multivariate Responses for Group Data

Bayesian Variable Selection Regression of Multivariate Responses for Group Data Bayesian Analysis 2017) 12, Number 4, pp. 1039 1067 Bayesian Variable Selection Regression of Multivariate Responses for Group Data B. Liquet,, K. Mengersen, A. N. Pettitt, and M. Sutton, Abstract. We

More information

Regularization Parameter Selection for a Bayesian Multi-Level Group Lasso Regression Model with Application to Imaging Genomics

Regularization Parameter Selection for a Bayesian Multi-Level Group Lasso Regression Model with Application to Imaging Genomics Regularization Parameter Selection for a Bayesian Multi-Level Group Lasso Regression Model with Application to Imaging Genomics arxiv:1603.08163v1 [stat.ml] 7 Mar 016 Farouk S. Nathoo, Keelin Greenlaw,

More information

Or How to select variables Using Bayesian LASSO

Or How to select variables Using Bayesian LASSO Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO On Bayesian Variable Selection

More information

25 : Graphical induced structured input/output models

25 : Graphical induced structured input/output models 10-708: Probabilistic Graphical Models 10-708, Spring 2013 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Meghana Kshirsagar (mkshirsa), Yiwen Chen (yiwenche) 1 Graph

More information

Estimating Sparse High Dimensional Linear Models using Global-Local Shrinkage

Estimating Sparse High Dimensional Linear Models using Global-Local Shrinkage Estimating Sparse High Dimensional Linear Models using Global-Local Shrinkage Daniel F. Schmidt Centre for Biostatistics and Epidemiology The University of Melbourne Monash University May 11, 2017 Outline

More information

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS023) p.3938 An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Vitara Pungpapong

More information

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables Qi Tang (Joint work with Kam-Wah Tsui and Sijian Wang) Department of Statistics University of Wisconsin-Madison Feb. 8,

More information

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS * Some contents are adapted from Dr. Hung Huang and Dr. Chengkai Li at UT Arlington Mingon Kang, Ph.D. Computer Science, Kennesaw State University Problems

More information

Integrated Anlaysis of Genomics Data

Integrated Anlaysis of Genomics Data Integrated Anlaysis of Genomics Data Elizabeth Jennings July 3, 01 Abstract In this project, we integrate data from several genomic platforms in a model that incorporates the biological relationships between

More information

Scalable MCMC for the horseshoe prior

Scalable MCMC for the horseshoe prior Scalable MCMC for the horseshoe prior Anirban Bhattacharya Department of Statistics, Texas A&M University Joint work with James Johndrow and Paolo Orenstein September 7, 2018 Cornell Day of Statistics

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Expression Data Exploration: Association, Patterns, Factors & Regression Modelling

Expression Data Exploration: Association, Patterns, Factors & Regression Modelling Expression Data Exploration: Association, Patterns, Factors & Regression Modelling Exploring gene expression data Scale factors, median chip correlation on gene subsets for crude data quality investigation

More information

Bayesian construction of perceptrons to predict phenotypes from 584K SNP data.

Bayesian construction of perceptrons to predict phenotypes from 584K SNP data. Bayesian construction of perceptrons to predict phenotypes from 584K SNP data. Luc Janss, Bert Kappen Radboud University Nijmegen Medical Centre Donders Institute for Neuroscience Introduction Genetic

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Probabilistic machine learning group, Aalto University Bayesian theory and methods, approximative integration, model

Probabilistic machine learning group, Aalto University  Bayesian theory and methods, approximative integration, model Aki Vehtari, Aalto University, Finland Probabilistic machine learning group, Aalto University http://research.cs.aalto.fi/pml/ Bayesian theory and methods, approximative integration, model assessment and

More information

Shrinkage Methods: Ridge and Lasso

Shrinkage Methods: Ridge and Lasso Shrinkage Methods: Ridge and Lasso Jonathan Hersh 1 Chapman University, Argyros School of Business hersh@chapman.edu February 27, 2019 J.Hersh (Chapman) Ridge & Lasso February 27, 2019 1 / 43 1 Intro and

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Bayesian Inference of Multiple Gaussian Graphical Models

Bayesian Inference of Multiple Gaussian Graphical Models Bayesian Inference of Multiple Gaussian Graphical Models Christine Peterson,, Francesco Stingo, and Marina Vannucci February 18, 2014 Abstract In this paper, we propose a Bayesian approach to inference

More information

Variable Selection in Structured High-dimensional Covariate Spaces

Variable Selection in Structured High-dimensional Covariate Spaces Variable Selection in Structured High-dimensional Covariate Spaces Fan Li 1 Nancy Zhang 2 1 Department of Health Care Policy Harvard University 2 Department of Statistics Stanford University May 14 2007

More information

25 : Graphical induced structured input/output models

25 : Graphical induced structured input/output models 10-708: Probabilistic Graphical Models 10-708, Spring 2016 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Raied Aljadaany, Shi Zong, Chenchen Zhu Disclaimer: A large

More information

Bayesian Grouped Horseshoe Regression with Application to Additive Models

Bayesian Grouped Horseshoe Regression with Application to Additive Models Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu 1,2, Daniel F. Schmidt 1, Enes Makalic 1, Guoqi Qian 2, John L. Hopper 1 1 Centre for Epidemiology and Biostatistics,

More information

Consistent high-dimensional Bayesian variable selection via penalized credible regions

Consistent high-dimensional Bayesian variable selection via penalized credible regions Consistent high-dimensional Bayesian variable selection via penalized credible regions Howard Bondell bondell@stat.ncsu.edu Joint work with Brian Reich Howard Bondell p. 1 Outline High-Dimensional Variable

More information

Multidimensional data analysis in biomedicine and epidemiology

Multidimensional data analysis in biomedicine and epidemiology in biomedicine and epidemiology Katja Ickstadt and Leo N. Geppert Faculty of Statistics, TU Dortmund, Germany Stakeholder Workshop 12 13 December 2017, PTB Berlin Supported by Deutsche Forschungsgemeinschaft

More information

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract Bayesian analysis of a vector autoregressive model with multiple structural breaks Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus Abstract This paper develops a Bayesian approach

More information

ESL Chap3. Some extensions of lasso

ESL Chap3. Some extensions of lasso ESL Chap3 Some extensions of lasso 1 Outline Consistency of lasso for model selection Adaptive lasso Elastic net Group lasso 2 Consistency of lasso for model selection A number of authors have studied

More information

Package LBLGXE. R topics documented: July 20, Type Package

Package LBLGXE. R topics documented: July 20, Type Package Type Package Package LBLGXE July 20, 2015 Title Bayesian Lasso for detecting Rare (or Common) Haplotype Association and their interactions with Environmental Covariates Version 1.2 Date 2015-07-09 Author

More information

Pre-Selection in Cluster Lasso Methods for Correlated Variable Selection in High-Dimensional Linear Models

Pre-Selection in Cluster Lasso Methods for Correlated Variable Selection in High-Dimensional Linear Models Pre-Selection in Cluster Lasso Methods for Correlated Variable Selection in High-Dimensional Linear Models Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider variable

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Nonparametric Bayes tensor factorizations for big data

Nonparametric Bayes tensor factorizations for big data Nonparametric Bayes tensor factorizations for big data David Dunson Department of Statistical Science, Duke University Funded from NIH R01-ES017240, R01-ES017436 & DARPA N66001-09-C-2082 Motivation Conditional

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Statistical Inference

Statistical Inference Statistical Inference Liu Yang Florida State University October 27, 2016 Liu Yang, Libo Wang (Florida State University) Statistical Inference October 27, 2016 1 / 27 Outline The Bayesian Lasso Trevor Park

More information

Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate

Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Lucas Janson, Stanford Department of Statistics WADAPT Workshop, NIPS, December 2016 Collaborators: Emmanuel

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Feature Selection via Block-Regularized Regression

Feature Selection via Block-Regularized Regression Feature Selection via Block-Regularized Regression Seyoung Kim School of Computer Science Carnegie Mellon University Pittsburgh, PA 3 Eric Xing School of Computer Science Carnegie Mellon University Pittsburgh,

More information

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Genet. Sel. Evol. 33 001) 443 45 443 INRA, EDP Sciences, 001 Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Louis Alberto GARCÍA-CORTÉS a, Daniel SORENSEN b, Note a

More information

Hypes and Other Important Developments in Statistics

Hypes and Other Important Developments in Statistics Hypes and Other Important Developments in Statistics Aad van der Vaart Vrije Universiteit Amsterdam May 2009 The Hype Sparsity For decades we taught students that to estimate p parameters one needs n p

More information

Large-scale Ordinal Collaborative Filtering

Large-scale Ordinal Collaborative Filtering Large-scale Ordinal Collaborative Filtering Ulrich Paquet, Blaise Thomson, and Ole Winther Microsoft Research Cambridge, University of Cambridge, Technical University of Denmark ulripa@microsoft.com,brmt2@cam.ac.uk,owi@imm.dtu.dk

More information

Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference

Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference Shunsuke Horii Waseda University s.horii@aoni.waseda.jp Abstract In this paper, we present a hierarchical model which

More information

Coupled Hidden Markov Models: Computational Challenges

Coupled Hidden Markov Models: Computational Challenges .. Coupled Hidden Markov Models: Computational Challenges Louis J. M. Aslett and Chris C. Holmes i-like Research Group University of Oxford Warwick Algorithms Seminar 7 th March 2014 ... Hidden Markov

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

A Bayesian Nonparametric Model for Predicting Disease Status Using Longitudinal Profiles

A Bayesian Nonparametric Model for Predicting Disease Status Using Longitudinal Profiles A Bayesian Nonparametric Model for Predicting Disease Status Using Longitudinal Profiles Jeremy Gaskins Department of Bioinformatics & Biostatistics University of Louisville Joint work with Claudio Fuentes

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Part 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior

Part 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior Tom Heskes joint work with Marcel van Gerven

More information

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010 Model-Averaged l 1 Regularization using Markov Chain Monte Carlo Model Composition Technical Report No. 541 Department of Statistics, University of Washington Chris Fraley and Daniel Percival August 22,

More information

On Bayesian Computation

On Bayesian Computation On Bayesian Computation Michael I. Jordan with Elaine Angelino, Maxim Rabinovich, Martin Wainwright and Yun Yang Previous Work: Information Constraints on Inference Minimize the minimax risk under constraints

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

Generalized Elastic Net Regression

Generalized Elastic Net Regression Abstract Generalized Elastic Net Regression Geoffroy MOURET Jean-Jules BRAULT Vahid PARTOVINIA This work presents a variation of the elastic net penalization method. We propose applying a combined l 1

More information

A new Hierarchical Bayes approach to ensemble-variational data assimilation

A new Hierarchical Bayes approach to ensemble-variational data assimilation A new Hierarchical Bayes approach to ensemble-variational data assimilation Michael Tsyrulnikov and Alexander Rakitko HydroMetCenter of Russia College Park, 20 Oct 2014 Michael Tsyrulnikov and Alexander

More information

STA 216, GLM, Lecture 16. October 29, 2007

STA 216, GLM, Lecture 16. October 29, 2007 STA 216, GLM, Lecture 16 October 29, 2007 Efficient Posterior Computation in Factor Models Underlying Normal Models Generalized Latent Trait Models Formulation Genetic Epidemiology Illustration Structural

More information

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined

More information

Bayesian methods in economics and finance

Bayesian methods in economics and finance 1/26 Bayesian methods in economics and finance Linear regression: Bayesian model selection and sparsity priors Linear Regression 2/26 Linear regression Model for relationship between (several) independent

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

GWAS IV: Bayesian linear (variance component) models

GWAS IV: Bayesian linear (variance component) models GWAS IV: Bayesian linear (variance component) models Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS IV: Bayesian

More information

November 2002 STA Random Effects Selection in Linear Mixed Models

November 2002 STA Random Effects Selection in Linear Mixed Models November 2002 STA216 1 Random Effects Selection in Linear Mixed Models November 2002 STA216 2 Introduction It is common practice in many applications to collect multiple measurements on a subject. Linear

More information

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage Lingrui Gan, Naveen N. Narisetty, Feng Liang Department of Statistics University of Illinois at Urbana-Champaign Problem Statement

More information

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I. Vector Autoregressive Model Vector Autoregressions II Empirical Macroeconomics - Lect 2 Dr. Ana Beatriz Galvao Queen Mary University of London January 2012 A VAR(p) model of the m 1 vector of time series

More information

Aspects of Feature Selection in Mass Spec Proteomic Functional Data

Aspects of Feature Selection in Mass Spec Proteomic Functional Data Aspects of Feature Selection in Mass Spec Proteomic Functional Data Phil Brown University of Kent Newton Institute 11th December 2006 Based on work with Jeff Morris, and others at MD Anderson, Houston

More information

Bayesian Inference of Interactions and Associations

Bayesian Inference of Interactions and Associations Bayesian Inference of Interactions and Associations Jun Liu Department of Statistics Harvard University http://www.fas.harvard.edu/~junliu Based on collaborations with Yu Zhang, Jing Zhang, Yuan Yuan,

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Chapter 17: Undirected Graphical Models

Chapter 17: Undirected Graphical Models Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)

More information

Selection-adjusted estimation of effect sizes

Selection-adjusted estimation of effect sizes Selection-adjusted estimation of effect sizes with an application in eqtl studies Snigdha Panigrahi 19 October, 2017 Stanford University Selective inference - introduction Selective inference Statistical

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public

More information

Bayes methods for categorical data. April 25, 2017

Bayes methods for categorical data. April 25, 2017 Bayes methods for categorical data April 25, 2017 Motivation for joint probability models Increasing interest in high-dimensional data in broad applications Focus may be on prediction, variable selection,

More information

Graphical Models for Collaborative Filtering

Graphical Models for Collaborative Filtering Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,

More information

Nearest Neighbor Gaussian Processes for Large Spatial Data

Nearest Neighbor Gaussian Processes for Large Spatial Data Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns

More information

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal

More information

An algorithm for the multivariate group lasso with covariance estimation

An algorithm for the multivariate group lasso with covariance estimation An algorithm for the multivariate group lasso with covariance estimation arxiv:1512.05153v1 [stat.co] 16 Dec 2015 Ines Wilms and Christophe Croux Leuven Statistics Research Centre, KU Leuven, Belgium Abstract

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Cross-sectional space-time modeling using ARNN(p, n) processes

Cross-sectional space-time modeling using ARNN(p, n) processes Cross-sectional space-time modeling using ARNN(p, n) processes W. Polasek K. Kakamu September, 006 Abstract We suggest a new class of cross-sectional space-time models based on local AR models and nearest

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

27: Case study with popular GM III. 1 Introduction: Gene association mapping for complex diseases 1

27: Case study with popular GM III. 1 Introduction: Gene association mapping for complex diseases 1 10-708: Probabilistic Graphical Models, Spring 2015 27: Case study with popular GM III Lecturer: Eric P. Xing Scribes: Hyun Ah Song & Elizabeth Silver 1 Introduction: Gene association mapping for complex

More information

On Gaussian Process Models for High-Dimensional Geostatistical Datasets

On Gaussian Process Models for High-Dimensional Geostatistical Datasets On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May

More information

Scale Mixture Modeling of Priors for Sparse Signal Recovery

Scale Mixture Modeling of Priors for Sparse Signal Recovery Scale Mixture Modeling of Priors for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Jason Palmer, Zhilin Zhang and Ritwik Giri Outline Outline Sparse

More information

Proteomics and Variable Selection

Proteomics and Variable Selection Proteomics and Variable Selection p. 1/55 Proteomics and Variable Selection Alex Lewin With thanks to Paul Kirk for some graphs Department of Epidemiology and Biostatistics, School of Public Health, Imperial

More information

Learning Gaussian Graphical Models with Unknown Group Sparsity

Learning Gaussian Graphical Models with Unknown Group Sparsity Learning Gaussian Graphical Models with Unknown Group Sparsity Kevin Murphy Ben Marlin Depts. of Statistics & Computer Science Univ. British Columbia Canada Connections Graphical models Density estimation

More information

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n

More information

Strong Lens Modeling (II): Statistical Methods

Strong Lens Modeling (II): Statistical Methods Strong Lens Modeling (II): Statistical Methods Chuck Keeton Rutgers, the State University of New Jersey Probability theory multiple random variables, a and b joint distribution p(a, b) conditional distribution

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Package IGG. R topics documented: April 9, 2018

Package IGG. R topics documented: April 9, 2018 Package IGG April 9, 2018 Type Package Title Inverse Gamma-Gamma Version 1.0 Date 2018-04-04 Author Ray Bai, Malay Ghosh Maintainer Ray Bai Description Implements Bayesian linear regression,

More information

Bi-level feature selection with applications to genetic association

Bi-level feature selection with applications to genetic association Bi-level feature selection with applications to genetic association studies October 15, 2008 Motivation In many applications, biological features possess a grouping structure Categorical variables may

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Monte Carlo in Bayesian Statistics

Monte Carlo in Bayesian Statistics Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview

More information

Partial factor modeling: predictor-dependent shrinkage for linear regression

Partial factor modeling: predictor-dependent shrinkage for linear regression modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework

More information

MULTILEVEL IMPUTATION 1

MULTILEVEL IMPUTATION 1 MULTILEVEL IMPUTATION 1 Supplement B: MCMC Sampling Steps and Distributions for Two-Level Imputation This document gives technical details of the full conditional distributions used to draw regression

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information

The STS Surgeon Composite Technical Appendix

The STS Surgeon Composite Technical Appendix The STS Surgeon Composite Technical Appendix Overview Surgeon-specific risk-adjusted operative operative mortality and major complication rates were estimated using a bivariate random-effects logistic

More information

Package horseshoe. November 8, 2016

Package horseshoe. November 8, 2016 Title Implementation of the Horseshoe Prior Version 0.1.0 Package horseshoe November 8, 2016 Description Contains functions for applying the horseshoe prior to highdimensional linear regression, yielding

More information

wissen leben WWU Münster

wissen leben WWU Münster MÜNSTER Sparsity Constraints in Bayesian Inversion Inverse Days conference in Jyväskylä, Finland. Felix Lucka 19.12.2012 MÜNSTER 2 /41 Sparsity Constraints in Inverse Problems Current trend in high dimensional

More information

Integrated Non-Factorized Variational Inference

Integrated Non-Factorized Variational Inference Integrated Non-Factorized Variational Inference Shaobo Han, Xuejun Liao and Lawrence Carin Duke University February 27, 2014 S. Han et al. Integrated Non-Factorized Variational Inference February 27, 2014

More information

Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis

Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Jeffrey S. Morris University of Texas, MD Anderson Cancer Center Joint wor with Marina Vannucci, Philip J. Brown,

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology

More information

Bayesian Grouped Horseshoe Regression with Application to Additive Models

Bayesian Grouped Horseshoe Regression with Application to Additive Models Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu, Daniel F. Schmidt, Enes Makalic, Guoqi Qian, and John L. Hopper Centre for Epidemiology and Biostatistics, Melbourne

More information

Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA)

Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Willem van den Boom Department of Statistics and Applied Probability National University

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information