Bayesian Learning of Sparse Gaussian Graphical Models. Duke University Durham, NC, USA. University of South Carolina Columbia, SC, USA I.

Similar documents
BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage

Robust Inverse Covariance Estimation under Noisy Measurements

Learning Gaussian Graphical Models with Unknown Group Sparsity

Approximation. Inderjit S. Dhillon Dept of Computer Science UT Austin. SAMSI Massive Datasets Opening Workshop Raleigh, North Carolina.

Sparse Permutation Invariant Covariance Estimation: Motivation, Background and Key Results

An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss

Permutation-invariant regularization of large covariance matrices. Liza Levina

Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference

Variational Principal Components

High-dimensional covariance estimation based on Gaussian graphical models

Learning Multiple Tasks with a Sparse Matrix-Normal Penalty

Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation

Inverse Covariance Estimation with Missing Data using the Concave-Convex Procedure

Gaussian Graphical Models and Graphical Lasso

High-dimensional Covariance Estimation Based On Gaussian Graphical Models

High-dimensional graphical model selection: Practical and information-theoretic limits

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models

Robust and sparse Gaussian graphical modelling under cell-wise contamination

Sparse Gaussian conditional random fields

Expectation Propagation Algorithm

Sparse inverse covariance estimation with the lasso

The Variational Gaussian Approximation Revisited

Regularized Parameter Estimation in High-Dimensional Gaussian Mixture Models

Lecture 3. Linear Regression II Bastian Leibe RWTH Aachen

Nonparametric Factor Analysis with Beta Process Priors

Or How to select variables Using Bayesian LASSO

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

High-dimensional graphical model selection: Practical and information-theoretic limits

Large-scale Ordinal Collaborative Filtering

Pattern Recognition and Machine Learning

Integrated Non-Factorized Variational Inference

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

Regularization Parameter Selection for a Bayesian Multi-Level Group Lasso Regression Model with Application to Imaging Genomics

Sparse and Locally Constant Gaussian Graphical Models

Automatic Rank Determination in Projective Nonnegative Matrix Factorization

Study Notes on the Latent Dirichlet Allocation

STA414/2104 Statistical Methods for Machine Learning II

Scale Mixture Modeling of Priors for Sparse Signal Recovery

Probabilistic Graphical Models

Comment on Article by Scutari

Data Mining Techniques

10708 Graphical Models: Homework 2

Estimation of Graphical Models with Shape Restriction

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

Group Sparse Non-negative Matrix Factorization for Multi-Manifold Learning

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

Statistical Inference

Learning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text

Undirected Graphical Models

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

25 : Graphical induced structured input/output models

A note on the group lasso and a sparse group lasso

Bayesian Grouped Horseshoe Regression with Application to Additive Models

Statistical Data Mining and Machine Learning Hilary Term 2016

Sparse Graph Learning via Markov Random Fields

Nonlinear Statistical Learning with Truncated Gaussian Graphical Models

Graphical Model Selection

Generalized Elastic Net Regression

Learning with L q<1 vs L 1 -norm regularisation with exponentially many irrelevant features

STA 4273H: Statistical Machine Learning

Bayesian Inference Course, WTCN, UCL, March 2013

Group Sparse Priors for Covariance Estimation

Introduction to graphical models: Lecture III

Bayesian linear regression

Shrinkage Tuning Parameter Selection in Precision Matrices Estimation

Probabilistic Graphical Models

Learning Binary Classifiers for Multi-Class Problem

Algorithmisches Lernen/Machine Learning

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables

Inequalities on partial correlations in Gaussian graphical models

Robust Inverse Covariance Estimation under Noisy Measurements

Machine Learning for Signal Processing Bayes Classification

Expectation Propagation for Approximate Bayesian Inference

Sparse Covariance Selection using Semidefinite Programming

Least Absolute Shrinkage is Equivalent to Quadratic Penalization

The lasso: some novel algorithms and applications

Graphical Models for Collaborative Filtering

An Improved 1-norm SVM for Simultaneous Classification and Variable Selection

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Learning discrete graphical models via generalized inverse covariance matrices

STA 4273H: Statistical Machine Learning

Sparse representation classification and positive L1 minimization

Chapter 17: Undirected Graphical Models

Bayesian Methods for Sparse Signal Recovery

Analysis Methods for Supersaturated Design: Some Comparisons

Bias-Variance Trade-Off in Hierarchical Probabilistic Models Using Higher-Order Feature Interactions

Convolutional Dictionary Learning and Feature Design

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Chris Bishop s PRML Ch. 8: Graphical Models

Least Squares Regression

11. Learning graphical models

Nonparametric Bayesian Methods (Gaussian Processes)

STA 4273H: Statistical Machine Learning

SINGLE-TASK AND MULTITASK SPARSE GAUSSIAN PROCESSES

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

arxiv: v1 [math.st] 13 Feb 2012

Consistent high-dimensional Bayesian variable selection via penalized credible regions

Behavioral Data Mining. Lecture 7 Linear and Logistic Regression

Transcription:

Bayesian Learning of Sparse Gaussian Graphical Models Minhua Chen, Hao Wang, Xuejun Liao and Lawrence Carin Electrical and Computer Engineering Department Duke University Durham, NC, USA Statistics Department University of South Carolina Columbia, SC, USA Abstract Sparse inverse covariance matrix modeling is an important tool for learning relationships among different variables in a Gaussian graph. Most existing algorithms are based on l regularization, with the regularization parameters tuned via cross-validation. In this paper, a Bayesian formulation of the problem is proposed, where the regularization parameters are inferred adaptively and cross-validation is avoided. Variational Bayes (VB) is used for the model inference. Results on simulated and real datasets validate the proposed approach. In addition, a graph extension algorithm is proposed to include a new variable in an existing graph, which can be used when separate testing data are available. I. INTRODUCTION There is significant interest in using limited data to infer relationships between different variables. A Gaussian Graphical Model (GGM) [] is an effective way to model such dependencies. Let X i R p be a p-dimensional random vector drawn from a multivariate Gaussian distribution, i.e., p(x i ) = N(X i ;,J ), i =,,,n. Without loss of generality, we assume the mean is subtracted in a pre-processing step. Here J is the inverse covariance matrix, also called a precision matrix or concentration matrix. According to the Hammersley-Clifford Theorem [], the conditional dependency of the variables in X i is encoded in the sparsity pattern of the inverse covariance matrix J. To see this more clearly, the distribution of a certain variable X si conditioned on the remaining variables X si can be written as p(x si X si ) = N(X si ; J ss J ss X si,j ss ), () where s denotes all variable indices excluding s, and J ss denotes the s th column in J excluding the element J ss. If some elements in J ss are zero, then the corresponding variables are

conditionally independent of variable s given all other variables, according to (). Inspired by the l -based sparse modeling in signal processing, sparse inverse covariance matrix estimation algorithms based on l regularization have been proposed recently [] [7]. In [], a neighborhood selection algorithm with LASSO regression [8] is applied to each node in the graph, reducing the original problem to multiple sparse linear regression problems. In [], an l penalty is directly imposed on the inverse covariance matrix to favor sparsity, so that the conditional independencies of the variables can be effectively identified. Motivated by the work on adaptive LASSO for regression [9], an adaptive LASSO penalty has also been considered for sparse inverse covariance matrix estimation []. It is shown [] that the adaptive version enjoys some desirable properties and gives better results. The idea of using adaptive LASSO penalty can also be found in a recent paper []. The difference between [], [] and our work is that we take a Bayesian approach, and our algorithm is stand-alone, while the above-mentioned methods require usage of the basic graphical LASSO algorithm from []. In another line of research, sparse Bayesian learning has been applied successfully in many areas, for example, compressive sensing and dictionary learning. The l regularization is equivalent to imposing a Laplacian prior on the variables to be estimated, which yields the Bayesian LASSO model [] for linear regression. In addition, a Bayesian adaptive LASSO model was proposed in [3] to enable adaptive shrinkage and avoid biased estimation. We propose a Bayesian hierarchical model for sparse inverse covariance matrix learning. Bayesian adaptive LASSO priors are imposed on off-diagonal elements of the inverse covariance matrix with different regularization parameters. The positive-definiteness constraint makes the estimation of a sparse inverse covariance matrix challenging. A first-order condition was derived in [] under which positive definiteness is guaranteed. This condition arises naturally in our Bayesian inference algorithm, hence positive definiteness is guaranteed in our algorithm. The objective function in [] is similar to our model. The main difference is that a Bayesian framework is used in our model, while an optimization approach is used in []. As for most optimization approaches, regularization parameters have to be chosen through cross-validation in []. This tuning process is avoided in Bayesian models, where sparseness-promoting priors with non-informative hyper-parameters are imposed. Bayesian hierarchical models have been adopted to infer a sparse GGM. In [] a dependency network was constructed by predicting each variable using a small number of other variables.

A Bayesian variable selection technique was used to select the predictors. This approach is equivalent to sparse Cholesky decomposition modeling of the inverse covariance matrix. A similar idea was exploited in [6]. In contrast, the Bayesian model proposed in this paper directly models the inverse covariance matrix. Due to the property of the Gaussian distribution, the dependency network can be expressed directly as in (), without the need for the Cholesky component. Another Bayesian model proposed recently for sparse GGM learning is [7]. This is a twostep approach. In the first step, a Bayesian hierarchical regression model was proposed to learn the dependency network, with sparseness-promoting priors imposed on the edges. In the second step, with the inferred graph structure, a standard optimization method (graphical LASSO []) was used to learn the inverse covariance matrix. There are two main differences between this approach and our approach. First, as mentioned in their paper, the Bayesian model used in the first step is a pseudo-likelihood model rather than a proper generative model for the data. However, our model is a valid generative model for the data. Secondly, as mentioned at the end of their paper, a future direction of their work is to combine the two-step process into one. Our model addresses this challenge, as we learn the structure of the graph and the inverse covariance matrix simultaneously, in one step. A unique aspect of our paper is that we study the graph-extension problem. Suppose a p p inverse covariance matrix is inferred using training data. Now given an additional testing variable or node, the question is how to grow the matrix to dimension (p+) (p+). In order for the extended graph to be consistent with the original graph, the covariance matrix for the original p variables is fixed as the training result, while a sparseness constraint is imposed on the extended part of the inverse covariance matrix. The resulting algorithm is a sparse regression for the additional variable using the original variables, with positive definiteness constraints. The remainder of the paper is organized as follows. In Section II we provide detailed specifications for the Bayesian Gaussian Graphical Model, and discuss variational Bayesian inference and computational details. In Section III we discuss model properties, extensions, and relationships to existing models. An extensive set of experimental results are presented in Section IV, followed by conclusions in Section V. 3

II. MODEL SPECIFICATION AND VARIATIONAL BAYESIAN INFERENCE A. Model construction Define X = [X,X,,X n ] R p n to be the data matrix with p variables and n sample points. A joint density function for the problem can be expressed as p(x,j) = p(x J)p(J) = n N(X i ;,J )p(j). () i= A Laplacian prior is imposed on the off-diagonal elements J st (s t) to induce sparsity []: p(j) = p s= t<s τγts exp( τγ ts J ts ) (3) with symmetry constraint J st = J ts ; τ is a global regularization parameter, and it is generally fixed to be one []. It is important to note that a separate parameter γ ts is used on each component of J, and therefore the components of J are not drawn i.i.d. from a single Laplace prior, but rather non-i.i.d., with component dependent parameters γ ts. This non-i.i.d. construction mitigates recent observations about the inconsistency of l regularizer (Lasso) [8] and the fact that an i.i.d. Laplace prior does not promote sparsity. The analog to using separate γ ts on each component of J is related to weighted (or adaptive) Lasso [9]. The log likelihood of the model in () can be expressed as ( ) logp(x,j) n p logdet(j) tr(σj) τγst J st n with Σ = n n i= X ix T i the sample covariance. This shows that our model is effectively imposing a weighted l regularization on the off-diagonal elements of J [], [7]. Different regularization parameters (γ st ) are used for J to foster adaptive regularization [9], []. Besides the Laplacian shrinkage prior, other priors could also be used, for example, the Student-t prior [9] and spike-and-slab prior []. The Laplacian prior in (3) is not analytic for Bayesian inference. Following [], we use a scale mixture of normals representation to avoid this problem: τγts exp( τγ ts J ts ) = N(J ts ;,τ αts )InvGa(α ts ;, γ ts )dα ts () where InvGa(x;g,h) = hg Γ(g) x g exp( h ) (x > ) denotes the inverse gamma distribution. x s= t<s ()

In addition, a gamma prior is imposed on γ ts to infer this regularization parameter adaptively within the Bayesian framework. In contrast, this parameter is found via cross-validation in the optimization based approaches [], []. The full model can be expressed as p(x,j,α,γ) = p(x J)p(J α)p(α γ)p(γ) (J S + )/C n p = N(X i ;,J ) N(J ts ;,τ αts )InvGa(α ts ;, γ ts )Ga(γ ts;a,b ) (J S + )/C i= s= t<s where (J S + ) defines a feasible region of the set of positive definite matrices, and C = p(j α)p(α γ)p(γ) (J S+ )djdαdγ is the normalizing constant. In practice, we do not need to evaluate this constant when performing VB inference. The Normal-InvGa- Ga prior on J st is called Bayesian adaptive LASSO, or Normal-Exponential-Gamma (NEG) prior [3]. Hyperparameters are set as a = b = 6, thereby imposing a flat (essentially non-informative) prior. (6) B. Variational Bayesian Inference We seek a variational Bayesian (VB) approximation to the posterior distribution []. In a variational Bayesian analysis, one approximates the posterior distribution as the product (factorization) of distributions with associated hyperparameters, and one optimizes these hyperparameters to minimize the Kullback-Leibler divergence between the actual posterior distribution and the factorized approximation. This minimization can be achieved by employing a variational representation, with details discussed in []. In this paper we employ the following factorization p(j,α,γ X) =. p δ(j ss Jss )δ(j ss J ss )q(α ss)q(γ ss ) (7) s= where again s denotes all variable indices except s. A point estimate (hence the delta functions in (7)) is used for the inverse covariance matrix, for convenience of inference. The density functions q(α ss ) and q(γ ss ) correspond to q(α ss ) = InvGa(α ts ;a ts,b ts ) (8) t<s q(γ ss ) = t<s Ga(γ ts ;c ts,d ts ) (9)

and we seek to infer the hyperparameters {a ts,b ts,c ts,d ts }, performed via an iterative procedure [] that is guaranteed to converge. Using identity N(X i ;,J ) = N(X si ; J s s J ssx si,j s s ) N(X si;,(j ss J ss J s s J ss) ) () the VB update equations are ) Update for J: (Jss,J ss ) = arg max J ss,j ss ( n log i= N(X si ; J s s J ssx si,j s s ) ) N(X si ;,(J ss J ss J s s J ss) ) N(J ss ;,τ diag (α ss )) (J S + ) q(α ss )δ(j s s J s s)dα ss dj s s () We first consider the unconstrained optimization problem without the term (J S + ). By taking derivatives with respect to (J ss,j ss ) simultaneously, we can obtain J ss = (nσ ssj s s +τdiag( α ss )) ( nσ ss ) () J ss = Σ ss +J ss J s s J ss (3) We then find that the above solution always lies in the feasible region (J S + ). See Section III-A for details. ) Update for α: q(α ss ) exp( log(n(j ss ;,τ diag (α ss )) = t s t s InvGaussian(α ts ; InvGa(α ts ;, γ ts )) q(γ ss)δ(j ss J ss)dγ ss dj ss ) γ ts τj ts, γ ts ) where h InvGaussian(x; g, h) = πx 3 exp( h(x g) ) g x 6

(x > ) denotes the inverse Gaussian distribution with mean x = g and x = g +h. 3) Update for γ: ( q(γ ss ) exp log( = t s t s Ga(γ ts ;a +,b + α ts ) InvGa(α ts ;, γ ts ) Ga(γ ts;a,b ))q(α ss )dα ss ) where Ga(x;g,h) = hg Γ(g) xg exp( hx) (x > ) denotes the Gamma distribution with x = g h. The above iterations only update the sth column of J, and we sweep all columns to update the whole inverse covariance matrix. The symbol on J is omitted for brevity in the following discussion. C. Computational details Computation of J s s is required in () in each iteration. However, by maintaining a global covariance matrix W = J, this matrix inversion can be avoided. The following matrix inversion identity is used to achieve this: A A A A B = A A B A A D with B = A A A A and D = A A A A. We can obtain J s s from the global covariance matrix W via identity J s s = W s s W ss W ss W ss and after applying () and (3), we can efficiently update W using the new J ss and J ss via identities: D W ss = (J ss J ssj s s J ss) ; W ss = J s s J ssw ss ; W s s = J s s +W ssw ss W ss Although the inversion is avoided, we do need to solve the linear systems in (). The overall algorithm is summarized in Algorithm. 7

Algorithm Bayesian Learning of Sparse GGM Input: Sample covariance Σ, sample size n. Initialize J, W = J, α and γ. repeat for s = to p do J s s = W s s W ss Wss W ss. J ss = (nσ ss J s s +τdiag( α ss )) ( nσ ss ). J ss = Σ ss +J ssj s s J ss. W ss = (J ss J ss J s s J ss). W ss = J W s s = J s s s s J ssw ss. +W ssw ss W ss. γ ts q(α ss ) = t s InvGaussian(α ts; τj ts, γ ts ). q(γ ss ) = t s Ga(γ ts;a +,b + α ts ). end for until Maximum number of iteration is reached. III. PROPERTIES AND EXTENSIONS A. Proof of Positive Definiteness and Convergence In the inference, matrix J is always symmetric since we imposed this symmetry constraint in the model. Equation (3) in the VB inference coincides with the first-order condition obtained in [], which guarantees positive definiteness of J. This is proved in the following theorem. Theorem. Let J () be initially a symmetric positive definite matrix, and Σ ss >, s. Then the variational Bayesian algorithm yields a sequence of updates J (),J (),J (),...,J (t),... that converges to a positive definite matrix J. Proof: In each iteration, whenever a certain column of J is updated using () and (3), the updated J can be written as J new = J s s J ss Σ J ss ss +J ssj s s J ss = J s s J J ss s s (J s s J ss ) + () Σ ss By induction, J s s is positive definite. Also, Σ ss >, as Σ ss is the sample variance on the s-th coordinate. Hence J new is nonnegative definite. However, det(j new ) = det(j s s )Σ ss >. Therefore, J new is positive definite. This proves that positive definiteness is preserved through all iterations. The variational Bayesian updates constitute a block coordinate descent that maximizes a lower 8

bound to the marginalized likelihood function. Because the solution of (J, q(α), q(γ)) in each iteration is unique, by general results on block coordinate descent [], the solutions must converge. As all updates of J are positive definite, the updates must also converge to a positive definite matrix. Q.E.D. B. Relation to Other Methods as According to () and (6), if we marginalize α, the proposed Bayesian model can be expressed p(x,j,γ) = n p N(X i ;,J ) i= s= t<s τγts exp( τγ ts J ts )Ga(γ ts ;a,b ) () and the log likelihood logp(x,j,γ) is the same as () except that we have an additional term for the prior on the weights γ ts. This underscores the connection of the proposed model with previous adaptive LASSO models [9], []. Via the VB representation in (7) we estimate a single (point) approximation to the components of the precision matrix, like in [9], []. However, rather than using cross validation to infer (point) estimates on the remaining parameters (weights and Lagrange multipliers in [9], [], we estimate posterior distributions on α ss and γ ss. The advantage of this appears to be that the fact that we do not estimate only single values for these latter parameters adds modeling robustness, and the Bayesian formulation has the advantage of not requiring cross validation. Concerning other previous methods, according to () and (), the likelihood function of J ss can be expressed as p(j ss ) exp( (nσ ssj ss J s s J ss +nσ ss J ss)) Equation () is derived based on this likelihood and the prior. This likelihood function is the same as the objective function used in []. It can be verified that exp( (nσ ssj ss J s s J ss +nσ ss J ss)) = exp( nj ss(β W s s β Σ ss β)) with W = J and β = W s s W ss = J ss Jss. The latter expression is also the objective function used in []. The difference of our approach is that Bayesian inference, instead of the l optimization approach, is used to infer J ss. Within the context of the Bayesian analysis we 9

do not require cross-validation, while such is required for the method in []. C. Graph Extension Suppose the above algorithm is used on a training data set X () R p n, and the inferred covariance matrix is W. Now given data from a new testing variable X () R n, we want to extend the original graph to include this new node. The augmented data is X = X() X () R (p+) n, and the corresponding sample covariance is Σ = n XX = Σ Σ. Denote Σ Σ the covariance and inverse covariance matrix for the augmented graph to be J and W, with the same partition as Σ. Then the likelihood of the augmented data can be expressed as p(x) = p(x () X () )p(x () ) = n i= N(X () i ; J J X() i,j )N(X() i ;,W ) In order for the extended graph to be consistent with the original one, the marginal distribution of X () i should remain the same. Thus W = (J J J J ) is fixed to be the training result W. By imposing a Laplacian prior on J like (6), the above problem becomes a sparse regression for the additional variable using the original ones. Using VB inference, the point estimate of J can be expressed as J = (nj Σ +τdiag( α )) ( nσ ) (6) The update of α and γ are similar to equations in Section II-B. The same positive definiteness constraint in (3) is imposed here: J = Σ +J J J, or equivalently W = (J J J J ) = Σ. Now given J, W and W, we would like to update J. Since JW = I, we have W J +W J = and W J +W J =. Using these two equations we obtain J (J W J )+W J =, and finally since J >, J = + +W J W J W (7) We observe that the covariance learned on the training data (W ) is used for the update of J. After J and J are inferred according to (6) and (7), J can be found as J = W + J J J, and it is still sparse.

IV. RESULTS For all experiments in this section, we first normalize the data matrix X to zero mean and unit standard deviation for each variable, then apply the sparse GGM learning algorithm. We may recover the inverse covariance matrix for the original data by using the above mean and standard deviation. The only hyperparameters that must be set are (a,b ) on the gamma prior, and as discussed above we set a = b = 6, as recommended in [9]. These same hyperparameters are used in all examples presented below, with no tuning. A. Simulation In this simulation study (taken from []), four kinds of sparse inverse covariance matrices are generated as follows, with J R p p denoting the inverse covariance matrix. ) (J ) ts =.7 t s. ) J ts = if t s = ;J ts =. if t s = ;J ts =. if t s = ;J ts =. if t s = 3;J ts =. if t s =. 3) J = B + δi, with B ts =. with probability α =. and otherwise; δ is chosen to make the condition number of J equal to p. ) J ts = if t s = ;J ts = if t s = ;J p = J p =.9. We choose the dimension p of the matrix J to be 3, and, and the number of data samples for each data matrix to be,, 8 and. For each (p,n) combination, we generate data matrices from the above four Gaussian models, each of dimension p n, and then apply the Bayesian sparse GGM learning algorithm to each data matrix, and average the performance across the runs. Two performance measures are defined: the Kullback-Leibler loss expressed as KL(J, J) = tr(j J) log(j J) p where J denotes the estimated inverse covariance matrix, and the l loss defined by the Frobenius norm L(J, J) = J J F. We used the graphical LASSO (GLASSO) algorithm [] as the baseline, and compared our Bayesian algorithm with it in Figure and 3. Examples of the the estimation are shown in Figure. The GLASSO algorithm requires tuning of the regularization parameter, so we used a candidate set of [.,.,.,.8,.,.,.,.8,.] for the regularization parameter, and employed ten-fold cross-validation to choose the best candidate in terms of model likelihood, and finally applied the GLASSO algorithm again on the whole dataset using the selected parameter.

From the results we can see that in most cases, the proposed algorithm outperforms the baseline algorithm under both KL and l measure. The gain over the baseline becomes more obvious as the number of samples increases. The main reason for this performance gain is that in our proposed model, each element of the inverse covariance matrix has a specific regularization parameter, making the shrinkage more flexible and the estimation more consistent [9], []. We also observe that the advantage is not obvious when the number of samples n is small and the model dimension p is large (e.g., n = and p = ), in which case no algorithm can give a reliable estimation. Truth Empirical Truth Empirical.8.8.6.6.. 3 3 3 3. 3 3. 3 3.8.8.6.6.. 3 3 3 3. 3 3. 3 3 Truth Empirical Truth Empirical........ 3 3 3 3 3 3 3 3........ 3 3 3 3 3 3 3 3 Fig.. Examples of simulation results (p = 3) for model to (from left to right and from top to bottom). Within each model, the true inverse covariance matrix, the empirical (maximum likelihood) estimation, the GLASSO estimation and the proposed estimation are displayed, from left to right and from top to bottom. B. Senate Data The Senate Voting Records Data from the 9th congress ( 6) is studied in this section. It was previously used in [3]. There are senators, 6 who are Democratic and 6 who are Republican. Each of the senators voted on 6 bills, with yes recorded as and no.

.. for Model with Dimension 3 8 6 for Model with Dimension for Model with Dimension.. 6 8 for Model with Dimension 3 6 8 for Model 3 with Dimension 3.. 6 8 for Model with Dimension 8 6 6 8 for Model 3 with Dimension 6 8 for Model with Dimension 3 6 8 for Model 3 with Dimension 6 6 8 for Model with Dimension 3.. 6 8 6 8 for Model with Dimension 8 6 6 8 6 8 for Model with Dimension 3 6 8 Fig.. KL loss (mean and standard deviation) for the baseline algorithm and the proposed algorithm. Each row represents one of the four models with different model dimension, and each subfigure shows the KL loss of the two algorithms as a function of the number of data samples. L Errors 3 L Errors for Model with Dimension3 L Errors L Errors for Model with Dimension L Errors L Errors for Model with Dimension L Errors 6 8 L Errors for Model with Dimension3 3 L Errors 6 8 L Errors for Model with Dimension 8 6 L Errors 6 8 L Errors for Model with Dimension 8 6 L Errors 6 8 L Errors for Model 3 with Dimension3 3 L Errors 6 8 L Errors for Model 3 with Dimension L Errors 6 8 L Errors for Model 3 with Dimension 3 L Errors 6 8 L Errors for Model with Dimension3 3 6 8 L Errors 6 8 L Errors for Model with Dimension 8 6 6 8 L Errors 6 8 L Errors for Model with Dimension 6 8 Fig. 3. l loss (mean and standard deviation) for the baseline algorithm and the proposed algorithm. Each row represents one of the four models with different model dimension, and each subfigure shows the l loss of the two algorithms as a function of the number of data samples. The missing values are imputed with, as was done in [3]. The sample inverse covariance matrix (i.e., the ML estimate of J) and the inferred sparse inverse covariance are plotted in Figure. The senators are ordered so that the first 6 are Democratic and the remaining 6 are 3

Republican. From the sample inverse covariance matrix, the partisan structure is not clear, but from the inferred sparse inverse covariance matrix, this partisanship becomes very clear. Hence our algorithm removes weak dependencies between variables, revealing latent structure in the data. Sample Inverse Covariance Inferred Inverse Covariance 3 6 7 8 9 6 8 8 6 6 3 6 7 8 9 6 8 Fig.. The sample inverse covariance matrix (left) and the inferred sparse inverse covariance matrix (right) for the senate data. The first 6 senators are Democratic and the next 6 are Republican. C. Face Data The Extended Yale Face Database [3] is a standard dataset with face images from 38 individuals. It was previously used in [] for face recognition. We randomly select half of the images as training data and randomly project the 9 68 dimensional image to dimensions. Algorithm is used to learn the face graph. In Figure, we show five randomly selected faces in the first column, and the top six faces linked to each of them in columns two to seven, according to J. We observe that the strongest links are from the same individuals. We can also view the linkage of individual groups quantitatively. Denote X s to be the sth image from individual l s. According to (), each image can be predicted by a sparse combination of the remaining ones, i.e., p(x s X s ) = N(X s ; t s ( J ss J ts )X t,j ss I). We define that X s is predicted to be from individual c s if the linkages from individual c s contribute most in predicting it, i.e., c s = argmin c X s t s,l t=c ( Jss J ts )X t. (8) Using the true label l s and the predicted label c s, we define a confusion matrix G as G uv = s:l s=u (c s = v)/ s:l s=u (9)

Subject Subject, Weight.3739 Subject, Weight.966 Subject 8, Weight.88 Subject, Weight.78 Subject, Weight.699 Subject, Weight.66 Subject 3 Subject 3, Weight.86 Subject 3, Weight.6979 Subject 36, Weight.39 Subject 3, Weight.33 Subject 3, Weight.38 Subject 3, Weight.9 Subject Subject, Weight.368 Subject 8, Weight.6 Subject, Weight.883 Subject 3, Weight.783 Subject, Weight.78 Subject, Weight.97 Subject Subject, Weight.68 Subject, Weight.6 Subject, Weight.39 Subject, Weight.3776 Subject, Weight.89 Subject 3, Weight.9 Subject Subject, Weight.89 Subject, Weight.386 Subject, Weight.896 Subject, Weight.97 Subject, Weight.633 Subject, Weight.38 Fig.. Example of faces (column one) and the top six links (column two to seven) in the inferred Gaussian graph. and display this Hinton map (the size of the boxes are proportional to the associated matrix value) in Figure 6 (left), with 96.6% of the faces grouped correctly. For a new testing image with unknown label, we apply the graph extension algorithm in Section III-C on the trained graph, and again use (8) and (9) to classify this new image. Our graph extension approach reduces to the sparse representation approach proposed in [], when the trained covariance matrix is simply replaced by the sample covariance. Comparable testing performance is obtained using our graph extension approach, as shown in the classification

Hinton map in Figure 6 (right) with an accuracy rate of 9.7%. The advantage of our approach is that we can explicitly construct a graph for the training data, as partially depicted in Figure. This is important when the primary goal is to learn the relationships between variables, as demonstrated on the cell line data in the next section. Individual Index Individual Index 3 3 3 3 3 3 Individual Index 3 3 Individual Index Fig. 6. Classification Hinton map of the individuals for the training (left) and testing (right) face dataset. The block-diagonal form is consistent with the way the faces were ordered. D. Cell Line Data Human Genome Diversity Cell Line Panel data was collected by [] to study human genomic variations among geographic regions. It contains6 individuals from seven geographic regions, with genotypes sampled at 377 autosomal microsatellite loci. Algorithm is applied to this dataset to learn a 6 6 dimensional sparse inverse covariance matrix. With the geographic labels for the individuals, we apply the same criterion in (8) and (9) to obtain a confusion matrix for the geographical regions, which is displayed in Figure 7. We observe a noticeable sharing among individuals from Central South Asia, Europe and Middle East, consistent with findings in [6] where a factor model is used for the analysis. This demonstrates that the sharing structure of the variables can be uncovered by the sparse GGM learning algorithm. E. Telephone Call Center Data The telephone call center data was used as a real data example for inverse covariance matrix estimation in []. The data is stored in matrix Y with Y ti = N ti +, where N ti is the number 6

Geographic Regions 3 6 7 3 6 7 Geographic Regions Fig. 7. Confusion matrix on cell line data. :Africa, :America, 3:Central South Asia, :East Asia, :Europe, 6:Middle East, 7:Oceania. of calls arriving at the call center for the interval t on day i (t =,,,;i =,,,39). Each data sample Y i R is modeled as a multivariate Gaussian as Y i N(Y i ;µ,j ). The data matrix is split into a training and a testing set. The training set contains data for the first days, and it is used for model estimation. The sample mean estimate is used for µ and the proposed Bayesian model estimate is used for J. For each data sample Y i R in the testing set (i = 6,,39), the last dimensions are assumed missing, and only the first dimensions are observed. To recover the missing data using the observed data and the trained model, the following equation (similar to ()) is used p(y i Y i ) = N(Y i ;µ J J (Y i µ ),J ) where Y i denotes the missing data and Y i denotes the observed data. The averaged performance measure is Error = 3 39 i=6 Y i Ỹi 7

where denotes the l norm, Y i is the ground truth and Ỹi = µ J J (Y i µ ) is the imputation according to the model. The inverse covariance matrix estimated using the proposed Bayesian method is plotted in Figure 8. This shows that data in nearby temporal intervals within one day are very dependent, and data temporally far away are less dependent. The prediction error in the testing set is.336, which is comparable to the adaptive LASSO result (.3) in [] and better than the graphical LASSO result (.39). In the adaptive LASSO method, firstly the shrinkage parameters are estimated using a consistent estimate for J, and then with the shrinkage parameters fixed, the final adaptive LASSO estimate for J is obtained. In contrast, our method infers J and t3he shrinkage parameters simultaneously in one step..338 3 6 7 8 9 6 8 3 Fig. 8. Inverse covariance matrix learned on the telephone call data. V. CONCLUSIONS A new Bayesian formulation has been proposed for sparse inverse covariance matrix estimation. The Laplacian prior is imposed on the off-diagonal elements of the inverse covariance matrix, and a Variational Bayesian (VB) inference algorithm is derived; the Laplacian prior is implemented in a hierarchical Bayesian manner. Properties of symmetry and positive definiteness are guaranteed in the estimation. The inferred sparse inverse covariance matrix can be interpreted as a sparse Gaussian graph, with conditional dependencies among variables expressed by the elements of the inferred matrix. In addition, a graph-extension algorithm is proposed to include 8

a new variable into an inferred graph, which can be used for learning on unseen testing data. Experiments on simulated and real datasets show that the proposed algorithm yields accurate and interpretable results. The main advantage of this Bayesian learning algorithm is that no cross-validation is needed for the regularization parameters, as we learn them adaptively in the Bayesian framework. This means that more data samples can be used for the estimation. This tuning-free algorithm can be readily used in a broad range of real-world applications, where the major goal is to learn relationships among different variables. It could also be used as a building block for more complicated hierarchical models. Our future work will apply the proposed method for missing data imputation [7] and manifold embedding problems. REFERENCES [] S. Lauritzen, Graphical models. Springer Verlag, 996. [] J. Friedman, T. Hastie, and R. Tibshirani, Sparse inverse covariance estimation with the graphical Lasso, Biostatistics, vol. 9, pp. 3, 8. [3] O. Banerjee, L. Ghaoui, and A. d Aspremont, Model selection through sparse maximum likelihood estimation for multivariate Gaussian or binary data, Journal of Machine Learning Research, vol. 9, pp. 8 6, 8. [] J. Duchi, S. Gould, and D. Koller, Projected subgradient methods for learning sparse Gaussians, in Conference on Uncertainty in Artificial Intelligence (UAI), 8. [] N. Meinshausen and P. Buhlmann, High-dimensional graphs and variable selection with the Lasso, The Annals of Statistics, vol. 3, pp. 36 6, 6. [6] M. Yuan, High dimensional inverse covariance matrix estimation via linear programming, Journal of Machine Learning Research, vol., pp. 6 86,. [7] M. Yuan and Y. Lin, Model selection and estimation in the Gaussian graphical model, Biometrika, vol. 9, no., pp. 9 3, 7. [8] R. Tibshirani, Regression shrinkage and selection via the Lasso, J. Royal. Statist. Soc B., vol. 8, no., pp. 67 88, 996. [9] H. Zou, The adaptive Lasso and its oracle properties, Journal of the American Statistical Association, vol., no. 76, pp. 8 9, 6. [] J. Fan, Y. Feng, and Y. Wu, Network exploration via the adaptive Lasso and SCAD penalties, Annals of Applied Statistics, vol. 3, no., pp., 9. [] Q. Liu and A. Ihler, Learning scale free networks by reweightedl regularization, in Proceedings of the th International Conference on Articial Intelligence and Statistics (AISTATS),. [] T. Park and G. Casella, The Bayesian Lasso, Journal of American Statistical Association, vol. 3, no. 8, pp. 68 686, 8. [3] J. Griffin and P. Brown, Bayesian adaptive Lassos with non-convex penalization, Technical Report,. 9

[] M. Yuan, Efficient computation of l regularized estimates in Gaussian graphical models, Journal of Computational and Graphical Statistics, vol. 7, pp. 89 86, 8. [] A. Dobra, C. Hans, B. Jones, J. Nevins, G. Yao, and M. West, Sparse graphical models for exploring gene expression data, Journal of Multivariate Analysis, vol. 9, pp. 96,. [6] A. Rothman, P. Bickel, E. Levina, and J. Zhu, Sparse permutation invariant covariance estimation, Electronic Journal of Statistics, vol., pp. 9, 8. [7] B. Marlin and K. Murphy, Sparse Gaussian graphical models with unknown block structure, in Proceedings of the 6th International Conference on Machine Learning (ICML 9), 9, pp. 7 7. [8] V. Cevher, Learning with compressible priors, in Advances in Neural Information Processing Systems (NIPS), 9. [9] M. Tipping, Sparse Bayesian learning and relevance vector machine, Journal of Machine Learning Research, vol., pp.,. [] H. Ishwaran and J. Rao, Spike and slab variable selection: Frequentist and Bayesian strategies, Annals of Statistics, vol. 33, no., pp. 73 773,. [] M. Beal, Variational algorithms for approximate Bayesian inference, Ph.D. dissertation, Gatsby Computational Neuroscience Unit, University College London, 3. [] D. P. Bertsekas, Nonlinear Programming (nd Edition). Athena Scientific, 999. [3] A. Georghiade, P. Belhumeur, and D. Kriegman, From few to many: Illumination cone models for face recognition under variable lighting and pose, IEEE Trans. PAMI, vol. 3, no. 6, pp. 63 66,. [] J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma, Robust face recognition via sparse representation, IEEE Trans. PAMI, vol. 3, no., pp. 7, 9. [] N. Rosenberg, J. Pritchard, J. Weber, H. Cann, K. Kidd, L. Zhivotovsky, and M. Feldman, Genetic structure of human populations, Science, vol. 98, pp. 38 38,. [6] J. Paisley and L. Carin, Nonparametric factor analysis with beta process priors, in The 6th International Conference on Machine Learning (ICML), 9. [7] N. Stadler and P. Buhlmann, Missing values: sparse inverse covariance estimation and an extension to sparse regression, Statistics and Computation,.