Factorization of Seperable and Patterned Covariance Matrices for Gibbs Sampling

Similar documents
A Single Series from the Gibbs Sampler Provides a False Sense of Security

MONTE CARLO METHODS. Hedibert Freitas Lopes

An Introduction to Multivariate Statistical Analysis

Session 3A: Markov chain Monte Carlo (MCMC)

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

The simple slice sampler is a specialised type of MCMC auxiliary variable method (Swendsen and Wang, 1987; Edwards and Sokal, 1988; Besag and Green, 1

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract

Bayesian Methods for Machine Learning

STAT 518 Intro Student Presentation

ELEC633: Graphical Models

Simulation of truncated normal variables. Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris

November 2002 STA Random Effects Selection in Linear Mixed Models

Partial factor modeling: predictor-dependent shrinkage for linear regression

STA 4273H: Statistical Machine Learning

Introduction to Machine Learning CMU-10701

Bayesian spatial hierarchical modeling for temperature extremes

Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units

The Wishart distribution Scaled Wishart. Wishart Priors. Patrick Breheny. March 28. Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/11

Consistent Bivariate Distribution

Bayesian time series classification

John Geweke Department of Economics University of Minnesota Minneapolis, MN First draft: April, 1991

Partially Collapsed Gibbs Samplers: Theory and Methods. Ever increasing computational power along with ever more sophisticated statistical computing

1. Introduction. Hang Qian 1 Iowa State University

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences

Bayesian Inference for the Multivariate Normal

Bayesian IV: the normal case with multiple endogenous variables

Alternative Bayesian Estimators for Vector-Autoregressive Models

Markov chain Monte Carlo methods for visual tracking

Partially Collapsed Gibbs Samplers: Theory and Methods

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Multivariate Statistical Analysis

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Part 6: Multivariate Normal and Linear Models

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Analysis of the Gibbs sampler for a model. related to James-Stein estimators. Jeffrey S. Rosenthal*

Gibbs Sampling in Linear Models #2

Department of Statistics

Gibbs Sampling in Endogenous Variables Models

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

Monte Carlo Integration using Importance Sampling and Gibbs Sampling

The Bayesian Approach to Multi-equation Econometric Model Estimation

Markov Chain Monte Carlo

Bayesian Linear Regression

Summary STK 4150/9150

TAMS39 Lecture 2 Multivariate normal distribution

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

University of Toronto Department of Statistics

MULTILEVEL IMPUTATION 1

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract

Learning latent structure in complex networks

Lecture 5: Spatial probit models. James P. LeSage University of Toledo Department of Economics Toledo, OH

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

STA 4273H: Statistical Machine Learning

A Fully Nonparametric Modeling Approach to. BNP Binary Regression

Bayesian Graphical Models for Structural Vector AutoregressiveMarch Processes 21, / 1

The lmm Package. May 9, Description Some improved procedures for linear mixed models

Package lmm. R topics documented: March 19, Version 0.4. Date Title Linear mixed models. Author Joseph L. Schafer

A MCMC Approach for Learning the Structure of Gaussian Acyclic Directed Mixed Graphs

Geometric ergodicity of the Bayesian lasso

Tutorial on Probabilistic Programming with PyMC3

An EM algorithm for Gaussian Markov Random Fields

Scaling up Bayesian Inference

Appendix: Modeling Approach

1 Data Arrays and Decompositions

Machine Learning in Simple Networks. Lars Kai Hansen

A Semi-parametric Bayesian Framework for Performance Analysis of Call Centers

Default Priors and Effcient Posterior Computation in Bayesian

Variational Principal Components

An ABC interpretation of the multiple auxiliary variable method

On the Fisher Bingham Distribution

The sbgcop Package. March 9, 2007

Bayesian Networks in Educational Assessment

STA 216, GLM, Lecture 16. October 29, 2007

Quantitative Trendspotting. Rex Yuxing Du and Wagner A. Kamakura. Web Appendix A Inferring and Projecting the Latent Dynamic Factors

MCMC algorithms for fitting Bayesian models

Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida

Bayesian inference for multivariate extreme value distributions

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Label Switching and Its Simple Solutions for Frequentist Mixture Models

R-INLA. Sam Clifford Su-Yun Kang Jeff Hsieh. 30 August Bayesian Research and Analysis Group 1 / 14

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

STA 4273H: Statistical Machine Learning

Wrapped Gaussian processes: a short review and some new results

ST 740: Linear Models and Multivariate Normal Inference

ABC methods for phase-type distributions with applications in insurance risk problems

A Process over all Stationary Covariance Kernels

Condensed Table of Contents for Introduction to Stochastic Search and Optimization: Estimation, Simulation, and Control by J. C.

Introduction to Machine Learning

Afternoon Meeting on Bayesian Computation 2018 University of Reading

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS

Sparse Factor-Analytic Probit Models

On a multivariate implementation of the Gibbs sampler

Outline. Clustering. Capturing Unobserved Heterogeneity in the Austrian Labor Market Using Finite Mixtures of Markov Chain Models

On Reparametrization and the Gibbs Sampler

8.2 Harmonic Regression and the Periodogram

Spatially Adaptive Smoothing Splines

Stat 516, Homework 1

Bayesian Analysis of Vector ARMA Models using Gibbs Sampling. Department of Mathematics and. June 12, 1996

Transcription:

Monte Carlo Methods Appl, Vol 6, No 3 (2000), pp 205 210 c VSP 2000 Factorization of Seperable and Patterned Covariance Matrices for Gibbs Sampling Daniel B Rowe H & SS, 228-77 California Institute of Technology Pasadena, California 91125 Abstract Recently the Gibbs sampler has become a very popular estimation technique especially in Bayesian Statistics In order to implement the Gibbs sampler, matrix factorizations must be computed which normally is not problematic When the dimension of the matrices to be factored is large, computation time increases to an amount to merit special attention I have found that when the matrices to be factored are separable or patterned, results from matrix theory can assist in computation time reduction Keywords: latent roots, latent vectors, random sample 1 Introduction Recently the Gibbs sampler has become one of the favored techniques for parameter estimation especially in Bayesian Statistics The Gibbs sampler is a sampling based approach to calculating marginal posterior distributions especially useful when the densities are not integrable in closed form or are too high a dimension for other methods The Gibbs sampler is a stochastic integration technique that draws samples from conditional densities and uses them to approximate the marginal densities In order to generate a random sample x from say a multivariate normal distribution with mean zero and covariance matrix Ω, denoted by N(0, Ω), we must first compute the factorization of the matrix Ω I will denote this factorization by Ω = UU We then generate a random y from a N(0,I) and transform it by Uy Then x = Uy is a random sample from N(0, Ω = UU ) (Press 1982, p 67) We similarly require matrix factorizations for random samples from Wishart or inverted Wishart distributions In most instances computing the factorization is not very time consuming, and if it is we allow for longer computation times Sometimes, we

206 Daniel B Rowe would like to perform the computations in minimal time I will consider the factorization of matrices that are separable or patterned 2 Seperable Matrices Quite often, the matrices to be factored are of the form Ω = Φ Ψ np np n n p p where denotes the Kroneker product Matrices of this form are called separable Separable matrices occur frequently in regression and time series A single observation vector consisting of observation vectors of smaller dimension, represented by x 1 x x = 2 (21) where the variance of observation vector i is given by the p p matrix, x n var(x i Ω,m,f,λ)=Ω ii = φ ii Ψ and the covariance between observation vectors i and j is given by the p p matrix cov(x i,x j Ω,m,f,λ)=Ω ij = φ ij Ψ has a separable covariance matrix If it is assumed that the observation vectors have a common variance, then this is equilavent to weak stationarity The factorization of the covariance matrix Ω without taking advantage of its separable nature can be very time consuming We take advantage of its separable form to reduce computation The factorization can be found by computing the latent roots and latent vectors (Press 1982) The matrix Ω can be expressed as Ω = ΓD λ Γ =(ΓD 1 2 λ )(ΓD 1 2 λ ) = UU (22) where Γ is an orthogonal matrix of latent vectors as columns and

Factorization of Seperable and Patterned Covariance Matrices 207 D λ =(λ 1,,λ k,,λ Np ) (23) is a matrix of latent roots of Ω This factorization can be performed numerically for a general covarince matrix Quite often, computing the factorization takes appreciable amounts of time especially when we have to calculate them many thousands of times for the Gibbs sampler We can save computation time by using the following theorem Theorem (Anderson 1994) Let the i th latent root of Φ be α i and the vector be u i =(u 1i,,u ni ), i = 1,,n and the j th latent root of Ψ be β j and the vector be v j = (v 1j,,v pj ), j = 1,,p Then, the k th or (ij) th latent root of Ω = Φ Ψ is and the vector is λ k = α i β j (24) γ k = u i v j =(u 1i y j,,u niy j ) (25) for k = 1,,np Now we can find Ω =(ΓD 1 1 2 λ )(ΓD 2 λ ) = UU The required computation is simplified by computing separate latent roots and vectors for Φ and Ψ and using the above theorem Sometimes, one or both of the matrices Φ and Ψ have structure When they have structure we can use the following results for latent roots and vectors of patterned matrices 3 Patterned Matrices We will consider the factorizations when we have structured or patterned matrices Let the patterned matrix either Φ or Ψ be WD δ W with dimensions q If A were the intraclass matrix a b b b a b b, (31) b a with a on the diagonal and b off the diagonal, then the latent roots are

208 Daniel B Rowe δ 1 = a +(N 1)b, δ 2 = = δ q = a b (32) and vectors are any q mutually orthogonal vectors of unit length, the first of which has components that are identical For example the columns of the following Helmert matrix W = 1 1 1 1 q 1 2 2 3 (q 1) q 1 q 1 1 1 2 2 3 0 2 2 3 1 q 0 0 (q 1) (q 1) q (33) If A were the following tri-diagonal matrix a b 0 b a b, (34) a b 0 b a corresponding to a distributed lag model with a on the diagonal and b on the super and sub diagonals, then the latent roots are ( ) lπ δ l = a + 2b cos,l= 1,,q (35) q + 1 and vectors are ( ( ) ( ) ( )) lπ 2lπ qlπ w l = sin, sin,,sin,l= 1,,q q + 1 q + 1 q + 1 (36) If A were the circular matrix a 1 a 2 a q a q a 1 a q 1, (37) a 2 a 3 a 1 with symmetry imposed on it so that it were a covariance matrix with applications to cyclical time series, then the latent roots are

Factorization of Seperable and Patterned Covariance Matrices 209 δ l = q s=1 [ ] 2π a s 1 cos (s 1)(l 1),l= 1,,q (38) q and the elements of W are w lm = 1 [ ] [ ]} 2π 2π {cos (l 1)(m 1) + sin (l 1)(m 1) (39) q q q If A were the Toeplitz matrix corresponding to a first order Markov correlation scheme 1 a a 2 a N 1 a 1 a a N 2 (310) a N 1 a N 2 1 with applications to time series, for example an AR(1), then the latent roots are δ 1 = 1, δ 2 = = δ q = 1 a 2 (311) but the a general form for the latent vectors w l is not known to the current author It has been found by the current author that a matrix with the above structure can be factorized by 1 0 0 0 a 1 a 2 0 0 T = a 2 a 1 a 2 1 a 2 a 3 a 2, (312) 1 a 2 where TT But if Φ = A where A is the first order Markov correlation matrix in Ω = Φ Ψ we still do not have an expression for Ω = UU An approximation to Ψ aψ a 2 Ψ a N 1 Ψ Ψ Ω, (313) Ψ can be made which is

210 Daniel B Rowe Ω = Ψ a b Ψ 0 a b Ψ 0, (314) where it is assumed that a b+1 is small enough to be neglected and a routine for exact factorization of band symmetric matrices can be used 4 Conclusion Factorization of separable or patterned matrices can be performed by taking advantage of the abovementioned theorem and results for latent roots and latent vectors Implementation of these factorization techniques will show that they are much faster than direct factorization when the dimension of the covariance matrix to be factorized becomes large References [1] Anderson, T W (1984) An Introduction to Multivariate Statistical Analysis John Wiley and Sons, Inc, New York, p 600 [2] Gelfand, A E and Smith, A F M (1990) Sampling based approaches to calculating marginal densities J Amer Stat Assoc 85, 398-409 [3] Geman, S and Geman, D (1984) Stochastic relaxation, Gibbs distributions and the Bayesian restoration of images IEEE Trans on pattern analysis and machine intelligence 6, 721-741 [4] Press, S J (1982) Applied Multivariate Analysis: Using Bayesian and Frequentist Methods of Inference Robert E Krieger Publishing Company, Malabar, Florida [5] Press, S J (1989) Bayesian Statistics: Principles, Models, and Applications John Wiley and Sons, Inc, New York