Nearest Neighbor Gaussian Processes for Large Spatial Data
|
|
- Aron Washington
- 5 years ago
- Views:
Transcription
1 Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, Department of Biostatistics, Bloomberg School of Public Health, Johns Hopkins University, Baltimore, Maryland. 2 Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles. 3 Departments of Forestry and Geography, Michigan State University, East Lansing, Michigan.
2 Low rank Gaussian Predictive Process Pros Proper Gaussian process Allows for coherent spatial interpolation at arbitrary resolution Can be used as prior for spatial random effects in any hierarchical setup for spatial data Computationally tractable 1
3 Low rank Gaussian Predictive Process Cons True w Full GP PP 64 knots Figure: Comparing full GP vs low-rank GP with 2500 locations Low rank models like the Predictive Process (PP) often tends to oversmooth Increasing the number of knots can fix this but will lead to heavy computation 1
4 Sparse matrices Idea: Use a sparse matrix instead of a low rank matrix to approximate the dense full GP covariance matrix Goals: Scalability: Both in terms of storage and computing inverse and determinants Closely approximate full GP inference Proper Gaussian process model like the Predictive Process 2
5 Cholesky factors Write a joint density p(w) = p(w 1, w 2,..., w n ) as: p(w 1 )p(w 2 w 1 )p(w 3 w 1, w 2 ) p(w n w 1, w 2,..., w n 1 ) For Gaussian distribution w N(0, C) this w 1 = 0 + η 1 ; w 2 = a 21 w 1 + η 2 ; w n = a n1 w 1 + a n2 w a n,n 1 w n 1 + η n ; 3
6 Cholesky factors Write a joint density p(w) = p(w 1, w 2,..., w n ) as: p(w 1 )p(w 2 w 1 )p(w 3 w 1, w 2 ) p(w n w 1, w 2,..., w n 1 ) For Gaussian distribution w N(0, C) this w 1 w 2 w 3. w n = a a 31 a a n1 a n2 a n3... a n,n 1 0 w 1 w 2 w 3. w n + η 1 η 2 η 3. η n = w = Aw + η; η N(0, D), where D = diag(d 1, d 2,..., d n ). 3
7 Cholesky factors Write a joint density p(w) = p(w 1, w 2,..., w n ) as: p(w 1 )p(w 2 w 1 )p(w 3 w 1, w 2 ) p(w n w 1, w 2,..., w n 1 ) For Gaussian distribution w N(0, C) this w w 1 η 1 w 2 a w 2 η 2 w 3 = a 31 a w 3 + η a n1 a n2 a n3... a n,n 1 0 w n = w = Aw + η; η N(0, D), where D = diag(d 1, d 2,..., d n ). Cholesky factorization: C 1 = (I A) D 1 (I A) w n η n 3
8 Cholesky factors w <i = (w 1, w 2,..., w i 1 ) c i = Cov(w i, w <i ), C i = Var(w <i ) i th row of A and d i = Var(η i ) are obtained from p(w i w <i ) as follows: Solve for a ij s from i 1 j=1 a ijw j = E(w i w <i ) = c i C 1 i w <i 4
9 Cholesky factors w <i = (w 1, w 2,..., w i 1 ) c i = Cov(w i, w <i ), C i = Var(w <i ) i th row of A and d i = Var(η i ) are obtained from p(w i w <i ) as follows: Solve for a ij s from i 1 j=1 a ijw j = E(w i w <i ) = c i C 1 i w <i d i = Var(w i w <i ) = σ 2 c i C 1 i c i 4
10 Cholesky factors w <i = (w 1, w 2,..., w i 1 ) c i = Cov(w i, w <i ), C i = Var(w <i ) i th row of A and d i = Var(η i ) are obtained from p(w i w <i ) as follows: Solve for a ij s from i 1 j=1 a ijw j = E(w i w <i ) = c i C 1 i w <i d i = Var(w i w <i ) = σ 2 c i C 1 i c i 4
11 Cholesky factors w <i = (w 1, w 2,..., w i 1 ) c i = Cov(w i, w <i ), C i = Var(w <i ) i th row of A and d i = Var(η i ) are obtained from p(w i w <i ) as follows: Solve for a ij s from i 1 j=1 a ijw j = E(w i w <i ) = c i C 1 i w <i d i = Var(w i w <i ) = σ 2 c i C 1 i c i For large i, inverting C i becomes slow The Cholesky factor approach for the full GP covariance matrix C does not offer any computational benefits 4
12 Cholesky Factors and Directed Acyclic Graphs (DAGs) Full dependency graph Number of non-zero entries (sparsity) of A equals number of arrows in the graph In particular: Sparsity of the i th row of A is same as the number of arrows towards i in the DAG 5
13 Introducing sparsity via graphical models Full dependency graph p(y 1 )p(y 2 y 1 )p(y 3 y 1, y 2 )p(y 4 y 1, y 2, y 3 ) p(y 5 y 1, y 2, y 3, y 4 )p(y 6 y 1, y 2,..., y 5 )p(y 7 y 1, y 2,..., y 6 ). 6
14 Introducing sparsity via graphical models 3 Nearest neighbor dependency graph p(y 1 )p(y 2 y 1 )p(y 3 y 1, y 2 )p(y 4 y 1, y 2, y 3 ) p(y 5 y 1, y 2, y 3, y 4 )p(y 6 y 1, y 2, y 3, y 4, y 5 )p(y 7 y 1, y 2, y 3, y 4, y 5, y 6 ) 6
15 Introducing sparsity via graphical models 3 Nearest neighbor dependency graph Create a sparse DAG by keeping at most m arrows pointing to each node Set a ij = 0 for all i, j which has no arrow between them Fixing a ij = 0 introduces conditional independence and w j drops out from the conditional set in p(w i {w k : l < i}) 7
16 Introducing sparsity via graphical models 3 Nearest neighbor dependency graph N(i) denote neighbor set of i, i.e., the set of nodes from which there are arrows to i a ij = 0 for j / N(i) and nonzero a ij s obtained by solving: E[w i w N(i) ] = a ij w j j N(i) The above linear system is only m m 7
17 Choosing neighbor sets Matern Covariance Function: C(s i, s j ) = 1 2 ν 1 Γ(ν) ( s i s j φ) ν K ν ( s i s j φ); φ > 0, ν > 0, 8
18 Choosing neighbor sets Spatial covariance functions decay with distance Vecchia (1988): N(s i ) = m nearest neighbors of s i in s 1, s 2,..., s i 1 Nearest points have highest correlations Theory: Screening effect Stein, 2002 We use Vecchia s choice of m-nearest neighbor Other choices proposed in Stein et al. (2004); Gramacy and Apley (2015); Guinness (2016) can also be used 9
19 Nearest neighbors 10
20 Nearest neighbors 10
21 Nearest neighbors 10
22 Sparse precision matrices The neighbor sets and the covariance function C(, ) define a sparse Cholesky factor A N(w 0, C) N(w 0, C) ; C 1 = (I A) D 1 (I A) A D C 1 det( C) = n i=1 D i, C 1 is sparse with O(nm 2 ) entries 11
23 Sparse precision matrices The neighbor sets and the covariance function C(, ) define a sparse Cholesky factor A N(w 0, C) N(w 0, C) ; C 1 = (I A) D 1 (I A) A D C 1 det( C) = n i=1 D i, C 1 is sparse with O(nm 2 ) entries 11
24 Extension to a Process We have defined w N(0, C) over the set of data locations S = {s 1, s 2,..., s n } For s / S, define N(s) as set of m-nearest neighbors of s in S Define w(s) = i:s i N(s) a i(s)w(s i ) + η(s) where η(s) ind N(0, d(s)) a i (s) and d(s) are once again obtained by solving m m system Well-defined GP over entire domain Nearest Neighbor GP (NNGP) Datta et al., JASA, (2016) 12
25 Hierarchical spatial regression with NNGP Spatial linear model y(s) = x(s) β + w(s) + ɛ(s) w(s) modeled as NNGP derived from a GP(0, C(,, σ 2, φ)) ɛ(s) iid N(0, τ 2 ) contributes to the nugget Priors for the parameters β, σ 2, τ 2 and φ Only difference from a full GP model is the NNGP prior w(s) 13
26 Hierarchical spatial regression with NNGP Full Bayesian Model N(y Xβ + w, τ 2 I) N(w 0, C(σ 2, φ)) N(β µ β, V β ) IG(τ 2 a τ, b τ ) IG(σ 2 a σ, b σ ) Unif (φ a φ, b φ ) Gibbs sampler: Conjugate full conditionals for β, τ 2, σ 2 and w(s i ) s Metropolis step for updating φ Posterior predictive distribution at any location using composition sampling: N(y(s) x(s) β + w(s), τ 2 I) N(w(s) a(s) w R, d(s)) p(w, β, τ 2, σ 2, φ y) d(w, β, τ 2, σ 2, φ) 14
27 Choosing m Run NNGP in parallel for few values of m Choose m based on model evaluation metrics Our results suggested that typically m 20 yielded excellent approximations to the full GP 15
28 Storage and computation Storage: Never needs to store n n distance matrix Stores smaller m m matrices Total storage requirements O(nm 2 ) Computation: Only involves inverting small m m matrices Total flop count per iteration of Gibbs sampler is O(nm 3 ) Since m n, NNGP offers great scalability for large datasets 16
29 Simulation experiments 2500 locations on a unit square y(s) = β 0 + β 1 x(s) + w(s) + ɛ(s) Single covariate x(s) generated as iid N(0, 1) Spatial effects generated from GP(0, σ 2 R(, φ)) R(, φ) is exponential correlation function with decay φ Candidate models: Full GP, Gaussian Predictive Process (GPP) with 64 knots and NNGP 17
30 Fitted Surfaces True w Full GP GPP 64 knots NNGP, m = 10 NNGP, m = 20 Figure: Univariate synthetic data analysis 18
31 Parameter estimates NNGP Predictive Process Full True m = 10 m = knots Gaussian Process β (0.62, 1.31) 1.03 (0.65, 1.34) 1.30 (0.54, 2.03) 1.03 (0.69, 1.34) β (4.99, 5.03) 5.01 (4.99, 5.03) 5.03 (4.99, 5.06) 5.01 (4.99, 5.03) σ (0.78, 1.23) 0.94 (0.77, 1.20) 1.29 (0.96, 2.00) 0.94 (0.76, 1.23) τ (0.08, 0.13) 0.10 (0.08, 0.13) 0.08 (0.04, 0.13) 0.10 (0.08, 0.12) φ (9.70, 16.77) (9.99, 17.15) 5.61 (3.48, 8.09) (9.92, 17.50) 19
32 Model evaluation NNGP Predictive Process Full m = 10 m = knots Gaussian Process DIC score RMSPE Run time (Minutes) NNGP performs at par with Full GP GPP oversmooths and performs much worse both in terms of parameter estimation and model comparison NNGP yields huge computational gains 20
33 Multivariate spatial data Point-referenced spatial data often come as multivariate measurements at each location. Examples: Environmental monitoring: stations yield measurements on ozone, NO, CO, and PM 2.5. Forestry: measurements of stand characteristics age, total biomass, and average tree diameter. Atmospheric modeling: at a given site we observe surface temperature, precipitation and wind speed We anticipate dependence between measurements at a particular location across locations 21
34 Multvariate spatial linear model Spatial linear model for q-variate spatial data: y i = x i (s)β i + w i (s) + ɛ i (s) for i = 1, 2,..., q ɛ(s) = (ɛ 1 (s), ɛ 2 (s),..., ɛ q (s)) N(0, E) where E is the q q noise matrix w(s) = (w 1 (s), w 2 (s),..., w q (s)) is modeled as a q-variate Gaussian process 22
35 Spatially varying coefficients Often the relationship between the (univariate) spatial response and covariates vary across the space The regression coefficients can then be modeled as spatial processes Spatially varying coefficient (SVC) model: y(s) = x(s) β(s) + ɛ(s) Even though the response can be univariate, β(s) is modeled as a p-variate GP 23
36 Multivariate GPs Cov(w(s i ), w(s j )) = C(s i, s j θ) a q q cross-covariance matrix Choices for the function C(, θ) Multivariate Matérn Linear model of co-regionalization For data observed at n locations, all choices lead to a dense nq nq matrix C = Cov(w(s 1 ), w(s 2 ),..., w(s n )) Not scalable when nq is large 24
37 Multivariate NNGPs Cholesky factor approach similar to the univariate case w(s 1 ) w(s 1 ) η(s 1 ) w(s 2 ) A w(s 2 ) η(s 2 ) w(s 3 ) = A 31 A w(s 3 ) + η(s 3 ) w(s n ) A n1 A n2 A n3... A n,n 1 0 w(s n ) η(s n ) = w = Aw + η; η N(0, D), D = diag(d 1, D 2,..., D n ). Only differences: w(s i ) and η(s i ) s are q 1 vectors and A ij and D i s are q q matrix 25
38 Multivariate NNGPs Choose neighbor sets N(i) for each location s i Set A ij = 0 if j / N(i) Solve for non-zero A ij s from the mq mq linear system: j N(i) A ijw(s j ) = E(w(s i ) {w(s j ) j N(i)}) Multivariate NNGP: w N(0, C) where C 1 = (I A) D 1 (I A) C 1 is sparse with O(nm 2 ) non-zero q q blocks det( C) = n i=1 det(d i ) Storage and computation needs remains linear in n 26
39 U.S. Forest biomass data Observed biomass NDVI Forest biomass data from measurements at 114,371 plots NDVI (greenness) is used to predict forest biomass 27
40 U.S. Forest biomass data Non Spatial Model Biomass = β 0 + β 1 NDVI + error, ˆβ 0 = 1.043, ˆβ 1 = Residuals Variogram of residuals Strong spatial pattern among residuals 28
41 Forest biomass dataset n 10 5 (Forest Biomass) full GP requires storage 40Gb and time 140 hrs per iteration. We use a spatially varying coefficients NNGP model Model Biomass(s) = β 0 (s) + β 1 (s)ndvi(s) + ɛ(s) w(s) = (β 0 (s), β 1 (s)) Bivariate NNGP(0, C(, θ)), m = 5 Time 6 seconds per iteration Full inferential output: 41 hours (25000 MCMC iterations) 29
42 Forest biomass data Observed biomass Fitted biomass β 0(s) β NDVI (s) 30
43 Reducing parameter dimensionality The Gibbs sampler algorithm for the NNGP updates w(s 1 ), w(s 2 ),..., w(s n ) sequentially Dimension of the MCMC for this sequential algorithm is O(n) If the number of data locations n is very large, this high-dimensional MCMC can converge slowly Although each iteration for the NNGP model will be very fast, many more MCMC iterations may be required 31
44 Collapsed NNGP Same model: y(s) = x(s) β + w(s) + ɛ(s) w(s) NNGP(0, C(, θ)) ɛ(s) iid N(0, τ 2 ) Vector form y N(Xβ + w, τ 2 I); w N(0, C(θ)) Collapsed model: Marginalizing out w, we have y N(Xβ, τ 2 I + C(θ)) 32
45 Collapsed NNGP Model y N(Xβ, τ 2 I + C(θ)) Only involves few parameters β, τ 2 and θ = (σ 2, φ) Drastically reduces the MCMC dimensionality Gibbs sampler updates are based on sparse linear systems using C 1 Improved MCMC convergence Can recover posterior distribution of w y Complexity of the algorithm depends on the design of the data locations and is not guaranteed to be O(n) 33
46 Response NNGP w(s) GP(0, C(, θ)) y(s) GP(x(s) β, Σ(, τ 2, θ)) Σ(s i, s j ) = C(s i, s j θ) + τ 2 δ(s i = s j ) (δ is Kronecker delta) We can directly derive the NNGP covariance function corresponding to Σ(, ) Σ is the NNGP covariance matrix for the n locations Response model: y N(Xβ, Σ) Storage and computations are guaranteed to be O(n) Low dimensional MCMC Improved convergence Cannot coherently recover w y 34
47 Comparison of NNGP models Sequential Collapsed Response O(n) time Yes No Yes Recovery of w y Yes Yes No Parameter dimensionality High Low Low 35
48 Comparison of NNGP models Figure: Run time per iteration as a function of number of locations for different NNGP models 36
49 Comparison of NNGP models Figure: MCMC convergence diagnostics using Gelman-Rubin shrink factor for different NNGP models for a simulated dataset 37
50 Summary of Nearest Neighbor Gaussian Processes Sparsity inducing Gaussian process Constructed from sparse Cholesky factors based on m nearest neighbors Scalability: Storage, inverse and determinant of NNGP covariance matrix are all O(n) Proper Gaussian process, allows for inference using hierarchical spatial models and predictions at arbitrary spatial resolution Closely approximates full GP inference, does not oversmooth like low rank models Extension to multivariate NNGP Collapsed and response NNGP models with improved MCMC convergence spnngp package in R for analyzing large spatial data using NNGP models 38
Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets
Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,
More informationOn Gaussian Process Models for High-Dimensional Geostatistical Datasets
On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May
More informationBayesian Modeling and Inference for High-Dimensional Spatiotemporal Datasets
Bayesian Modeling and Inference for High-Dimensional Spatiotemporal Datasets Sudipto Banerjee University of California, Los Angeles, USA Based upon projects involving: Abhirup Datta (Johns Hopkins University)
More informationHierarchical Modelling for Univariate Spatial Data
Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department
More informationIntroduction to Geostatistics
Introduction to Geostatistics Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns Hopkins University, Baltimore,
More informationHierarchical Modelling for Univariate Spatial Data
Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.
More informationHierarchical Modeling for Multivariate Spatial Data
Hierarchical Modeling for Multivariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department
More informationHierarchical Modeling for Univariate Spatial Data
Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This
More informationBayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes
Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Andrew O. Finley 1 and Sudipto Banerjee 2 1 Department of Forestry & Department of Geography, Michigan
More informationHierarchical Modeling for Spatio-temporal Data
Hierarchical Modeling for Spatio-temporal Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of
More informationBayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes
Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota,
More informationHierarchical Modelling for Multivariate Spatial Data
Hierarchical Modelling for Multivariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Point-referenced spatial data often come as
More informationBayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes
Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Alan Gelfand 1 and Andrew O. Finley 2 1 Department of Statistical Science, Duke University, Durham, North
More informationGaussian predictive process models for large spatial data sets.
Gaussian predictive process models for large spatial data sets. Sudipto Banerjee, Alan E. Gelfand, Andrew O. Finley, and Huiyan Sang Presenters: Halley Brantley and Chris Krut September 28, 2015 Overview
More informationBayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes
Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Andrew O. Finley Department of Forestry & Department of Geography, Michigan State University, Lansing
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee September 03 05, 2017 Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles Linear Regression Linear regression is,
More informationA Note on the comparison of Nearest Neighbor Gaussian Process (NNGP) based models
A Note on the comparison of Nearest Neighbor Gaussian Process (NNGP) based models arxiv:1811.03735v1 [math.st] 9 Nov 2018 Lu Zhang UCLA Department of Biostatistics Lu.Zhang@ucla.edu Sudipto Banerjee UCLA
More informationspbayes: An R Package for Univariate and Multivariate Hierarchical Point-referenced Spatial Models
spbayes: An R Package for Univariate and Multivariate Hierarchical Point-referenced Spatial Models Andrew O. Finley 1, Sudipto Banerjee 2, and Bradley P. Carlin 2 1 Michigan State University, Departments
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationIntroduction to Spatial Data and Models
Introduction to Spatial Data and Models Researchers in diverse areas such as climatology, ecology, environmental health, and real estate marketing are increasingly faced with the task of analyzing data
More informationAn Additive Gaussian Process Approximation for Large Spatio-Temporal Data
An Additive Gaussian Process Approximation for Large Spatio-Temporal Data arxiv:1801.00319v2 [stat.me] 31 Oct 2018 Pulong Ma Statistical and Applied Mathematical Sciences Institute and Duke University
More informationGaussian Processes 1. Schedule
1 Schedule 17 Jan: Gaussian processes (Jo Eidsvik) 24 Jan: Hands-on project on Gaussian processes (Team effort, work in groups) 31 Jan: Latent Gaussian models and INLA (Jo Eidsvik) 7 Feb: Hands-on project
More informationModelling Multivariate Spatial Data
Modelling Multivariate Spatial Data Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. June 20th, 2014 1 Point-referenced spatial data often
More informationHierarchical Modelling for Univariate and Multivariate Spatial Data
Hierarchical Modelling for Univariate and Multivariate Spatial Data p. 1/4 Hierarchical Modelling for Univariate and Multivariate Spatial Data Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota
More informationA Divide-and-Conquer Bayesian Approach to Large-Scale Kriging
A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging Cheng Li DSAP, National University of Singapore Joint work with Rajarshi Guhaniyogi (UC Santa Cruz), Terrance D. Savitsky (US Bureau of Labor
More informationHierarchical Modeling for non-gaussian Spatial Data
Hierarchical Modeling for non-gaussian Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department
More informationMultivariate spatial modeling
Multivariate spatial modeling Point-referenced spatial data often come as multivariate measurements at each location Chapter 7: Multivariate Spatial Modeling p. 1/21 Multivariate spatial modeling Point-referenced
More informationBayesian Linear Regression
Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective
More informationSpatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields
Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields 1 Introduction Jo Eidsvik Department of Mathematical Sciences, NTNU, Norway. (joeid@math.ntnu.no) February
More informationModels for spatial data (cont d) Types of spatial data. Types of spatial data (cont d) Hierarchical models for spatial data
Hierarchical models for spatial data Based on the book by Banerjee, Carlin and Gelfand Hierarchical Modeling and Analysis for Spatial Data, 2004. We focus on Chapters 1, 2 and 5. Geo-referenced data arise
More informationIntroduction to Spatial Data and Models
Introduction to Spatial Data and Models Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry
More informationIntroduction to Spatial Data and Models
Introduction to Spatial Data and Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry
More informationHierarchical Modelling for non-gaussian Spatial Data
Hierarchical Modelling for non-gaussian Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2
More informationEfficient algorithms for Bayesian Nearest Neighbor. Gaussian Processes
Efficient algorithms for Bayesian Nearest Neighbor Gaussian Processes arxiv:1702.00434v3 [stat.co] 3 Mar 2018 Andrew O. Finley 1, Abhirup Datta 2, Bruce C. Cook 3, Douglas C. Morton 3, Hans E. Andersen
More informationIntroduction to Spatial Data and Models
Introduction to Spatial Data and Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics,
More informationPractical Bayesian Modeling and Inference for Massive Spatial Datasets On Modest Computing Environments
Practical Bayesian Modeling and Inference for Massive Spatial Datasets On Modest Computing Environments Lu Zhang UCLA Department of Biostatistics arxiv:1802.00495v1 [stat.me] 1 Feb 2018 Lu.Zhang@ucla.edu
More informationHierarchical Modeling for Spatial Data
Bayesian Spatial Modelling Spatial model specifications: P(y X, θ). Prior specifications: P(θ). Posterior inference of model parameters: P(θ y). Predictions at new locations: P(y 0 y). Model comparisons.
More informationDynamically updated spatially varying parameterisations of hierarchical Bayesian models for spatially correlated data
Dynamically updated spatially varying parameterisations of hierarchical Bayesian models for spatially correlated data Mark Bass and Sujit Sahu University of Southampton, UK June 4, 06 Abstract Fitting
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationThe Matrix Reloaded: Computations for large spatial data sets
The Matrix Reloaded: Computations for large spatial data sets Doug Nychka National Center for Atmospheric Research The spatial model Solving linear systems Matrix multiplication Creating sparsity Sparsity,
More informationModel Assessment and Comparisons
Model Assessment and Comparisons Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationA Fully Nonparametric Modeling Approach to. BNP Binary Regression
A Fully Nonparametric Modeling Approach to Binary Regression Maria Department of Applied Mathematics and Statistics University of California, Santa Cruz SBIES, April 27-28, 2012 Outline 1 2 3 Simulation
More informationBayesian spatial hierarchical modeling for temperature extremes
Bayesian spatial hierarchical modeling for temperature extremes Indriati Bisono Dr. Andrew Robinson Dr. Aloke Phatak Mathematics and Statistics Department The University of Melbourne Maths, Informatics
More informationarxiv: v1 [stat.me] 28 Dec 2017
A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging Raarshi Guhaniyogi, Cheng Li, Terrance D. Savitsky 3, and Sanvesh Srivastava 4 arxiv:7.9767v [stat.me] 8 Dec 7 Department of Applied Mathematics
More informationThe Matrix Reloaded: Computations for large spatial data sets
The Matrix Reloaded: Computations for large spatial data sets The spatial model Solving linear systems Matrix multiplication Creating sparsity Doug Nychka National Center for Atmospheric Research Sparsity,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationProbabilistic Graphical Models
2016 Robert Nowak Probabilistic Graphical Models 1 Introduction We have focused mainly on linear models for signals, in particular the subspace model x = Uθ, where U is a n k matrix and θ R k is a vector
More informationPart 6: Multivariate Normal and Linear Models
Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationSome notes on efficient computing and setting up high performance computing environments
Some notes on efficient computing and setting up high performance computing environments Andrew O. Finley Department of Forestry, Michigan State University, Lansing, Michigan. April 17, 2017 1 Efficient
More informationThe linear model is the most fundamental of all serious statistical models encompassing:
Linear Regression Models: A Bayesian perspective Ingredients of a linear model include an n 1 response vector y = (y 1,..., y n ) T and an n p design matrix (e.g. including regressors) X = [x 1,..., x
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationCovariate-Assisted Variable Ranking
Covariate-Assisted Variable Ranking Tracy Ke Department of Statistics Harvard University WHOA-PSI@St. Louis, Sep. 8, 2018 1/18 Sparse linear regression Y = X β + z, X R n,p, z N(0, σ 2 I n ) Signals (nonzero
More informationBayesian data analysis in practice: Three simple examples
Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to
More informationHierarchical Modeling for Univariate Spatial Data
Univariate spatial models Spatial Domain Hierarchical Modeling for Univariate Spatial Data Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis,
More informationMetropolis-Hastings Algorithm
Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to
More informationAreal data models. Spatial smoothers. Brook s Lemma and Gibbs distribution. CAR models Gaussian case Non-Gaussian case
Areal data models Spatial smoothers Brook s Lemma and Gibbs distribution CAR models Gaussian case Non-Gaussian case SAR models Gaussian case Non-Gaussian case CAR vs. SAR STAR models Inference for areal
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationLecture 23. Spatio-temporal Models. Colin Rundel 04/17/2017
Lecture 23 Spatio-temporal Models Colin Rundel 04/17/2017 1 Spatial Models with AR time dependence 2 Example - Weather station data Based on Andrew Finley and Sudipto Banerjee s notes from National Ecological
More informationMULTILEVEL IMPUTATION 1
MULTILEVEL IMPUTATION 1 Supplement B: MCMC Sampling Steps and Distributions for Two-Level Imputation This document gives technical details of the full conditional distributions used to draw regression
More informationOn Some Computational, Modeling and Design Issues in Bayesian Analysis of Spatial Data
On Some Computational, Modeling and Design Issues in Bayesian Analysis of Spatial Data A DISSERTATION SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL OF THE UNIVERSITY OF MINNESOTA BY Qian Ren IN PARTIAL
More informationA new Hierarchical Bayes approach to ensemble-variational data assimilation
A new Hierarchical Bayes approach to ensemble-variational data assimilation Michael Tsyrulnikov and Alexander Rakitko HydroMetCenter of Russia College Park, 20 Oct 2014 Michael Tsyrulnikov and Alexander
More informationSub-kilometer-scale space-time stochastic rainfall simulation
Picture: Huw Alexander Ogilvie Sub-kilometer-scale space-time stochastic rainfall simulation Lionel Benoit (University of Lausanne) Gregoire Mariethoz (University of Lausanne) Denis Allard (INRA Avignon)
More informationBayesian linear regression
Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationNonparametric Bayesian Methods
Nonparametric Bayesian Methods Debdeep Pati Florida State University October 2, 2014 Large spatial datasets (Problem of big n) Large observational and computer-generated datasets: Often have spatial and
More informationA Framework for Daily Spatio-Temporal Stochastic Weather Simulation
A Framework for Daily Spatio-Temporal Stochastic Weather Simulation, Rick Katz, Balaji Rajagopalan Geophysical Statistics Project Institute for Mathematics Applied to Geosciences National Center for Atmospheric
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationJournal of Statistical Software
JSS Journal of Statistical Software April 2007, Volume 19, Issue 4. http://www.jstatsoft.org/ spbayes: An R Package for Univariate and Multivariate Hierarchical Point-referenced Spatial Models Andrew O.
More informationGaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency
Gaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency Chris Paciorek March 11, 2005 Department of Biostatistics Harvard School of
More informationFusing point and areal level space-time data. data with application to wet deposition
Fusing point and areal level space-time data with application to wet deposition Alan Gelfand Duke University Joint work with Sujit Sahu and David Holland Chemical Deposition Combustion of fossil fuel produces
More informationComparing Non-informative Priors for Estimation and Prediction in Spatial Models
Environmentrics 00, 1 12 DOI: 10.1002/env.XXXX Comparing Non-informative Priors for Estimation and Prediction in Spatial Models Regina Wu a and Cari G. Kaufman a Summary: Fitting a Bayesian model to spatial
More informationST 740: Linear Models and Multivariate Normal Inference
ST 740: Linear Models and Multivariate Normal Inference Alyson Wilson Department of Statistics North Carolina State University November 4, 2013 A. Wilson (NCSU STAT) Linear Models November 4, 2013 1 /
More informationHierarchical Linear Models
Hierarchical Linear Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin The linear regression model Hierarchical Linear Models y N(Xβ, Σ y ) β σ 2 p(β σ 2 ) σ 2 p(σ 2 ) can be extended
More informationScalable MCMC for the horseshoe prior
Scalable MCMC for the horseshoe prior Anirban Bhattacharya Department of Statistics, Texas A&M University Joint work with James Johndrow and Paolo Orenstein September 7, 2018 Cornell Day of Statistics
More informationWrapped Gaussian processes: a short review and some new results
Wrapped Gaussian processes: a short review and some new results Giovanna Jona Lasinio 1, Gianluca Mastrantonio 2 and Alan Gelfand 3 1-Università Sapienza di Roma 2- Università RomaTRE 3- Duke University
More informationBayesian Learning in Undirected Graphical Models
Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ Work with: Iain Murray and Hyun-Chul
More informationBayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL
1 Bayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL 2 MOVING AVERAGE SPATIAL MODELS Kernel basis representation for spatial processes z(s) Define m basis functions
More informationMultivariate spatial models and the multikrig class
Multivariate spatial models and the multikrig class Stephan R Sain, IMAGe, NCAR ENAR Spring Meetings March 15, 2009 Outline Overview of multivariate spatial regression models Case study: pedotransfer functions
More informationMotivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University
Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined
More informationStatistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling
1 / 27 Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling Melih Kandemir Özyeğin University, İstanbul, Turkey 2 / 27 Monte Carlo Integration The big question : Evaluate E p(z) [f(z)]
More informationGaussian Process Regression
Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process
More informationGeostatistical Modeling for Large Data Sets: Low-rank methods
Geostatistical Modeling for Large Data Sets: Low-rank methods Whitney Huang, Kelly-Ann Dixon Hamil, and Zizhuang Wu Department of Statistics Purdue University February 22, 2016 Outline Motivation Low-rank
More informationThe Bayesian approach to inverse problems
The Bayesian approach to inverse problems Youssef Marzouk Department of Aeronautics and Astronautics Center for Computational Engineering Massachusetts Institute of Technology ymarz@mit.edu, http://uqgroup.mit.edu
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationTowards a Complete Picture of Covariance Functions on Spheres Cross Time
Towards a Complete Picture of Covariance Functions on Spheres Cross Time Philip White 1 and Emilio Porcu 2 arxiv:1807.04272v1 [stat.me] 11 Jul 2018 1 Department of Statistical Science, Duke University,
More informationBayesian non-parametric model to longitudinally predict churn
Bayesian non-parametric model to longitudinally predict churn Bruno Scarpa Università di Padova Conference of European Statistics Stakeholders Methodologists, Producers and Users of European Statistics
More information20: Gaussian Processes
10-708: Probabilistic Graphical Models 10-708, Spring 2016 20: Gaussian Processes Lecturer: Andrew Gordon Wilson Scribes: Sai Ganesh Bandiatmakuri 1 Discussion about ML Here we discuss an introduction
More informationA Bayesian perspective on GMM and IV
A Bayesian perspective on GMM and IV Christopher A. Sims Princeton University sims@princeton.edu November 26, 2013 What is a Bayesian perspective? A Bayesian perspective on scientific reporting views all
More informationCBMS Lecture 1. Alan E. Gelfand Duke University
CBMS Lecture 1 Alan E. Gelfand Duke University Introduction to spatial data and models Researchers in diverse areas such as climatology, ecology, environmental exposure, public health, and real estate
More informationParameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets
Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets Matthias Katzfuß Advisor: Dr. Noel Cressie Department of Statistics The Ohio State University
More informationSTAT 518 Intro Student Presentation
STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible
More informationHastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model
UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced
More informationVariable Selection in Structured High-dimensional Covariate Spaces
Variable Selection in Structured High-dimensional Covariate Spaces Fan Li 1 Nancy Zhang 2 1 Department of Health Care Policy Harvard University 2 Department of Statistics Stanford University May 14 2007
More informationA Bayesian Treatment of Linear Gaussian Regression
A Bayesian Treatment of Linear Gaussian Regression Frank Wood December 3, 2009 Bayesian Approach to Classical Linear Regression In classical linear regression we have the following model y β, σ 2, X N(Xβ,
More informationGauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA
JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter
More information