Comparing Non-informative Priors for Estimation and Prediction in Spatial Models

Size: px
Start display at page:

Download "Comparing Non-informative Priors for Estimation and Prediction in Spatial Models"

Transcription

1 Environmentrics 00, 1 12 DOI: /env.XXXX Comparing Non-informative Priors for Estimation and Prediction in Spatial Models Regina Wu a and Cari G. Kaufman a Summary: Fitting a Bayesian model to spatial data requires specifying prior distributions for parameters. Current literature focuses on the performance of certain non-informative prior distributions in the context of parameter estimation, but often times, what is of main interest is accuracy of predictions. In this paper, we carry out a simulation study comparing two commonly used non-informative priors for spatial models, the reference prior and Jeffreys-rule prior, and find that even though the reference prior outperforms the Jeffreys-rule prior for parameter estimation, the two perform almost identically in terms of mean square prediction errors and coverage rates in prediction. We do find a systematic relationship between prior choice and prediction credible interval length and we explain why this happens, relating it to previous results on parameter estimation. Our results show that unlike parameter estimation, prior choice makes little difference in prediction. Keywords: reference prior; Jeffreys-rule prior; spatial data; Bayesian kriging 1. INTRODUCTION Spatial data are often modeled by Gaussian random fields where the mean and covariance parameters are unknown. This uncertainty can be accounted for by taking a Bayesian approach and placing priors distributions on the parameters in the assumed model. Choosing a prior distribution can be difficult when the parameters are hard to interpret or when there is little prior knowledge available, so it is useful to have default objective priors which avoid a Department of Statistics, University of California, Berkeley, CA 94720, U.S.A Correspondence to: Regina Wu, Department of Statistics, University of California, Berkeley, CA 94720, U.S.A wuregina88@gmail.com This paper has been submitted for consideration for publication in Environmetrics

2 Environmetrics R. Wu and C. G. Kaufman the need for subjective specifications. The motivation of Berger et al., 2001 is to present non-informative prior distributions that yield proper posterior distributions, as many commonly used priors fail to do. Berger et al., 2001 recommend using the reference prior over the Jeffreys-rule prior, independence Jeffreys prior, and constant prior. In addition to having a proper posterior, Berger et al., 2001 show the reference prior outperforms the Jeffreys-rule prior in terms of parameter estimation as it produces credible intervals for the parameters with close to nominal coverage. Papers such as Kazianka and Pilz, 2012, Ren et al., 2012, and Oliveira, 2007 have extended the study of objective Bayes inference for spatially correlated data by including nugget effects to account for the possibility of measurement errors. When examining frequentist coverage of parameters such as and noise to signal ratio η, the results agree with Berger et al., 2001 where the Jeffreys-rule prior produces credible intervals for the parameters with low coverage rates, especially as p and get larger. The current literature focuses on prior performance with regards to parameter estimation. The motivation for this paper stems from the fact that in geostatistics parameter estimation is usually not performed in isolation, but for the end goal of making predictions of the underlying process at unobserved locations. Hence, it is interesting to look at how the priors compare for predictions. From a Bayesian standpoint, predictions can easily be generated using kriging procedures (see e.g. Handcock and Stein, 1993), but prior choice is again a question. We evaluate the predictive performance of the reference prior and Jeffreys-rule prior introduced in Berger et al.,

3 Comparing Non-informative Priors Environmetrics 2. THE MODEL AND PRIORS 2.1. The Model Let {Z(s), s D} denote a Gaussian random field, with the n observations denoted by Z = (Z(s 1 ),..., Z(s n )) where s 1,..., s n are known distinct sampling locations in D. Our goal is to examine predictions made for Z pred = (Z(s 1),..., Z(s m)), where s 1,..., s m are new locations in D. The mean function of the Gaussian field is E{Z(s)} = β f(s) where β = (β 1,..., β p ) R p are unknown regression parameters and f(s) = (f 1 (s),..., f p (s)) are known locationdependent covariates. The covariance function is cov{z(s), Z(u)} = σ 2 K ( s u ) where denotes Euclidean distance, σ 2 = var{z(s)}, K ( s u ) = corr{z(s), Z(u)} is an isotropic correlation function, and = ( 1,..., c ) Θ R c is the vector of parameters controlling the range of correlation and the smoothness of the field. The likelihood of the model parameters using observed data z = (z 1,..., z n ) where z i represents the observed value of Z(s i ) is { } 1 L(β, σ 2, ; z) = (2πσ 2 ) n/2 Σ 1/2 exp 2σ (z 2 Xβ) Σ 1 (z Xβ) (1) where X is a full rank n p matrix defined by X ij = f j (s i ) and Σ is a positive definite n n matrix defined by Σ,ij = K ( s i s j ) Θ. We focus on the use of spherical, power exponential, rational quadratic, and Matérn covariance functions. Generating samples from the posterior distribution of is most efficiently carried out using the integrated likelihood L I (, z) defined by L(β, σ 2, ; z)π(β, σ 2, )dβdσ 2 = L I (; z)π() R p (0, ) 3

4 Environmetrics R. Wu and C. G. Kaufman where π(β, σ 2, ) and π() denote the prior distributions for (β, σ 2, ) and respectively. Further details on the priors will be given in Section 2.2. Specifically, we have L I (; z) = Σ 1 2 X Σ 1 X 1 2 (S 2 ) n p 2 +a 1 (2) where S 2 = (z X ˆβ ) Σ 1 (z X ˆβ ) is the generalized residual sum of squares and ˆβ = (X Σ 1 X) 1 X Σ 1 z is the generalized least squares estimator of β given. Furthermore, we can write the posterior predictive distribution conditional on Z and, which is the multivariate t-distribution with df = n p + 2a 2 Z pred Z, t n p+2a 2 (Ẑpred, V pred ) (3) with mean Ẑpred and variance V pred defined as Ẑ pred = X pred β + k predσ 1 (z Xβ ) V pred = 1 df 2 S2 V where X pred is a full rank m p matrix defined by X pred,ij = f j (s i ), k pred is a n m matrix with k pred,ij = K ( s i s j ), K pred is a m m matrix with K pred,ij = K ( s i s j ), and V = K pred k pred Σ 1 k pred + [X pred X Σ 1 k pred] (X Σ 1 X) 1 [X pred X Σ 1 k pred] Reference and Jeffreys-rule Priors The prior distributions for (β, σ 2, ) Ω R p (0, ) (0, ) presented in Berger et al., 2001 take the form π(β, σ 2, ) π() (σ 2 ) a. (4) They derive the reference prior for (σ 2, ) which is the Jeffreys-rule prior using a marginal 4

5 Comparing Non-informative Priors Environmetrics model with integrated likelihood defined for (σ 2, ). The reference prior π R (β, σ 2, ) is of the form (4) with a = 1 and π R () { tr[w 2 ] 1 } 1 n p (tr[w ]) 2 2 where W = (( )Σ )Σ 1 P Σ and P Σ = I X(X Σ 1 X) 1 X Σ 1. The independence Jeffreys prior π J1 (β, σ 2, ) is defined by a = 1 and π J1 () {tr[u 2 ] 1n (tr[u ]) 2 } 1 2 where U = (( )Σ )Σ 1. And lastly, the Jeffreys-rule prior πj2 (β, σ 2, ) is defined by a = 1 + p 2 and πj2 () X Σ 1 X 1 2 π J1 () Berger et al., 2001 show that the reference prior and Jeffreys-rule prior always yield proper posteriors under sampling distribution (1), but the independence Jeffreys prior is proper only when 1 is not a column of X. 3. SIMULATION STUDY 3.1. Simulation Set-Up Berger et al., 2001 compare the reference prior and Jeffreys-rule prior by examining the empirical coverage of equal-tailed Bayesian credible intervals for the range parameter and their average log lengths. Both priors produce intervals with reasonable coverage when only an intercept term is included in the mean of the Gaussian process, but the coverage of the Jeffreys-rule prior is very low when additional terms are added, particularly when the spatial correlation is also strong. The authors attribute the poor performance to the increase in degrees of freedom to the posterior for σ, which causes a shift towards smaller σ values. 5

6 Environmetrics R. Wu and C. G. Kaufman This shift is said to cause the 95% credible sets from the marginal posterior distribution for to have poor coverage rates. We replicate and extend the simulation study in Berger et al., 2001 to compare mean squared errors in addition to coverage rates, and posterior predictions of the Gaussian process in addition to inference about. Our simulated spatial data modeled under 6 different Gaussian random fields Z( ) uses the same choices of mean and covariance functions in Berger et al., The mean function E{Z(s)} is either the constant value 0.15 (p=1) or the function x.1y + 9x 2 xy + 1.2y 2 (p=6), and the Matérn covariance function is C(d) which simplifies to the exponential case C(d) =.12exp{ d/} when ν =.5. We let values in the covariance vary across.2,.5, or 1.0. These values give correlations which decay to.05 at distances of.5992, , and For each of the six combinations, we simulate 3000 sets of data at equally-spaced locations in the union of a training set S t = {(x, y) x, y {0, 0.25, 0.5, 0.75, 1}} = {s 1,..., s 25 } and a validation set S v = {(x, y) x, y {.125, 0.375, 0.625, 0.875}} = {s 1,..., s 16}. For each of the 3000 sample replications z j in our training set S t, we sample 1000 values of from the marginal posterior p( z j ) using a random walk Metropolis Hastings (MH) sampler. Candidates are sampled from a normal distribution centered at the previous value, and the first 200 values of each chain are removed for burn-in. The sampling procedure can be carried out using the integrated likelihood L I (; z j ) in (2). For each dataset z j, we take each sample made from p( z j ) and draw from the corresponding multivariate t-distribution p(z pred z j, ) given in (3). The resulting 800 draws are therefore a sample from p(z pred z j ), from which we can calculate the posterior means z pred,j, a vector of length 16. Calculations of mean squared prediction error (MSPE), frequent coverage, and credible length are based on just one arbitrary prediction location s k = (.625,.375). We compute MSPE by averaging over the 3000 replications. Bayesian equaltailed 95% credible intervals (L j, U j ) are calculated for each dataset z j. The frequentist 6

7 Comparing Non-informative Priors Environmetrics coverage is calculated as p = 1 estimate of j=1 I L j <z pred,j <U j. This has an associated standard error p(1 p). Lastly, we calculate the average credible interval lengths Simulation Results and Interpretation Our first parameter setting follows the same set-up as in Berger et al., 2001, in which the data are simulated under range parameter ν =.5 and this is treated as known. Results in Table 1 show that the MSPE s are very similar between the two priors. For p = 1, the Jeffreysrule prior has higher coverage rates than the reference prior and for p = 6 the reference prior has higher coverage rates than the Jeffreys-rule prior. It seems that this ordering is driven by the credible interval length. The Jeffreys-rule prior performs well here, in contrast to the results for parameter estimation in Berger et al., 2001, in which it had very poor performance, especially when p = 6. It should be noted that differences in interval length and coverage when doing predictions are, practically speaking, very small, in contrast to the large differences observed for parameter estimation in Berger et al., [Table 1 about here.] The similarity in predictive performance across priors is further illustrated in the forest plots of Figure 1. In each plot, the same random subset of 10 out of the 3000 datasets is taken and for each prior we show the predictive samples at the location s k = (.625,.375) in the validation set S v. The samples are centered by the true value for easier comparison and the posterior means are plotted along with 95% equal-tailed credible intervals. The striking similarities between predictions made from the two priors imply that prior choice has little effect on predictive inference. [Figure 1 about here.] Although the similarity between the priors in terms of prediction is strong, we still seek to understand the root of the differences in credible interval lengths, since this is the suspected 7

8 Environmetrics R. Wu and C. G. Kaufman cause of the differences in frequentist coverage rates. In particular, we will look for differences in p( Z) across priors that may be related to differences in spread of p(z pred Z). This relationship is somewhat subtle, depending on both the marginal posterior p( Z) for the different priors, as well as the way a particular influences V ar[z pred,jk Z j, ] for a given prior. For a dataset Z j, the prediction interval for Z pred,jk should be wider for those with larger V ar[z pred,jk Z j ]. We can decompose this to V ar[z pred,jk Z j ] = E (V ar[z pred,jk Z j, ]) + V ar (E[Z pred,jk Z j, ]) (5) where V ar[z pred,jk Z j, ] and E[Z pred,jk Z j, ] can be calculated using the t-distribution in (3) for a given Z j and sample. The outer expectation and variance are with respect to the posterior distribution p( Z j ). Therefore, we can approximate each term in the sum using the empirical mean and variance of these terms calculated using the MCMC samples. We find that the E (V ar[z pred,jk Z j, ]) component consistently dominates the sum, with the V ar (E[Z pred,jk Z j, ]) component having negligible magnitude in comparison. Furthermore, no consistent ordering is found in V ar (E[Z pred,jk Z j, ]) between priors, but E (V ar[z pred,jk Z j, ]) follows the ordering of the priors with regards to credible interval lengths observed earlier in Table 1. To understand why this ordering occurs, we focus on the predictions when p = 6, since this is when the difference between the priors is largest. We illustrate the result for a randomly chosen dataset and the same arbitrary prediction location s k = (.625,.375). A sequence of s are plotted with the corresponding V ar[z pred,jk Z j, ] values. The resulting curves are shown in Figure 2, with a = 1 corresponding to the reference prior and a = 1 + p 2 corresponding to the Jeffreys-rule prior. Other differences between these two priors are not important in calculating V ar (E[Z pred,jk Z j, ]). Specifically, for given Z j and, the values of V ar[z pred,jk Z j, ] differ only by the multiplicative constant n 2 n p 2 across the different 8

9 Comparing Non-informative Priors Environmetrics priors. Along the x-axis are rug plots of the posterior samples under each prior and along the y-axis are the rug plots of the corresponding V ar[z pred,jk Z j, ] values for each prior. The horizontal lines indicate the estimated value of E(V ar[z pred,jk Z j, ]) under each prior. [Figure 2 about here.] The curves in Figure 2 indicate that for a particular prior, having larger values correspond to smaller V ar[z pred,jk Z j, ] values. Judging by the rug plot, if V ar[z pred,jk Z j, ] did not differ by the multiplicative constant, the reference prior should have the lowest E(V ar[z pred,jk Z j, ]) since it has slightly larger samples on average, but instead we see it has the highest E(V ar[z pred,jk Z j, ]). This is due to the upward shift of the true V ar[z pred,jk Z j, ] curve under the reference prior which completely drowns out the effect of having larger samples. Hence, E(V ar[z pred,jk Z j, ]) is larger under the reference prior than under the Jeffreys-rule prior, which is exactly the ordering of the credible interval lengths seen in the p = 6 case in Table 1 and in the ordering of the E(V ar[z pred,jk Z j, ]) component of Equation (5). When the simulation is re-run under a larger range parameter of ν = 1.5, the credible interval lengths is higher for the Jeffreys-rule prior than the reference prior in the p = 6 case. Looking at Figure 3, we see that the reference prior still has the largest samples on average, but this time the curve for V ar[z pred,jk Z j, ] under the reference prior is only shifted up slightly. Because of the closeness of the curve to that of the Jeffreys-rule, the effect of having larger samples is present and the expected ordering of E(V ar[z pred,jk Z j, ]) is maintained unlike in the ν =.5 case. [Figure 3 about here.] 9

10 Environmetrics R. Wu and C. G. Kaufman 4. ILLUSTRATION Unlike global climate models (GCMs), regional climate models (RCMs) are used to model climate system evolution in a limited area through discretized versions of physical processes. The higher resolution in RCMs is beneficial as it better captures the impact of local features such as lakes and mountains as well as subgrid-scale atmospheric processes that can only be approximated in GCMs. Because the RCMs have limited area, they require boundary conditions which are commonly provided by the output of GCMs (Kaufman and Sain, 2010). To illustrate the conclusions from the simulation study, we observe average summer temperatures with model output from a RCM called CHRM with boundary conditions from a GCM called HADAM as shown in Figure 4. This data consists of 2663 average summer temperatures. Temperatures at 671 equally spaced locations (25%) are used as the training set and the remaining 1,992 locations (75%) are used as the validation set. The three covariates are longitude (lon), latitude (lat), and elevation (elev). We model the mean as E{Z(s)} = β 1 + β 2 lon + β 3 lat + β 4 elev + β 5 elev 2 (p = 5) and the covariance as the exponential covariance function. For each prior, we sample 6000 values of from the marginal posterior p( z) using a random walk Metropolis Hastings sampler as before. The first 1000 values of each chain are removed for burn-in and every 10th sample is kept leaving 500 values. For each remaining, we obtain sample predictions at the 1,992 locations by drawing from the corresponding multivariate t-distribution p(z pred z, ). The MSPE and average credible interval lengths are calculated by averaging over the 1,992 locations. The observed coverage rate is calculated as ˆp = 1 1,992 1,992 k=1 I L k <z pred,k <U k. Table 2 shows that the two priors perform almost identically in terms of prediction. In fact, there are no significant differences in observed coverage nor in MSPE across priors. 10

11 Comparing Non-informative Priors Environmetrics [Figure 4 about here.] [Table 2 about here.] 5. DISCUSSION In this paper, we extended the current literature on prior performance in parameter estimation into prediction. Although the reference prior is the recommended prior in Berger et al., 2001, a key finding of this paper is that the reference prior and Jeffreys-rule prior perform almost identically for prediction. We discovered a systematic relationship between prior choice and credible interval lengths based on the posterior predictive distribution. Specifically, we decompose V ar[z pred,jk Z j ] to find the ordering is driven by the E(V ar[z pred,jk Z j, ]) component in (5). The observed ordering in credible interval lengths is dependent on the degree of separation between the a = 1 curve and a = 1 + p 2 curves seen in Figure 2 and Figure 3. The larger the separation, the less we see the effect of the reference prior having larger values. In our simulation, we assumed the range parameter ν is known, so it would be interesting to see the effects when the range parameter is fixed at an incorrect value or estimated. Furthermore, it would be interesting to extend the simulation to see how well the priors perform for prediction when there is a nugget present. REFERENCES Berger JO, Oliveira VD, Sansó B, Objective Bayesian analysis of spatially correlated data. Journal of the American Statistical Association 96(456): Handcock MS, Stein ML, A Bayesian analysis of kriging. Technometrics 35(4):

12 Environmetrics R. Wu and C. G. Kaufman Kaufman CG, Sain SR, Bayesian functional ANOVA modeling using gaussian process prior distributions. Bayesian Analysis 5(1): Kazianka H, Pilz J, Objective bayesian analysis of spatial data with uncertain nugget and range parameters. The Canadian Journal of Statistics 40(2): Oliveira VD, Objective bayesian analysis of spatial data with measurement error. The Canadian Journal of Statistics 35(2): Ren C, Sun D, He C, Objective bayesian analysis for a spatial model with nuggest effects. Journal of Statistical Planning and Inference 142(7):

13 FIGURES Environmetrics FIGURES Figure 1. Forest plots of centered prediction samples grouped by dataset (Dashed Line with circle: reference prior, Solid Line with mark x: Jeffreys-rule prior). =0.2 =0.5 =1 p= p=

14 Environmetrics FIGURES Figure 2. Plot of against V ar[z pred,jk Z j, ]. Upper curve calculated using a = 1 and lower curve using a = 1 + p 2. Rug plot of samples along x-axis and rug plot of corresponding V ar[z pred,jk Z j, ] values along y-axis. Horizontal lines indicate E(V ar[z pred,jk Z j, ]) values calculated from MCMC samples. (Dashed line: reference prior, Solid line: Jeffreys-rule prior) Var[Z pred, jk Z j, ] Jeffreys rule prior Reference prior Reference prior Jeffreys rule prior 14

15 FIGURES Environmetrics Figure 3. Plot of against V ar[z pred,jk Z j, ] values. Upper curve calculated using a = 1 and lower curve using a = 1 + p 2. Rug plot of samples along x-axis and rug plot of corresponding V ar[z pred,jk Z j, ] values along y-axis. Horizontal lines indicate E(V ar[z pred,jk Z j, ]) values calculated from MCMC samples. (Dashed line: reference prior, Solid line: Jeffreys-rule prior) Var[Z pred, jk Z j, ] Jeffreys rule prior Reference prior Reference prior Jeffreys rule prior 15

16 Environmetrics FIGURES Figure 4. Average summer temperatures (C ) of the Prudence Project experiment between

17 TABLES Environmetrics TABLES Table 1. Mean-squared prediction error, frequentist coverage of predictions, and credible interval length results from simulation =.2 =.5 = 1 Mean Squared Prediction Error (and S.E.) p = 1, ν =.5 Reference prior.0729 (.0019).0332 (.0008).0169 (.0004) Jeffreys-rule prior.0732 (.0019).0334 (.0008).0170 (.0004) Prediction reference prior.0728 (.0019).0332 (.0008).0169 (.0004) p = 6, ν =.5 Reference prior.0742 (.0019).0330 (.0008).0174 (.0005) Jeffreys-rule prior.0762 (.0019).0346 (.0009).0184 (.0005) Prediction reference prior.0745 (.0019).0333 (.0008).0176 (.0005) Frequentist Coverage (and S.E.) p = 1, ν =.5 Reference prior.9480 (.0041).9580 (.0037).9647 (.0034) Jeffreys-rule prior.9483 (.0040).9593 (.0036).9660 (.0033) Prediction reference prior.9433 (.0042).9530 (.0039).9613 (.0035) p = 6, ν =.5 Reference prior.9513 (.0039).9660 (.0033).9603 (.0036) Jeffreys-rule Prior.9243 (.0048).9380 (.0044).9413 (.0043) Prediction reference prior.9180(.0050).9310 (.0046).9357 (.0045) Credible Interval Lengths (and S.E.) p = 1, ν =.5 Reference prior (.0034).7746 (.0023).5649 (.0017) Jeffreys-rule prior (.0034).7807 (.0024).5673 (.0017) Prediction reference prior (.0032).7613 (.0023).5554 (.0017) p = 6, ν =.5 Reference prior (.0035).8048 (.0025).5857 (.0018) Jeffreys-rule prior (.0032).7382 (.0023).5386 (.0017) Prediction reference prior (.0030).7061 (.0022).5140 (.0016) 17

18 Environmetrics TABLES Table 2. Mean-squared prediction error, observed coverage of predictions, and credible interval length results from illustration Mean Squared Prediction Error (and S.E.) Reference prior.2156(.0214) Jeffreys-rule prior.2150 (.0214) Observed Coverage (and S.E.) Reference prior.9487 (.0049) Jeffreys-rule prior.9483 (.0049) Credible Interval Lengths (and S.E.) Reference prior (.0070) Jeffreys-rule prior (.0069) 18

Comparing Non-informative Priors for Estimation and. Prediction in Spatial Models

Comparing Non-informative Priors for Estimation and. Prediction in Spatial Models Comparing Non-informative Priors for Estimation and Prediction in Spatial Models Vigre Semester Report by: Regina Wu Advisor: Cari Kaufman January 31, 2010 1 Introduction Gaussian random fields with specified

More information

Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study

Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study Gunter Spöck, Hannes Kazianka, Jürgen Pilz Department of Statistics, University of Klagenfurt, Austria hannes.kazianka@uni-klu.ac.at

More information

Bayesian Transformed Gaussian Random Field: A Review

Bayesian Transformed Gaussian Random Field: A Review Bayesian Transformed Gaussian Random Field: A Review Benjamin Kedem Department of Mathematics & ISR University of Maryland College Park, MD (Victor De Oliveira, David Bindel, Boris and Sandra Kozintsev)

More information

Statistics for analyzing and modeling precipitation isotope ratios in IsoMAP

Statistics for analyzing and modeling precipitation isotope ratios in IsoMAP Statistics for analyzing and modeling precipitation isotope ratios in IsoMAP The IsoMAP uses the multiple linear regression and geostatistical methods to analyze isotope data Suppose the response variable

More information

Hierarchical Modeling for Univariate Spatial Data

Hierarchical Modeling for Univariate Spatial Data Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Bayesian spatial prediction of the site index in the study of the Missouri Ozark Forest Ecosystem Project

Bayesian spatial prediction of the site index in the study of the Missouri Ozark Forest Ecosystem Project Computational Statistics and Data Analysis 52 (2008) 3749 3764 www.elsevier.com/locate/csda Bayesian spatial prediction of the site index in the study of the Missouri Ozark Forest Ecosystem Project Xiaoqian

More information

Chapter 4 - Fundamentals of spatial processes Lecture notes

Chapter 4 - Fundamentals of spatial processes Lecture notes Chapter 4 - Fundamentals of spatial processes Lecture notes Geir Storvik January 21, 2013 STK4150 - Intro 2 Spatial processes Typically correlation between nearby sites Mostly positive correlation Negative

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Chapter 4 - Fundamentals of spatial processes Lecture notes

Chapter 4 - Fundamentals of spatial processes Lecture notes TK4150 - Intro 1 Chapter 4 - Fundamentals of spatial processes Lecture notes Odd Kolbjørnsen and Geir Storvik January 30, 2017 STK4150 - Intro 2 Spatial processes Typically correlation between nearby sites

More information

g-priors for Linear Regression

g-priors for Linear Regression Stat60: Bayesian Modeling and Inference Lecture Date: March 15, 010 g-priors for Linear Regression Lecturer: Michael I. Jordan Scribe: Andrew H. Chan 1 Linear regression and g-priors In the last lecture,

More information

Overall Objective Priors

Overall Objective Priors Overall Objective Priors Jim Berger, Jose Bernardo and Dongchu Sun Duke University, University of Valencia and University of Missouri Recent advances in statistical inference: theory and case studies University

More information

Hierarchical Modeling for Spatial Data

Hierarchical Modeling for Spatial Data Bayesian Spatial Modelling Spatial model specifications: P(y X, θ). Prior specifications: P(θ). Posterior inference of model parameters: P(θ y). Predictions at new locations: P(y 0 y). Model comparisons.

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Model Selection for Geostatistical Models

Model Selection for Geostatistical Models Model Selection for Geostatistical Models Richard A. Davis Colorado State University http://www.stat.colostate.edu/~rdavis/lectures Joint work with: Jennifer A. Hoeting, Colorado State University Andrew

More information

ST 740: Markov Chain Monte Carlo

ST 740: Markov Chain Monte Carlo ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:

More information

A Bayesian perspective on GMM and IV

A Bayesian perspective on GMM and IV A Bayesian perspective on GMM and IV Christopher A. Sims Princeton University sims@princeton.edu November 26, 2013 What is a Bayesian perspective? A Bayesian perspective on scientific reporting views all

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee September 03 05, 2017 Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles Linear Regression Linear regression is,

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

The linear model is the most fundamental of all serious statistical models encompassing:

The linear model is the most fundamental of all serious statistical models encompassing: Linear Regression Models: A Bayesian perspective Ingredients of a linear model include an n 1 response vector y = (y 1,..., y n ) T and an n p design matrix (e.g. including regressors) X = [x 1,..., x

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.

More information

ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS

ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, N.C.,

More information

COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION

COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION SEAN GERRISH AND CHONG WANG 1. WAYS OF ORGANIZING MODELS In probabilistic modeling, there are several ways of organizing models:

More information

spbayes: An R Package for Univariate and Multivariate Hierarchical Point-referenced Spatial Models

spbayes: An R Package for Univariate and Multivariate Hierarchical Point-referenced Spatial Models spbayes: An R Package for Univariate and Multivariate Hierarchical Point-referenced Spatial Models Andrew O. Finley 1, Sudipto Banerjee 2, and Bradley P. Carlin 2 1 Michigan State University, Departments

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Analysing geoadditive regression data: a mixed model approach

Analysing geoadditive regression data: a mixed model approach Analysing geoadditive regression data: a mixed model approach Institut für Statistik, Ludwig-Maximilians-Universität München Joint work with Ludwig Fahrmeir & Stefan Lang 25.11.2005 Spatio-temporal regression

More information

Multivariate Gaussian Random Fields with SPDEs

Multivariate Gaussian Random Fields with SPDEs Multivariate Gaussian Random Fields with SPDEs Xiangping Hu Daniel Simpson, Finn Lindgren and Håvard Rue Department of Mathematics, University of Oslo PASI, 214 Outline The Matérn covariance function and

More information

Models for spatial data (cont d) Types of spatial data. Types of spatial data (cont d) Hierarchical models for spatial data

Models for spatial data (cont d) Types of spatial data. Types of spatial data (cont d) Hierarchical models for spatial data Hierarchical models for spatial data Based on the book by Banerjee, Carlin and Gelfand Hierarchical Modeling and Analysis for Spatial Data, 2004. We focus on Chapters 1, 2 and 5. Geo-referenced data arise

More information

Cross-covariance Functions for Tangent Vector Fields on the Sphere

Cross-covariance Functions for Tangent Vector Fields on the Sphere Cross-covariance Functions for Tangent Vector Fields on the Sphere Minjie Fan 1 Tomoko Matsuo 2 1 Department of Statistics University of California, Davis 2 Cooperative Institute for Research in Environmental

More information

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b) LECTURE 5 NOTES 1. Bayesian point estimators. In the conventional (frequentist) approach to statistical inference, the parameter θ Θ is considered a fixed quantity. In the Bayesian approach, it is considered

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

BAYESIAN KRIGING AND BAYESIAN NETWORK DESIGN

BAYESIAN KRIGING AND BAYESIAN NETWORK DESIGN BAYESIAN KRIGING AND BAYESIAN NETWORK DESIGN Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, N.C., U.S.A. J. Stuart Hunter Lecture TIES 2004

More information

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department

More information

Climate Change: the Uncertainty of Certainty

Climate Change: the Uncertainty of Certainty Climate Change: the Uncertainty of Certainty Reinhard Furrer, UZH JSS, Geneva Oct. 30, 2009 Collaboration with: Stephan Sain - NCAR Reto Knutti - ETHZ Claudia Tebaldi - Climate Central Ryan Ford, Doug

More information

Default priors and model parametrization

Default priors and model parametrization 1 / 16 Default priors and model parametrization Nancy Reid O-Bayes09, June 6, 2009 Don Fraser, Elisabeta Marras, Grace Yun-Yi 2 / 16 Well-calibrated priors model f (y; θ), F(y; θ); log-likelihood l(θ)

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Bruno Sansó. Department of Applied Mathematics and Statistics University of California Santa Cruz bruno

Bruno Sansó. Department of Applied Mathematics and Statistics University of California Santa Cruz   bruno Bruno Sansó Department of Applied Mathematics and Statistics University of California Santa Cruz http://www.ams.ucsc.edu/ bruno Climate Models Climate Models use the equations of motion to simulate changes

More information

Uncertainty and regional climate experiments

Uncertainty and regional climate experiments Uncertainty and regional climate experiments Stephan R. Sain Geophysical Statistics Project Institute for Mathematics Applied to Geosciences National Center for Atmospheric Research Boulder, CO Linda Mearns,

More information

Non-gaussian spatiotemporal modeling

Non-gaussian spatiotemporal modeling Dec, 2008 1/ 37 Non-gaussian spatiotemporal modeling Thais C O da Fonseca Joint work with Prof Mark F J Steel Department of Statistics University of Warwick Dec, 2008 Dec, 2008 2/ 37 1 Introduction Motivation

More information

Consistent high-dimensional Bayesian variable selection via penalized credible regions

Consistent high-dimensional Bayesian variable selection via penalized credible regions Consistent high-dimensional Bayesian variable selection via penalized credible regions Howard Bondell bondell@stat.ncsu.edu Joint work with Brian Reich Howard Bondell p. 1 Outline High-Dimensional Variable

More information

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment Ben Shaby SAMSI August 3, 2010 Ben Shaby (SAMSI) OFS adjustment August 3, 2010 1 / 29 Outline 1 Introduction 2 Spatial

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Carl N. Morris. University of Texas

Carl N. Morris. University of Texas EMPIRICAL BAYES: A FREQUENCY-BAYES COMPROMISE Carl N. Morris University of Texas Empirical Bayes research has expanded significantly since the ground-breaking paper (1956) of Herbert Robbins, and its province

More information

Nearest Neighbor Gaussian Processes for Large Spatial Data

Nearest Neighbor Gaussian Processes for Large Spatial Data Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns

More information

Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Mela. P.

Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Mela. P. Frailty Modeling for Spatially Correlated Survival Data, with Application to Infant Mortality in Minnesota By: Sudipto Banerjee, Melanie M. Wall, Bradley P. Carlin November 24, 2014 Outlines of the talk

More information

Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields

Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields 1 Introduction Jo Eidsvik Department of Mathematical Sciences, NTNU, Norway. (joeid@math.ntnu.no) February

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School

More information

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter

More information

Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes

Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes TTU, October 26, 2012 p. 1/3 Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes Hao Zhang Department of Statistics Department of Forestry and Natural Resources Purdue University

More information

Bayesian Methods in Multilevel Regression

Bayesian Methods in Multilevel Regression Bayesian Methods in Multilevel Regression Joop Hox MuLOG, 15 september 2000 mcmc What is Statistics?! Statistics is about uncertainty To err is human, to forgive divine, but to include errors in your design

More information

What s New in Econometrics. Lecture 13

What s New in Econometrics. Lecture 13 What s New in Econometrics Lecture 13 Weak Instruments and Many Instruments Guido Imbens NBER Summer Institute, 2007 Outline 1. Introduction 2. Motivation 3. Weak Instruments 4. Many Weak) Instruments

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging

A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging Cheng Li DSAP, National University of Singapore Joint work with Rajarshi Guhaniyogi (UC Santa Cruz), Terrance D. Savitsky (US Bureau of Labor

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary

More information

STATISTICAL MODELS FOR QUANTIFYING THE SPATIAL DISTRIBUTION OF SEASONALLY DERIVED OZONE STANDARDS

STATISTICAL MODELS FOR QUANTIFYING THE SPATIAL DISTRIBUTION OF SEASONALLY DERIVED OZONE STANDARDS STATISTICAL MODELS FOR QUANTIFYING THE SPATIAL DISTRIBUTION OF SEASONALLY DERIVED OZONE STANDARDS Eric Gilleland Douglas Nychka Geophysical Statistics Project National Center for Atmospheric Research Supported

More information

Some Curiosities Arising in Objective Bayesian Analysis

Some Curiosities Arising in Objective Bayesian Analysis . Some Curiosities Arising in Objective Bayesian Analysis Jim Berger Duke University Statistical and Applied Mathematical Institute Yale University May 15, 2009 1 Three vignettes related to John s work

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

Introduction to Spatial Data and Models

Introduction to Spatial Data and Models Introduction to Spatial Data and Models Researchers in diverse areas such as climatology, ecology, environmental health, and real estate marketing are increasingly faced with the task of analyzing data

More information

Bayesian Gaussian / Linear Models. Read Sections and 3.3 in the text by Bishop

Bayesian Gaussian / Linear Models. Read Sections and 3.3 in the text by Bishop Bayesian Gaussian / Linear Models Read Sections 2.3.3 and 3.3 in the text by Bishop Multivariate Gaussian Model with Multivariate Gaussian Prior Suppose we model the observed vector b as having a multivariate

More information

Spatio-temporal precipitation modeling based on time-varying regressions

Spatio-temporal precipitation modeling based on time-varying regressions Spatio-temporal precipitation modeling based on time-varying regressions Oleg Makhnin Department of Mathematics New Mexico Tech Socorro, NM 87801 January 19, 2007 1 Abstract: A time-varying regression

More information

Bayesian inference for factor scores

Bayesian inference for factor scores Bayesian inference for factor scores Murray Aitkin and Irit Aitkin School of Mathematics and Statistics University of Newcastle UK October, 3 Abstract Bayesian inference for the parameters of the factor

More information

Hierarchical Modelling for Multivariate Spatial Data

Hierarchical Modelling for Multivariate Spatial Data Hierarchical Modelling for Multivariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Point-referenced spatial data often come as

More information

Empirical Bayes methods for the transformed Gaussian random field model with additive measurement errors

Empirical Bayes methods for the transformed Gaussian random field model with additive measurement errors 1 Empirical Bayes methods for the transformed Gaussian random field model with additive measurement errors Vivekananda Roy Evangelos Evangelou Zhengyuan Zhu CONTENTS 1.1 Introduction......................................................

More information

Introduction to Geostatistics

Introduction to Geostatistics Introduction to Geostatistics Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns Hopkins University, Baltimore,

More information

Statistical analysis of regional climate models. Douglas Nychka, National Center for Atmospheric Research

Statistical analysis of regional climate models. Douglas Nychka, National Center for Atmospheric Research Statistical analysis of regional climate models. Douglas Nychka, National Center for Atmospheric Research National Science Foundation Olso workshop, February 2010 Outline Regional models and the NARCCAP

More information

A STATISTICAL TECHNIQUE FOR MODELLING NON-STATIONARY SPATIAL PROCESSES

A STATISTICAL TECHNIQUE FOR MODELLING NON-STATIONARY SPATIAL PROCESSES A STATISTICAL TECHNIQUE FOR MODELLING NON-STATIONARY SPATIAL PROCESSES JOHN STEPHENSON 1, CHRIS HOLMES, KERRY GALLAGHER 1 and ALEXANDRE PINTORE 1 Dept. Earth Science and Engineering, Imperial College,

More information

Making rating curves - the Bayesian approach

Making rating curves - the Bayesian approach Making rating curves - the Bayesian approach Rating curves what is wanted? A best estimate of the relationship between stage and discharge at a given place in a river. The relationship should be on the

More information

STAT763: Applied Regression Analysis. Multiple linear regression. 4.4 Hypothesis testing

STAT763: Applied Regression Analysis. Multiple linear regression. 4.4 Hypothesis testing STAT763: Applied Regression Analysis Multiple linear regression 4.4 Hypothesis testing Chunsheng Ma E-mail: cma@math.wichita.edu 4.4.1 Significance of regression Null hypothesis (Test whether all β j =

More information

Modelling geoadditive survival data

Modelling geoadditive survival data Modelling geoadditive survival data Thomas Kneib & Ludwig Fahrmeir Department of Statistics, Ludwig-Maximilians-University Munich 1. Leukemia survival data 2. Structured hazard regression 3. Mixed model

More information

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Jeffrey N. Rouder Francis Tuerlinckx Paul L. Speckman Jun Lu & Pablo Gomez May 4 008 1 The Weibull regression model

More information

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall Machine Learning Gaussian Mixture Models Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall 2012 1 The Generative Model POV We think of the data as being generated from some process. We assume

More information

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n

More information

Estimating marginal likelihoods from the posterior draws through a geometric identity

Estimating marginal likelihoods from the posterior draws through a geometric identity Estimating marginal likelihoods from the posterior draws through a geometric identity Johannes Reichl Energy Institute at the Johannes Kepler University Linz E-mail for correspondence: reichl@energieinstitut-linz.at

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Multi-resolution models for large data sets

Multi-resolution models for large data sets Multi-resolution models for large data sets Douglas Nychka, National Center for Atmospheric Research National Science Foundation NORDSTAT, Umeå, June, 2012 Credits Steve Sain, NCAR Tia LeRud, UC Davis

More information

Part 8: GLMs and Hierarchical LMs and GLMs

Part 8: GLMs and Hierarchical LMs and GLMs Part 8: GLMs and Hierarchical LMs and GLMs 1 Example: Song sparrow reproductive success Arcese et al., (1992) provide data on a sample from a population of 52 female song sparrows studied over the course

More information

Learning the hyper-parameters. Luca Martino

Learning the hyper-parameters. Luca Martino Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth

More information

Bayesian analysis of conditional autoregressive models

Bayesian analysis of conditional autoregressive models Ann Inst Stat Math 202) 64:07 33 DOI 0.007/s0463-00-0298- Bayesian analysis of conditional autoregressive models Victor De Oliveira Received: 8 December 2008 / Revised: 4 January 200 / Published online:

More information

Gaussian Processes for Computer Experiments

Gaussian Processes for Computer Experiments Gaussian Processes for Computer Experiments Jeremy Oakley School of Mathematics and Statistics, University of Sheffield www.jeremy-oakley.staff.shef.ac.uk 1 / 43 Computer models Computer model represented

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33 Hypothesis Testing Econ 690 Purdue University Justin L. Tobias (Purdue) Testing 1 / 33 Outline 1 Basic Testing Framework 2 Testing with HPD intervals 3 Example 4 Savage Dickey Density Ratio 5 Bartlett

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Alan Gelfand 1 and Andrew O. Finley 2 1 Department of Statistical Science, Duke University, Durham, North

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo 1 Motivation 1.1 Bayesian Learning Markov Chain Monte Carlo Yale Chang In Bayesian learning, given data X, we make assumptions on the generative process of X by introducing hidden variables Z: p(z): prior

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Foundations of Statistical Inference

Foundations of Statistical Inference Foundations of Statistical Inference Julien Berestycki Department of Statistics University of Oxford MT 2015 Julien Berestycki (University of Oxford) SB2a MT 2015 1 / 16 Lecture 16 : Bayesian analysis

More information

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative

More information

Bayesian Analysis of Latent Variable Models using Mplus

Bayesian Analysis of Latent Variable Models using Mplus Bayesian Analysis of Latent Variable Models using Mplus Tihomir Asparouhov and Bengt Muthén Version 2 June 29, 2010 1 1 Introduction In this paper we describe some of the modeling possibilities that are

More information

A Framework for Daily Spatio-Temporal Stochastic Weather Simulation

A Framework for Daily Spatio-Temporal Stochastic Weather Simulation A Framework for Daily Spatio-Temporal Stochastic Weather Simulation, Rick Katz, Balaji Rajagopalan Geophysical Statistics Project Institute for Mathematics Applied to Geosciences National Center for Atmospheric

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

Math 423/533: The Main Theoretical Topics

Math 423/533: The Main Theoretical Topics Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)

More information