Bayesian Estimation of Expected Cell Counts by Using R

Size: px
Start display at page:

Download "Bayesian Estimation of Expected Cell Counts by Using R"

Transcription

1 Bayesian Estimation of Expected Cell Counts by Using R Haydar Demirhan 1 and Canan Hamurkaroglu 2 Department of Statistics, Hacettepe University, Beytepe, 06800, Ankara, Turkey Abstract In this article, we give a script that is working on the R software for the Bayesian estimation of expected cell counts of a contingency table over a given log linear model The script promotes the usage and applicability of the Bayesian estimation of expected cell counts We illustrate usage and interpretation of the resulting outputs over an example Keywords:Log linear models; WinBugs; Script; Maximum likelihood estimation 1 Introduction Log linear modelling is frequently used in the analysis of the contingency tables When researchers want to include their subjective information into the analysis, Bayesian point of view is a way to include all available information In the log linear modelling, model parameters and expected cell counts of the considered contingency table can be estimated using the Bayesian approaches For each of the purposes, a prior distribution is determined and it is combined with the corresponding likelihood function using the Bayes Theorem Bayesian estimation of the model parameters over a given log linear model is considered by Leighty&Johnson [1] King&Brooks [2] use the prior specified for the log linear model parameters to elicit a prior for the expected cell counts They show consistency of the transition from model parameters to the expected cell counts in representation of the prior information The advantage of using the information on the model parameters for the estimation of expected cell counts is that one should determine prior information for less number of parameters for all of the log linear models except the saturated model Demirhan&Hamurkaroglu [3] show that the transition from prior distribution of model parameters to that of the expected cell counts, which is given by King&Brooks [2], is not appropriate for posterior inferences due to the singularity problem of the variance matrix of the prior distribution induced on the expected cell counts In addition, Demirhan&Hamurkaroglu [3] give an approach to solve the singularity problem They used Metropolis-Hastings (MH) algorithm for posterior inferences instead of the Gibbs sampling, because of the difficulties in the derivation of full conditional distributions The approach given by Demirhan&Hamurkaroglu [3] is not easy to implement for the researchers, who are not familiar with the Bayesian approaches Moreover, one should write his own computer program to implement the MH algorithm for posterior inferences In this article, we give an R script that takes mean vector of prior distribution of the model parameters, 1 haydarde@hacettepeedutr 2 caca@hacettepeedutr 1

2 the belief in the prior information and considered contingency table, makes the transition by using the approach of Demirhan&Hamurkaroglu [3], then draws posterior inferences by using the Gibbs sampling over the well known Bayesian analysis software WinBugs [4] WinBugs does not require the determination of the full conditional distributions by the user [5] The function bugs(), given by Gelman [6], calls WinBugs from R for the Gibbs sampling We use R to make required calculations for the transition and determination of the parameters of prior distribution induced on the expected cell counts, and the bugs() for the posterior inferences By this way, the Gibbs sampling is used instead of the MH algorithm with the approach of Demirhan&Hamurkaroglu [3] Log linear models are frequently used for the analysis of the contingency tables in the medical and social researches In general, the researchers from these fields have little information on the Bayesian estimation Even if one has enough information to understand Bayesian techniques, there are severe difficulties in the calculation of posterior estimates There is not any software that is able to conduct most of the proposed Bayesian approaches Therefore, one has to prepare a computer program to draw posterior inferences in most of the Bayesian approaches Because of these difficulties, researchers are not using the Bayesian approaches that are very useful and practical for their research The given script is a ready to use tool for the mentioned estimation purposes We believe that it will promote the usage of the approach of Demirhan&Hamurkaroglu [3] The approaches given by Leighty&Johnson [1], King&Brooks [2] and Demirhan& Hamurkaroglu [3] are mentioned in Section 2 The R script and discussions on it are given in Section 3 An example is given to clarify the usage and outputs of the script in Section 4 2 Bayesian estimation of the expected cell counts It is assumed that variables constituting the considered contingency table are nominal The log linear is defined as follows: log n = X m β m (1) where β m includes main and interaction effect parameters of a model m X m is corresponding design matrix, composed of -1, 0 and 1 values, and includes a column for each parameter in β m n is the vector of expected cell counts Leighty&Johnson [1] use multivariate normal distribution with parameters µ m and Σ m as the prior distribution of the model parameters µ m is used to represent the prior information on the parameters For example, if one has an information on one of the interaction parameters between two variables that it is negative and weak Then the relevant element of the µ m is chosen as negative and close to zero Σ m represents the researchers belief in the induced prior information They give a two staged approach for the specification of the entries of Σ m At the first stage it is taken as Σ m = αc m = αci m (2) where I m is the identity matrix dimension of p = dim(β m ); and to make C 1 m agree with the precision of the likelihood it is chosen as c = p/tr(v 1 1 b ), where V b is the inverse of variance matrix of the maximum likelihood (ML) estimates [1] On this point we faced with another 2

3 difficulty that the researcher should construct the design matrix by himself and obtain relevant ML estimates using a statistical software before the determination of prior distribution Our script carries on these issues for the user The α, seen in (2), is used to represent the prior belief in the induced prior information Distribution of the α is given by the second stage It is taken that τ = 1/(1 + α) and τ uniform(0, 1) to make calculations and express the belief in the prior easier Close to zero values of τ represent high disbelief and close to one values represent high belief in the prior When the τ is close to zero, the information coming from the sample dominates the prior information In the opposite case, prior information dominates the information provided by the sample King&Brooks [2] give an approach to obtain a prior distribution for the expected cell counts by using the prior distribution of the model parameters and (1) If β m follows N(µ m, Σ m ), then log n follows N(µ, Σ) from (1), where µ = X m µ m and Σ = X m Σ m X T m Proofs of these results are given in [2] As the result, joint prior distribution of the expected cell counts follows a multivariate log normal distribution with parameters µ and Σ, if Σ is a positive definite matrix Demirhan&Hamurkaroglu [3] show that Σ is singular, and hence it is not positive definite Thus, using the transition with this setting is not possible Demirhan&Hamurkaroglu [3] propose an approach to overcome the singularity problem They define the scale matrix of the multivariate normal distribution as Σ = Σ + hi, where h is a constant They show that Σ is positive definite, if h is taken as tr(σ)/dim(σ) e min (Σ), where e min (Σ), tr(σ) and dim(σ) denote the smallest eigenvalue, the trace and the dimension of Σ, respectively Then, Demirhan&Hamurkaroglu [3] use the approach of Leighty&Johnson [1] to specify the variance matrix of the prior distribution of the model parameters As the result, h is taken as following: h = α [ tr(x m C m X T m)/dim(σ) e min (X m C m X T m) ] = αv (3) Demirhan&Hamurkaroglu [3] state that because h is dependent to τ, h has some effect on the degree of belief, however, it has not got any effect on the degree of belief in the prior apart from τ Diagonal elements of Σ increases by the addition of h So, if the diagonal elements of Σ was decreased as much as the value of h before the addition of it then the effect of h may be neutralized on the average Let the value of random variable τ specified to represent our degree of belief in the prior be τ g and the decreased value for τ is obtained by using τ g, be τ s Demirhan&Hamurkaroglu [3] show that if τ s is defined as the following: τ s = τ g ( tr(xm C m X T m)/dim(σ) )(1 τ g ) + τ g, (4) tr(x m C m X T m)/dim(σ) + v effect of h is minimized and also the degree of belief in the prior information is not changed Joint prior distribution of the expected cell counts given τ and h is as follows: p(n τ, h) ( ) { 1 exp 1 } n k 2 (log n µ)t (Σ ) 1 (log n µ), 0 < n k n (5) k K 3

4 In (5), K denotes the considered contingency table and k corresponds to each cell of the table Details of this notation style are given by King&Brooks [2] Under the Poisson sampling plan, the kernel of log likelihood function is as follows: l(n y) = k K y k log n k, (6) where y k is the observed frequency of the cell k Joint posterior distribution of the expected cell counts is obtained from (5) and (6) as follows: p(n τ, h) ( ) { 1 exp 1 } n k 2 (log n µ)t (Σ ) 1 (log n µ) y k log n k (7) k K k K 3 The R script for the Bayesian estimation of expected cell counts The script works on the posterior distribution given in (7), with the prior scheme mentioned in the Section 2 It requires the τ s value, µ m, relevant contingency table and considered model to work It obtains ML estimates of the model parameters under the given model, then calculates elements of µ, h and entries of Σ to specify the prior distribution After the specification of the prior distribution, it calls WinBugs to implement the Gibbs sampling for posterior inferences Obtained posterior estimates, some of the percentiles of the marginal posterior distributions of the expected cell counts and scale reduction factors corresponding to each parameter are displayed by the R All of the statistics available in the procedure bugs, which are mentioned by Gelman[6], can be displayed in console windows of R The script is given below We added line numbers for clarifications 1 library(matrix) 2 library(r2winbugs) 3 library(brugs) 4 tau_g< tablo<-readtable("x://obsdat",header=true, fill=true, blanklinesskip =TRUE) 6 #dataframe(tablo$y,asfactor(tablo$gender),asfactor(tablo$edu),asfactor( tablo$age)) 7 gm<-glm(tablo$y ~ asfactor(tablo$gender) + asfactor(tablo$edu) + asfactor( tablo$age)+asfactor(tablo$gender) * asfactor(tablo$edu) + asfactor( tablo$gender) * asfactor(tablo$age) + asfactor(tablo$edu) * asfactor( tablo$age), family=poisson(log), x=true) 8 tas<-gm$x 9 p<-ncol(tas) 10 r<-nrow(tas) 11 tas_dev<-t(tas) 12 tas_tas_dev<-tas %*% tas_dev 13 prm_kest<-gm$coefficients 14 g1<-tas %*% prm_kest 15 bgs_kest<-exp(g1) 16 d_bgs_kest<-matrix(0,r,r) 17 for(i in 1:r){d_bgs_kest[i,i]=bgs_kest[i]} 18 g2<-tas_dev %*% d_bgs_kest 19 g3<-g2 %*% tas 20 VarCov<-inv(g3) 21 iz_v<-sum(diag(varcov)) 22 C<-diag(p/iz_V,r,r) %*% tas_tas_dev 4

5 23 iz_c<-sum(diag(c)) 24 v<-(iz_c/p)-min(eigen(c,onlyvalues=true)$values) 25 tau_s<-tau_g/((((iz_c/p)/((iz_c/p)+v))*(1-tau_g))+tau_g) 26 alpha<-(1-tau_s)/tau_s 27 h<-alpha*v 28 cov<-diag(p/(alpha*iz_v),r,r) %*% tas_tas_dev + diag(h,r,r) 29 mu_m<-asmatrix(readtable("x://mu_mdat")) 30 mu<-tas %*% mu_m 31 y<-asmatrix(tablo$y) 32 k<-tablo$d_say[1] 33 nn<-vector(mode="numeric", length=r) 34 girdi<-list("r","cov","mu") 35 basl<-function(){list(nn=runif(r,0,1))} 36 pars<-list("nn") 37 sonucsim<-bugs(girdi,basl,pars,"x://modelbug",nchains=3,niter=5000, DIC=FALSE) 38 print(sonucsim) A x:// is seen in the lines 5, 29 and 37 It should be replaced with the relevant path before running the script Lines 1-3 load the required R libraries R2WinBUGS and BRugs are given by Gelman [6] Lines 4 and 5 get the inputs from the user Definitions about the variables are made in the line 6 ML estimates of the log linear parameter are obtained at the line 7 over the model that contains all of the two factor interactions Design matrix X m is taken from the glm procedure of the R at the line 8 Some calculations are carried on at the lines 9-12 ML estimates of the model parameters are taken from the glm procedure at the line 13 Required calculations to obtain v, seen in (3), are done at the lines The v is obtained at the line 24 The τ s, mentioned in (4), is calculated at the line 25 Precision of the prior distribution of the expected cell counts is calculated at the line 26, and the value of h is obtained at the line 27 X, µ m and µ are obtained at the lines 28, 29 and 30 respectively Lines prepare inputs of the bugs procedure Line 37 calls WinBugs for implementation of the Gibbs sampling Outputs generated by WinBugs is taken back to R environment and displayed in R console window by the line 38 The script uses the files obsdat, mu mdat and modelbug Contents of these files are given in the appendix 4 An example We give an example to illustrate the outputs of the script We used gender, age and level of education cross tabulation, which is a cross tabulation The data is taken from the web site of the Turkish Statistical Institute Considered hierarchical model does not contain only three factor interaction term The script given in the Section 3 generates three chains each of which has 5000 iterations Obtained outputs from the implementation of the script are as following: Inference for Bugs model at "x://modelbug", fit using winbugs, 3 chains, each with 5000 iterations (first 500 discarded) nsims = 1500 iterations saved mean sd 25% 25% 50% 75% 975% Rhat neff nn[1] nn[2]

6 nn[3] nn[130] nn[131] nn[132] For each parameter, neff is a crude measure of effective sample size, and Rhat is the potential scale reduction factor (at convergence, Rhat=1) It is seen from the output that first 500 iterations were discarded to reduce the autocorrelation Posterior estimates of the expected cell counts are given in the mean column Estimated standard deviations are under the sd column Next five columns correspond to the 25, 25, 50, 75 and 975th percentiles Potential scale reduction factors are given in the Rhat column It can be concluded that the convergence is achieved for a parameter, if the relevant Rhat value is close to 1 and less than 12 [7] 5 Conclusions We consider the application of the approach, which is on the Bayesian estimation of the expected cell counts of a contingency table over a given log linear model, given by Demirhan& Hamurkaroglu [3] on the R software We prepare a script for this aim The script is practically reasonable for the aims of non-bayesian statisticians or non-statistician researchers For the mentioned estimation purposes one should construct the relevant design matrix, obtain ML estimates, calculate parameters of the prior distribution of the expected cell counts, prepare a computer program to implement the MH algorithm, calculate convergence measures such as Rhat, percentiles of each marginal posterior distribution Following this list is a hard work for most of the researchers Our script provides an easy way One should coy and paste the script into the console window of R, prepare obsdat, which contains observed table, mu mdat, which contains prior information on the model parameters and create a file modelbug by coping it from the Appendix C to obtain Bayesian estimates of the expected cell counts R and WinBugs softwares can be obtained from their internet sites free of charge [8, 4] This is also promotes the usage of the script In conclusion, the given script increases applicability of the approach of the writers, and it is a ready to use tool for all researchers from various fields References [1] Leighty, RM, Johnson, WJ, 1990, A Bayesian loglinear model analysis of categorical data Journal of Official Statistics, 6, 2, [2] King, R, Brooks, SP, 2001, Prior induction in log-linear models for general contingency table analysis The Annals of Statistics, 29, 3, [3] Demirhan, H, Hamurkaroglu, C, 2006, A Bayesian Approach to the Estimation of Expected Cell Counts by Using Log Linear Models Communications in Statistics - Theory and Methods, 35, 2,

7 [4] WinBugs, 2007, Available online at: wwwmrc-bsucamacuk/bugs/winbugs/contentsshtml (accessed 10 June 2007) [5] Spiegelhalter, D, Thomas, A, Best, N, Lunn, D, 2003, WinBugs user manual, v14 Available online at: wwwmrc-bsucamacuk/bugs (accessed 10 June 2007) [6] Gelman, A, 2006, bugsr: functions for running WinBugs and OpenBugs from R Available online at: wwwstatcolumbiaedu/ gelman/bugsr (accessed 10 June 2007) [7] Gelman, A, 1996, Inference and monitoring convergence, in Markov Chain Monte Carlo (eds WR Gilks, S Richardson, DJ Spiegelhalter), Chapman&Hall/CRC, London [8] R project, 2007, Available online at: wwwr-projectorg (accessed 10 June 2007) Appendix A1 Content of the obsdat file Content of the obsdat file is as following: Gender Edu Age y A2 Content of the mu mdat file mu mdat includes the prior information about the value of each parameter one under the other The content is as following: A3 Content of the modelbug file The modelbug contains a model code in WinBugs language The content is as following: model{for(i in 1:r) { y[i,1]~dpois(nn[i]) nn[i]~dlnorm(mu[i,1],cov[i,i]) } } 7

Advanced Statistical Modelling

Advanced Statistical Modelling Markov chain Monte Carlo (MCMC) Methods and Their Applications in Bayesian Statistics School of Technology and Business Studies/Statistics Dalarna University Borlänge, Sweden. Feb. 05, 2014. Outlines 1

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

The Bayesian Approach to Multi-equation Econometric Model Estimation

The Bayesian Approach to Multi-equation Econometric Model Estimation Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation

More information

Bayesian Networks in Educational Assessment

Bayesian Networks in Educational Assessment Bayesian Networks in Educational Assessment Estimating Parameters with MCMC Bayesian Inference: Expanding Our Context Roy Levy Arizona State University Roy.Levy@asu.edu 2017 Roy Levy MCMC 1 MCMC 2 Posterior

More information

eqr094: Hierarchical MCMC for Bayesian System Reliability

eqr094: Hierarchical MCMC for Bayesian System Reliability eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

DIC: Deviance Information Criterion

DIC: Deviance Information Criterion (((( Welcome Page Latest News DIC: Deviance Information Criterion Contact us/bugs list WinBUGS New WinBUGS examples FAQs DIC GeoBUGS DIC (Deviance Information Criterion) is a Bayesian method for model

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Bayesian Inference for Regression Parameters

Bayesian Inference for Regression Parameters Bayesian Inference for Regression Parameters 1 Bayesian inference for simple linear regression parameters follows the usual pattern for all Bayesian analyses: 1. Form a prior distribution over all unknown

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Hierarchical Linear Models

Hierarchical Linear Models Hierarchical Linear Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin The linear regression model Hierarchical Linear Models y N(Xβ, Σ y ) β σ 2 p(β σ 2 ) σ 2 p(σ 2 ) can be extended

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

Bayesian modelling. Hans-Peter Helfrich. University of Bonn. Theodor-Brinkmann-Graduate School

Bayesian modelling. Hans-Peter Helfrich. University of Bonn. Theodor-Brinkmann-Graduate School Bayesian modelling Hans-Peter Helfrich University of Bonn Theodor-Brinkmann-Graduate School H.-P. Helfrich (University of Bonn) Bayesian modelling Brinkmann School 1 / 22 Overview 1 Bayesian modelling

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

Statistics 220 Bayesian Data Analysis

Statistics 220 Bayesian Data Analysis Statistics 220 Bayesian Data Analysis Mark E. Irwin Department of Statistics Harvard University Spring Term Thursday, February 3, 2005 - Tuesday, May 17, 2005 Copyright c 2005 by Mark E. Irwin Personnel

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

WinBUGS : part 2. Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert. Gabriele, living with rheumatoid arthritis

WinBUGS : part 2. Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert. Gabriele, living with rheumatoid arthritis WinBUGS : part 2 Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert Gabriele, living with rheumatoid arthritis Agenda 2! Hierarchical model: linear regression example! R2WinBUGS Linear Regression

More information

36-463/663Multilevel and Hierarchical Models

36-463/663Multilevel and Hierarchical Models 36-463/663Multilevel and Hierarchical Models From Bayes to MCMC to MLMs Brian Junker 132E Baker Hall brian@stat.cmu.edu 1 Outline Bayesian Statistics and MCMC Distribution of Skill Mastery in a Population

More information

Ronald Christensen. University of New Mexico. Albuquerque, New Mexico. Wesley Johnson. University of California, Irvine. Irvine, California

Ronald Christensen. University of New Mexico. Albuquerque, New Mexico. Wesley Johnson. University of California, Irvine. Irvine, California Texts in Statistical Science Bayesian Ideas and Data Analysis An Introduction for Scientists and Statisticians Ronald Christensen University of New Mexico Albuquerque, New Mexico Wesley Johnson University

More information

Bayesian Inference and Decision Theory

Bayesian Inference and Decision Theory Bayesian Inference and Decision Theory Instructor: Kathryn Blackmond Laskey Room 2214 ENGR (703) 993-1644 Office Hours: Tuesday and Thursday 4:30-5:30 PM, or by appointment Spring 2018 Unit 6: Gibbs Sampling

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

MIT /30 Gelman, Carpenter, Hoffman, Guo, Goodrich, Lee,... Stan for Bayesian data analysis

MIT /30 Gelman, Carpenter, Hoffman, Guo, Goodrich, Lee,... Stan for Bayesian data analysis MIT 1985 1/30 Stan: a program for Bayesian data analysis with complex models Andrew Gelman, Bob Carpenter, and Matt Hoffman, Jiqiang Guo, Ben Goodrich, and Daniel Lee Department of Statistics, Columbia

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Markov Chain Monte Carlo in Practice

Markov Chain Monte Carlo in Practice Markov Chain Monte Carlo in Practice Edited by W.R. Gilks Medical Research Council Biostatistics Unit Cambridge UK S. Richardson French National Institute for Health and Medical Research Vilejuif France

More information

Bayesian Methods in Multilevel Regression

Bayesian Methods in Multilevel Regression Bayesian Methods in Multilevel Regression Joop Hox MuLOG, 15 september 2000 mcmc What is Statistics?! Statistics is about uncertainty To err is human, to forgive divine, but to include errors in your design

More information

Item Parameter Calibration of LSAT Items Using MCMC Approximation of Bayes Posterior Distributions

Item Parameter Calibration of LSAT Items Using MCMC Approximation of Bayes Posterior Distributions R U T C O R R E S E A R C H R E P O R T Item Parameter Calibration of LSAT Items Using MCMC Approximation of Bayes Posterior Distributions Douglas H. Jones a Mikhail Nediak b RRR 7-2, February, 2! " ##$%#&

More information

SAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software

SAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software SAMSI Astrostatistics Tutorial More Markov chain Monte Carlo & Demo of Mathematica software Phil Gregory University of British Columbia 26 Bayesian Logical Data Analysis for the Physical Sciences Contents:

More information

MCMC Methods: Gibbs and Metropolis

MCMC Methods: Gibbs and Metropolis MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution

More information

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,

More information

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1 Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data

More information

Bayesian Nonparametric Regression for Diabetes Deaths

Bayesian Nonparametric Regression for Diabetes Deaths Bayesian Nonparametric Regression for Diabetes Deaths Brian M. Hartman PhD Student, 2010 Texas A&M University College Station, TX, USA David B. Dahl Assistant Professor Texas A&M University College Station,

More information

ST 740: Markov Chain Monte Carlo

ST 740: Markov Chain Monte Carlo ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

MALA versus Random Walk Metropolis Dootika Vats June 4, 2017

MALA versus Random Walk Metropolis Dootika Vats June 4, 2017 MALA versus Random Walk Metropolis Dootika Vats June 4, 2017 Introduction My research thus far has predominantly been on output analysis for Markov chain Monte Carlo. The examples on which I have implemented

More information

Bayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014

Bayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014 Bayesian Prediction of Code Output ASA Albuquerque Chapter Short Course October 2014 Abstract This presentation summarizes Bayesian prediction methodology for the Gaussian process (GP) surrogate representation

More information

PIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS

PIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS PIER HLM Course July 30, 2011 Howard Seltman Discussion Guide for Bayes and BUGS 1. Classical Statistics is based on parameters as fixed unknown values. a. The standard approach is to try to discover,

More information

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017 Chalmers April 6, 2017 Bayesian philosophy Bayesian philosophy Bayesian statistics versus classical statistics: War or co-existence? Classical statistics: Models have variables and parameters; these are

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Latent Class Analysis for Models with Error of Measurement Using Log-Linear Models and An Application to Women s Liberation Data

Latent Class Analysis for Models with Error of Measurement Using Log-Linear Models and An Application to Women s Liberation Data Journal of Data Science 9(2011), 43-54 Latent Class Analysis for Models with Error of Measurement Using Log-Linear Models and An Application to Women s Liberation Data Haydar Demirhan Hacettepe University

More information

Markov Chain Monte Carlo and Applied Bayesian Statistics

Markov Chain Monte Carlo and Applied Bayesian Statistics Markov Chain Monte Carlo and Applied Bayesian Statistics Trinity Term 2005 Prof. Gesine Reinert Markov chain Monte Carlo is a stochastic simulation technique that is very useful for computing inferential

More information

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference 1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE

More information

Monte Carlo integration

Monte Carlo integration Monte Carlo integration Eample of a Monte Carlo sampler in D: imagine a circle radius L/ within a square of LL. If points are randoml generated over the square, what s the probabilit to hit within circle?

More information

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci

More information

Doing Bayesian Integrals

Doing Bayesian Integrals ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to

More information

Summary of Talk Background to Multilevel modelling project. What is complex level 1 variation? Tutorial dataset. Method 1 : Inverse Wishart proposals.

Summary of Talk Background to Multilevel modelling project. What is complex level 1 variation? Tutorial dataset. Method 1 : Inverse Wishart proposals. Modelling the Variance : MCMC methods for tting multilevel models with complex level 1 variation and extensions to constrained variance matrices By Dr William Browne Centre for Multilevel Modelling Institute

More information

MCMC Review. MCMC Review. Gibbs Sampling. MCMC Review

MCMC Review. MCMC Review. Gibbs Sampling. MCMC Review MCMC Review http://jackman.stanford.edu/mcmc/icpsr99.pdf http://students.washington.edu/fkrogsta/bayes/stat538.pdf http://www.stat.berkeley.edu/users/terry/classes/s260.1998 /Week9a/week9a/week9a.html

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

10. Exchangeability and hierarchical models Objective. Recommended reading

10. Exchangeability and hierarchical models Objective. Recommended reading 10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.

More information

Multivariate Normal & Wishart

Multivariate Normal & Wishart Multivariate Normal & Wishart Hoff Chapter 7 October 21, 2010 Reading Comprehesion Example Twenty-two children are given a reading comprehsion test before and after receiving a particular instruction method.

More information

Bayesian Phylogenetics:

Bayesian Phylogenetics: Bayesian Phylogenetics: an introduction Marc A. Suchard msuchard@ucla.edu UCLA Who is this man? How sure are you? The one true tree? Methods we ve learned so far try to find a single tree that best describes

More information

Probabilistic Machine Learning

Probabilistic Machine Learning Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent

More information

BUGS Bayesian inference Using Gibbs Sampling

BUGS Bayesian inference Using Gibbs Sampling BUGS Bayesian inference Using Gibbs Sampling Glen DePalma Department of Statistics May 30, 2013 www.stat.purdue.edu/~gdepalma 1 / 20 Bayesian Philosophy I [Pearl] turned Bayesian in 1971, as soon as I

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

Cluster investigations using Disease mapping methods International workshop on Risk Factors for Childhood Leukemia Berlin May

Cluster investigations using Disease mapping methods International workshop on Risk Factors for Childhood Leukemia Berlin May Cluster investigations using Disease mapping methods International workshop on Risk Factors for Childhood Leukemia Berlin May 5-7 2008 Peter Schlattmann Institut für Biometrie und Klinische Epidemiologie

More information

Markov Networks.

Markov Networks. Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function

More information

Markov Chain Monte Carlo, Numerical Integration

Markov Chain Monte Carlo, Numerical Integration Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical

More information

Monetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results

Monetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results Monetary and Exchange Rate Policy Under Remittance Fluctuations Technical Appendix and Additional Results Federico Mandelman February In this appendix, I provide technical details on the Bayesian estimation.

More information

STAT 425: Introduction to Bayesian Analysis

STAT 425: Introduction to Bayesian Analysis STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 2) Fall 2017 1 / 19 Part 2: Markov chain Monte

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Spatial Analysis of Incidence Rates: A Bayesian Approach

Spatial Analysis of Incidence Rates: A Bayesian Approach Spatial Analysis of Incidence Rates: A Bayesian Approach Silvio A. da Silva, Luiz L.M. Melo and Ricardo Ehlers July 2004 Abstract Spatial models have been used in many fields of science where the data

More information

Unfolding techniques for neutron spectrometry

Unfolding techniques for neutron spectrometry Uncertainty Assessment in Computational Dosimetry: A comparison of Approaches Unfolding techniques for neutron spectrometry Physikalisch-Technische Bundesanstalt Braunschweig, Germany Contents of the talk

More information

Statistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS

Statistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS Statistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS 1. (a) The posterior mean estimate of α is 14.27, and the posterior mean for the standard deviation of

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Karl Oskar Ekvall Galin L. Jones University of Minnesota March 12, 2019 Abstract Practically relevant statistical models often give rise to probability distributions that are analytically

More information

Weakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0

More information

A Lightweight Guide on Gibbs Sampling and JAGS

A Lightweight Guide on Gibbs Sampling and JAGS A Lightweight Guide on Gibbs Sampling and JAGS Gianpaolo Coro Istituto di Scienza e Tecnologie dell Informazione A. Faedo, Pisa, Italy 1 Summary In this document we give some insight about how Gibbs Sampling

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture

More information

Downloaded from:

Downloaded from: Camacho, A; Kucharski, AJ; Funk, S; Breman, J; Piot, P; Edmunds, WJ (2014) Potential for large outbreaks of Ebola virus disease. Epidemics, 9. pp. 70-8. ISSN 1755-4365 DOI: https://doi.org/10.1016/j.epidem.2014.09.003

More information

A Review of Pseudo-Marginal Markov Chain Monte Carlo

A Review of Pseudo-Marginal Markov Chain Monte Carlo A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the

More information

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation PRE 905: Multivariate Analysis Spring 2014 Lecture 4 Today s Class The building blocks: The basics of mathematical

More information

Kobe University Repository : Kernel

Kobe University Repository : Kernel Kobe University Repository : Kernel タイトル Title 著者 Author(s) 掲載誌 巻号 ページ Citation 刊行日 Issue date 資源タイプ Resource Type 版区分 Resource Version 権利 Rights DOI URL Note on the Sampling Distribution for the Metropolis-

More information

Bayesian phylogenetics. the one true tree? Bayesian phylogenetics

Bayesian phylogenetics. the one true tree? Bayesian phylogenetics Bayesian phylogenetics the one true tree? the methods we ve learned so far try to get a single tree that best describes the data however, they admit that they don t search everywhere, and that it is difficult

More information

Bayesian hypothesis testing for the distribution of insurance claim counts using the Gibbs sampler

Bayesian hypothesis testing for the distribution of insurance claim counts using the Gibbs sampler Journal of Computational Methods in Sciences and Engineering 5 (2005) 201 214 201 IOS Press Bayesian hypothesis testing for the distribution of insurance claim counts using the Gibbs sampler Athanassios

More information

ComputationalToolsforComparing AsymmetricGARCHModelsviaBayes Factors. RicardoS.Ehlers

ComputationalToolsforComparing AsymmetricGARCHModelsviaBayes Factors. RicardoS.Ehlers ComputationalToolsforComparing AsymmetricGARCHModelsviaBayes Factors RicardoS.Ehlers Laboratório de Estatística e Geoinformação- UFPR http://leg.ufpr.br/ ehlers ehlers@leg.ufpr.br II Workshop on Statistical

More information

Sawtooth Software. CVA/HB Technical Paper TECHNICAL PAPER SERIES

Sawtooth Software. CVA/HB Technical Paper TECHNICAL PAPER SERIES Sawtooth Software TECHNICAL PAPER SERIES CVA/HB Technical Paper Copyright 2002, Sawtooth Software, Inc. 530 W. Fir St. Sequim, WA 98382 (360) 681-2300 www.sawtoothsoftware.com The CVA/HB Technical Paper

More information

Bayes Estimation in Meta-analysis using a linear model theorem

Bayes Estimation in Meta-analysis using a linear model theorem University of Wollongong Research Online Applied Statistics Education and Research Collaboration (ASEARC) - Conference Papers Faculty of Engineering and Information Sciences 2012 Bayes Estimation in Meta-analysis

More information

David Giles Bayesian Econometrics

David Giles Bayesian Econometrics David Giles Bayesian Econometrics 5. Bayesian Computation Historically, the computational "cost" of Bayesian methods greatly limited their application. For instance, by Bayes' Theorem: p(θ y) = p(θ)p(y

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

A Bayesian mixture model for term re-occurrence and burstiness

A Bayesian mixture model for term re-occurrence and burstiness A Bayesian mixture model for term re-occurrence and burstiness Avik Sarkar 1, Paul H Garthwaite 2, Anne De Roeck 1 1 Department of Computing, 2 Department of Statistics The Open University Milton Keynes,

More information

Markov chain Monte Carlo

Markov chain Monte Carlo 1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop

More information

U.P., India Dr.Vijay Kumar, Govt. College for Women, M.A. Road Srinagar (J & K), India ABSTRACT

U.P., India Dr.Vijay Kumar, Govt. College for Women, M.A. Road Srinagar (J & K), India ABSTRACT BAYESIAN ESTIMATION OF THE PARAMETER OF GENERALIZED EXPONENTIAL DISTRIBUTION USING MARKOV CHAIN MONTE CARLO METHOD IN OPEN BUGS FOR INFORMATIVE SET OF PRIORS Dr. Mahmood Alam Khan, Dr. Aijaz Ahmed Hakkak,

More information

Bayesian GLMs and Metropolis-Hastings Algorithm

Bayesian GLMs and Metropolis-Hastings Algorithm Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,

More information

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to

More information

Darwin Uy Math 538 Quiz 4 Dr. Behseta

Darwin Uy Math 538 Quiz 4 Dr. Behseta Darwin Uy Math 538 Quiz 4 Dr. Behseta 1) Section 16.1.4 talks about how when sample size gets large, posterior distributions become approximately normal and centered at the MLE. This is largely due to

More information

Bayesian data analysis in practice: Three simple examples

Bayesian data analysis in practice: Three simple examples Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to

More information

Bayesian time series classification

Bayesian time series classification Bayesian time series classification Peter Sykacek Department of Engineering Science University of Oxford Oxford, OX 3PJ, UK psyk@robots.ox.ac.uk Stephen Roberts Department of Engineering Science University

More information

Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material

Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material Quentin Frederik Gronau 1, Monique Duizer 1, Marjan Bakker

More information

The lmm Package. May 9, Description Some improved procedures for linear mixed models

The lmm Package. May 9, Description Some improved procedures for linear mixed models The lmm Package May 9, 2005 Version 0.3-4 Date 2005-5-9 Title Linear mixed models Author Original by Joseph L. Schafer . Maintainer Jing hua Zhao Description Some improved

More information

Bivariate Degradation Modeling Based on Gamma Process

Bivariate Degradation Modeling Based on Gamma Process Bivariate Degradation Modeling Based on Gamma Process Jinglun Zhou Zhengqiang Pan Member IAENG and Quan Sun Abstract Many highly reliable products have two or more performance characteristics (PCs). The

More information

A Bayesian Approach for Sample Size Determination in Method Comparison Studies

A Bayesian Approach for Sample Size Determination in Method Comparison Studies A Bayesian Approach for Sample Size Determination in Method Comparison Studies Kunshan Yin a, Pankaj K. Choudhary a,1, Diana Varghese b and Steven R. Goodman b a Department of Mathematical Sciences b Department

More information

Spatial Statistics Chapter 4 Basics of Bayesian Inference and Computation

Spatial Statistics Chapter 4 Basics of Bayesian Inference and Computation Spatial Statistics Chapter 4 Basics of Bayesian Inference and Computation So far we have discussed types of spatial data, some basic modeling frameworks and exploratory techniques. We have not discussed

More information

Appendix A Probability Distributions

Appendix A Probability Distributions Appendix A Probability Distributions A.1 Discrete Distributions A.1.1 Binomial Distribution The binomial distribution (Bernoulli model) describes the number of failures, X, in n independent Bernoulli trials.

More information

Forward Problems and their Inverse Solutions

Forward Problems and their Inverse Solutions Forward Problems and their Inverse Solutions Sarah Zedler 1,2 1 King Abdullah University of Science and Technology 2 University of Texas at Austin February, 2013 Outline 1 Forward Problem Example Weather

More information

Report and Opinion 2016;8(6) Analysis of bivariate correlated data under the Poisson-gamma model

Report and Opinion 2016;8(6)   Analysis of bivariate correlated data under the Poisson-gamma model Analysis of bivariate correlated data under the Poisson-gamma model Narges Ramooz, Farzad Eskandari 2. MSc of Statistics, Allameh Tabatabai University, Tehran, Iran 2. Associate professor of Statistics,

More information