Bayesian Estimation of Expected Cell Counts by Using R
|
|
- Griselda Gibbs
- 5 years ago
- Views:
Transcription
1 Bayesian Estimation of Expected Cell Counts by Using R Haydar Demirhan 1 and Canan Hamurkaroglu 2 Department of Statistics, Hacettepe University, Beytepe, 06800, Ankara, Turkey Abstract In this article, we give a script that is working on the R software for the Bayesian estimation of expected cell counts of a contingency table over a given log linear model The script promotes the usage and applicability of the Bayesian estimation of expected cell counts We illustrate usage and interpretation of the resulting outputs over an example Keywords:Log linear models; WinBugs; Script; Maximum likelihood estimation 1 Introduction Log linear modelling is frequently used in the analysis of the contingency tables When researchers want to include their subjective information into the analysis, Bayesian point of view is a way to include all available information In the log linear modelling, model parameters and expected cell counts of the considered contingency table can be estimated using the Bayesian approaches For each of the purposes, a prior distribution is determined and it is combined with the corresponding likelihood function using the Bayes Theorem Bayesian estimation of the model parameters over a given log linear model is considered by Leighty&Johnson [1] King&Brooks [2] use the prior specified for the log linear model parameters to elicit a prior for the expected cell counts They show consistency of the transition from model parameters to the expected cell counts in representation of the prior information The advantage of using the information on the model parameters for the estimation of expected cell counts is that one should determine prior information for less number of parameters for all of the log linear models except the saturated model Demirhan&Hamurkaroglu [3] show that the transition from prior distribution of model parameters to that of the expected cell counts, which is given by King&Brooks [2], is not appropriate for posterior inferences due to the singularity problem of the variance matrix of the prior distribution induced on the expected cell counts In addition, Demirhan&Hamurkaroglu [3] give an approach to solve the singularity problem They used Metropolis-Hastings (MH) algorithm for posterior inferences instead of the Gibbs sampling, because of the difficulties in the derivation of full conditional distributions The approach given by Demirhan&Hamurkaroglu [3] is not easy to implement for the researchers, who are not familiar with the Bayesian approaches Moreover, one should write his own computer program to implement the MH algorithm for posterior inferences In this article, we give an R script that takes mean vector of prior distribution of the model parameters, 1 haydarde@hacettepeedutr 2 caca@hacettepeedutr 1
2 the belief in the prior information and considered contingency table, makes the transition by using the approach of Demirhan&Hamurkaroglu [3], then draws posterior inferences by using the Gibbs sampling over the well known Bayesian analysis software WinBugs [4] WinBugs does not require the determination of the full conditional distributions by the user [5] The function bugs(), given by Gelman [6], calls WinBugs from R for the Gibbs sampling We use R to make required calculations for the transition and determination of the parameters of prior distribution induced on the expected cell counts, and the bugs() for the posterior inferences By this way, the Gibbs sampling is used instead of the MH algorithm with the approach of Demirhan&Hamurkaroglu [3] Log linear models are frequently used for the analysis of the contingency tables in the medical and social researches In general, the researchers from these fields have little information on the Bayesian estimation Even if one has enough information to understand Bayesian techniques, there are severe difficulties in the calculation of posterior estimates There is not any software that is able to conduct most of the proposed Bayesian approaches Therefore, one has to prepare a computer program to draw posterior inferences in most of the Bayesian approaches Because of these difficulties, researchers are not using the Bayesian approaches that are very useful and practical for their research The given script is a ready to use tool for the mentioned estimation purposes We believe that it will promote the usage of the approach of Demirhan&Hamurkaroglu [3] The approaches given by Leighty&Johnson [1], King&Brooks [2] and Demirhan& Hamurkaroglu [3] are mentioned in Section 2 The R script and discussions on it are given in Section 3 An example is given to clarify the usage and outputs of the script in Section 4 2 Bayesian estimation of the expected cell counts It is assumed that variables constituting the considered contingency table are nominal The log linear is defined as follows: log n = X m β m (1) where β m includes main and interaction effect parameters of a model m X m is corresponding design matrix, composed of -1, 0 and 1 values, and includes a column for each parameter in β m n is the vector of expected cell counts Leighty&Johnson [1] use multivariate normal distribution with parameters µ m and Σ m as the prior distribution of the model parameters µ m is used to represent the prior information on the parameters For example, if one has an information on one of the interaction parameters between two variables that it is negative and weak Then the relevant element of the µ m is chosen as negative and close to zero Σ m represents the researchers belief in the induced prior information They give a two staged approach for the specification of the entries of Σ m At the first stage it is taken as Σ m = αc m = αci m (2) where I m is the identity matrix dimension of p = dim(β m ); and to make C 1 m agree with the precision of the likelihood it is chosen as c = p/tr(v 1 1 b ), where V b is the inverse of variance matrix of the maximum likelihood (ML) estimates [1] On this point we faced with another 2
3 difficulty that the researcher should construct the design matrix by himself and obtain relevant ML estimates using a statistical software before the determination of prior distribution Our script carries on these issues for the user The α, seen in (2), is used to represent the prior belief in the induced prior information Distribution of the α is given by the second stage It is taken that τ = 1/(1 + α) and τ uniform(0, 1) to make calculations and express the belief in the prior easier Close to zero values of τ represent high disbelief and close to one values represent high belief in the prior When the τ is close to zero, the information coming from the sample dominates the prior information In the opposite case, prior information dominates the information provided by the sample King&Brooks [2] give an approach to obtain a prior distribution for the expected cell counts by using the prior distribution of the model parameters and (1) If β m follows N(µ m, Σ m ), then log n follows N(µ, Σ) from (1), where µ = X m µ m and Σ = X m Σ m X T m Proofs of these results are given in [2] As the result, joint prior distribution of the expected cell counts follows a multivariate log normal distribution with parameters µ and Σ, if Σ is a positive definite matrix Demirhan&Hamurkaroglu [3] show that Σ is singular, and hence it is not positive definite Thus, using the transition with this setting is not possible Demirhan&Hamurkaroglu [3] propose an approach to overcome the singularity problem They define the scale matrix of the multivariate normal distribution as Σ = Σ + hi, where h is a constant They show that Σ is positive definite, if h is taken as tr(σ)/dim(σ) e min (Σ), where e min (Σ), tr(σ) and dim(σ) denote the smallest eigenvalue, the trace and the dimension of Σ, respectively Then, Demirhan&Hamurkaroglu [3] use the approach of Leighty&Johnson [1] to specify the variance matrix of the prior distribution of the model parameters As the result, h is taken as following: h = α [ tr(x m C m X T m)/dim(σ) e min (X m C m X T m) ] = αv (3) Demirhan&Hamurkaroglu [3] state that because h is dependent to τ, h has some effect on the degree of belief, however, it has not got any effect on the degree of belief in the prior apart from τ Diagonal elements of Σ increases by the addition of h So, if the diagonal elements of Σ was decreased as much as the value of h before the addition of it then the effect of h may be neutralized on the average Let the value of random variable τ specified to represent our degree of belief in the prior be τ g and the decreased value for τ is obtained by using τ g, be τ s Demirhan&Hamurkaroglu [3] show that if τ s is defined as the following: τ s = τ g ( tr(xm C m X T m)/dim(σ) )(1 τ g ) + τ g, (4) tr(x m C m X T m)/dim(σ) + v effect of h is minimized and also the degree of belief in the prior information is not changed Joint prior distribution of the expected cell counts given τ and h is as follows: p(n τ, h) ( ) { 1 exp 1 } n k 2 (log n µ)t (Σ ) 1 (log n µ), 0 < n k n (5) k K 3
4 In (5), K denotes the considered contingency table and k corresponds to each cell of the table Details of this notation style are given by King&Brooks [2] Under the Poisson sampling plan, the kernel of log likelihood function is as follows: l(n y) = k K y k log n k, (6) where y k is the observed frequency of the cell k Joint posterior distribution of the expected cell counts is obtained from (5) and (6) as follows: p(n τ, h) ( ) { 1 exp 1 } n k 2 (log n µ)t (Σ ) 1 (log n µ) y k log n k (7) k K k K 3 The R script for the Bayesian estimation of expected cell counts The script works on the posterior distribution given in (7), with the prior scheme mentioned in the Section 2 It requires the τ s value, µ m, relevant contingency table and considered model to work It obtains ML estimates of the model parameters under the given model, then calculates elements of µ, h and entries of Σ to specify the prior distribution After the specification of the prior distribution, it calls WinBugs to implement the Gibbs sampling for posterior inferences Obtained posterior estimates, some of the percentiles of the marginal posterior distributions of the expected cell counts and scale reduction factors corresponding to each parameter are displayed by the R All of the statistics available in the procedure bugs, which are mentioned by Gelman[6], can be displayed in console windows of R The script is given below We added line numbers for clarifications 1 library(matrix) 2 library(r2winbugs) 3 library(brugs) 4 tau_g< tablo<-readtable("x://obsdat",header=true, fill=true, blanklinesskip =TRUE) 6 #dataframe(tablo$y,asfactor(tablo$gender),asfactor(tablo$edu),asfactor( tablo$age)) 7 gm<-glm(tablo$y ~ asfactor(tablo$gender) + asfactor(tablo$edu) + asfactor( tablo$age)+asfactor(tablo$gender) * asfactor(tablo$edu) + asfactor( tablo$gender) * asfactor(tablo$age) + asfactor(tablo$edu) * asfactor( tablo$age), family=poisson(log), x=true) 8 tas<-gm$x 9 p<-ncol(tas) 10 r<-nrow(tas) 11 tas_dev<-t(tas) 12 tas_tas_dev<-tas %*% tas_dev 13 prm_kest<-gm$coefficients 14 g1<-tas %*% prm_kest 15 bgs_kest<-exp(g1) 16 d_bgs_kest<-matrix(0,r,r) 17 for(i in 1:r){d_bgs_kest[i,i]=bgs_kest[i]} 18 g2<-tas_dev %*% d_bgs_kest 19 g3<-g2 %*% tas 20 VarCov<-inv(g3) 21 iz_v<-sum(diag(varcov)) 22 C<-diag(p/iz_V,r,r) %*% tas_tas_dev 4
5 23 iz_c<-sum(diag(c)) 24 v<-(iz_c/p)-min(eigen(c,onlyvalues=true)$values) 25 tau_s<-tau_g/((((iz_c/p)/((iz_c/p)+v))*(1-tau_g))+tau_g) 26 alpha<-(1-tau_s)/tau_s 27 h<-alpha*v 28 cov<-diag(p/(alpha*iz_v),r,r) %*% tas_tas_dev + diag(h,r,r) 29 mu_m<-asmatrix(readtable("x://mu_mdat")) 30 mu<-tas %*% mu_m 31 y<-asmatrix(tablo$y) 32 k<-tablo$d_say[1] 33 nn<-vector(mode="numeric", length=r) 34 girdi<-list("r","cov","mu") 35 basl<-function(){list(nn=runif(r,0,1))} 36 pars<-list("nn") 37 sonucsim<-bugs(girdi,basl,pars,"x://modelbug",nchains=3,niter=5000, DIC=FALSE) 38 print(sonucsim) A x:// is seen in the lines 5, 29 and 37 It should be replaced with the relevant path before running the script Lines 1-3 load the required R libraries R2WinBUGS and BRugs are given by Gelman [6] Lines 4 and 5 get the inputs from the user Definitions about the variables are made in the line 6 ML estimates of the log linear parameter are obtained at the line 7 over the model that contains all of the two factor interactions Design matrix X m is taken from the glm procedure of the R at the line 8 Some calculations are carried on at the lines 9-12 ML estimates of the model parameters are taken from the glm procedure at the line 13 Required calculations to obtain v, seen in (3), are done at the lines The v is obtained at the line 24 The τ s, mentioned in (4), is calculated at the line 25 Precision of the prior distribution of the expected cell counts is calculated at the line 26, and the value of h is obtained at the line 27 X, µ m and µ are obtained at the lines 28, 29 and 30 respectively Lines prepare inputs of the bugs procedure Line 37 calls WinBugs for implementation of the Gibbs sampling Outputs generated by WinBugs is taken back to R environment and displayed in R console window by the line 38 The script uses the files obsdat, mu mdat and modelbug Contents of these files are given in the appendix 4 An example We give an example to illustrate the outputs of the script We used gender, age and level of education cross tabulation, which is a cross tabulation The data is taken from the web site of the Turkish Statistical Institute Considered hierarchical model does not contain only three factor interaction term The script given in the Section 3 generates three chains each of which has 5000 iterations Obtained outputs from the implementation of the script are as following: Inference for Bugs model at "x://modelbug", fit using winbugs, 3 chains, each with 5000 iterations (first 500 discarded) nsims = 1500 iterations saved mean sd 25% 25% 50% 75% 975% Rhat neff nn[1] nn[2]
6 nn[3] nn[130] nn[131] nn[132] For each parameter, neff is a crude measure of effective sample size, and Rhat is the potential scale reduction factor (at convergence, Rhat=1) It is seen from the output that first 500 iterations were discarded to reduce the autocorrelation Posterior estimates of the expected cell counts are given in the mean column Estimated standard deviations are under the sd column Next five columns correspond to the 25, 25, 50, 75 and 975th percentiles Potential scale reduction factors are given in the Rhat column It can be concluded that the convergence is achieved for a parameter, if the relevant Rhat value is close to 1 and less than 12 [7] 5 Conclusions We consider the application of the approach, which is on the Bayesian estimation of the expected cell counts of a contingency table over a given log linear model, given by Demirhan& Hamurkaroglu [3] on the R software We prepare a script for this aim The script is practically reasonable for the aims of non-bayesian statisticians or non-statistician researchers For the mentioned estimation purposes one should construct the relevant design matrix, obtain ML estimates, calculate parameters of the prior distribution of the expected cell counts, prepare a computer program to implement the MH algorithm, calculate convergence measures such as Rhat, percentiles of each marginal posterior distribution Following this list is a hard work for most of the researchers Our script provides an easy way One should coy and paste the script into the console window of R, prepare obsdat, which contains observed table, mu mdat, which contains prior information on the model parameters and create a file modelbug by coping it from the Appendix C to obtain Bayesian estimates of the expected cell counts R and WinBugs softwares can be obtained from their internet sites free of charge [8, 4] This is also promotes the usage of the script In conclusion, the given script increases applicability of the approach of the writers, and it is a ready to use tool for all researchers from various fields References [1] Leighty, RM, Johnson, WJ, 1990, A Bayesian loglinear model analysis of categorical data Journal of Official Statistics, 6, 2, [2] King, R, Brooks, SP, 2001, Prior induction in log-linear models for general contingency table analysis The Annals of Statistics, 29, 3, [3] Demirhan, H, Hamurkaroglu, C, 2006, A Bayesian Approach to the Estimation of Expected Cell Counts by Using Log Linear Models Communications in Statistics - Theory and Methods, 35, 2,
7 [4] WinBugs, 2007, Available online at: wwwmrc-bsucamacuk/bugs/winbugs/contentsshtml (accessed 10 June 2007) [5] Spiegelhalter, D, Thomas, A, Best, N, Lunn, D, 2003, WinBugs user manual, v14 Available online at: wwwmrc-bsucamacuk/bugs (accessed 10 June 2007) [6] Gelman, A, 2006, bugsr: functions for running WinBugs and OpenBugs from R Available online at: wwwstatcolumbiaedu/ gelman/bugsr (accessed 10 June 2007) [7] Gelman, A, 1996, Inference and monitoring convergence, in Markov Chain Monte Carlo (eds WR Gilks, S Richardson, DJ Spiegelhalter), Chapman&Hall/CRC, London [8] R project, 2007, Available online at: wwwr-projectorg (accessed 10 June 2007) Appendix A1 Content of the obsdat file Content of the obsdat file is as following: Gender Edu Age y A2 Content of the mu mdat file mu mdat includes the prior information about the value of each parameter one under the other The content is as following: A3 Content of the modelbug file The modelbug contains a model code in WinBugs language The content is as following: model{for(i in 1:r) { y[i,1]~dpois(nn[i]) nn[i]~dlnorm(mu[i,1],cov[i,i]) } } 7
Advanced Statistical Modelling
Markov chain Monte Carlo (MCMC) Methods and Their Applications in Bayesian Statistics School of Technology and Business Studies/Statistics Dalarna University Borlänge, Sweden. Feb. 05, 2014. Outlines 1
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationThe Bayesian Approach to Multi-equation Econometric Model Estimation
Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation
More informationBayesian Networks in Educational Assessment
Bayesian Networks in Educational Assessment Estimating Parameters with MCMC Bayesian Inference: Expanding Our Context Roy Levy Arizona State University Roy.Levy@asu.edu 2017 Roy Levy MCMC 1 MCMC 2 Posterior
More informationeqr094: Hierarchical MCMC for Bayesian System Reliability
eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationDIC: Deviance Information Criterion
(((( Welcome Page Latest News DIC: Deviance Information Criterion Contact us/bugs list WinBUGS New WinBUGS examples FAQs DIC GeoBUGS DIC (Deviance Information Criterion) is a Bayesian method for model
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationBayesian Inference for Regression Parameters
Bayesian Inference for Regression Parameters 1 Bayesian inference for simple linear regression parameters follows the usual pattern for all Bayesian analyses: 1. Form a prior distribution over all unknown
More informationIntroduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016
Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An
More informationHierarchical Linear Models
Hierarchical Linear Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin The linear regression model Hierarchical Linear Models y N(Xβ, Σ y ) β σ 2 p(β σ 2 ) σ 2 p(σ 2 ) can be extended
More informationContents. Part I: Fundamentals of Bayesian Inference 1
Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian
More informationBayesian modelling. Hans-Peter Helfrich. University of Bonn. Theodor-Brinkmann-Graduate School
Bayesian modelling Hans-Peter Helfrich University of Bonn Theodor-Brinkmann-Graduate School H.-P. Helfrich (University of Bonn) Bayesian modelling Brinkmann School 1 / 22 Overview 1 Bayesian modelling
More informationMultilevel Statistical Models: 3 rd edition, 2003 Contents
Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction
More informationStatistics 220 Bayesian Data Analysis
Statistics 220 Bayesian Data Analysis Mark E. Irwin Department of Statistics Harvard University Spring Term Thursday, February 3, 2005 - Tuesday, May 17, 2005 Copyright c 2005 by Mark E. Irwin Personnel
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationWinBUGS : part 2. Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert. Gabriele, living with rheumatoid arthritis
WinBUGS : part 2 Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert Gabriele, living with rheumatoid arthritis Agenda 2! Hierarchical model: linear regression example! R2WinBUGS Linear Regression
More information36-463/663Multilevel and Hierarchical Models
36-463/663Multilevel and Hierarchical Models From Bayes to MCMC to MLMs Brian Junker 132E Baker Hall brian@stat.cmu.edu 1 Outline Bayesian Statistics and MCMC Distribution of Skill Mastery in a Population
More informationRonald Christensen. University of New Mexico. Albuquerque, New Mexico. Wesley Johnson. University of California, Irvine. Irvine, California
Texts in Statistical Science Bayesian Ideas and Data Analysis An Introduction for Scientists and Statisticians Ronald Christensen University of New Mexico Albuquerque, New Mexico Wesley Johnson University
More informationBayesian Inference and Decision Theory
Bayesian Inference and Decision Theory Instructor: Kathryn Blackmond Laskey Room 2214 ENGR (703) 993-1644 Office Hours: Tuesday and Thursday 4:30-5:30 PM, or by appointment Spring 2018 Unit 6: Gibbs Sampling
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationMIT /30 Gelman, Carpenter, Hoffman, Guo, Goodrich, Lee,... Stan for Bayesian data analysis
MIT 1985 1/30 Stan: a program for Bayesian data analysis with complex models Andrew Gelman, Bob Carpenter, and Matt Hoffman, Jiqiang Guo, Ben Goodrich, and Daniel Lee Department of Statistics, Columbia
More informationBayesian Linear Regression
Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective
More informationMarkov Chain Monte Carlo in Practice
Markov Chain Monte Carlo in Practice Edited by W.R. Gilks Medical Research Council Biostatistics Unit Cambridge UK S. Richardson French National Institute for Health and Medical Research Vilejuif France
More informationBayesian Methods in Multilevel Regression
Bayesian Methods in Multilevel Regression Joop Hox MuLOG, 15 september 2000 mcmc What is Statistics?! Statistics is about uncertainty To err is human, to forgive divine, but to include errors in your design
More informationItem Parameter Calibration of LSAT Items Using MCMC Approximation of Bayes Posterior Distributions
R U T C O R R E S E A R C H R E P O R T Item Parameter Calibration of LSAT Items Using MCMC Approximation of Bayes Posterior Distributions Douglas H. Jones a Mikhail Nediak b RRR 7-2, February, 2! " ##$%#&
More informationSAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software
SAMSI Astrostatistics Tutorial More Markov chain Monte Carlo & Demo of Mathematica software Phil Gregory University of British Columbia 26 Bayesian Logical Data Analysis for the Physical Sciences Contents:
More informationMCMC Methods: Gibbs and Metropolis
MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution
More informationProbabilistic Graphical Models Lecture 17: Markov chain Monte Carlo
Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,
More informationParameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1
Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data
More informationBayesian Nonparametric Regression for Diabetes Deaths
Bayesian Nonparametric Regression for Diabetes Deaths Brian M. Hartman PhD Student, 2010 Texas A&M University College Station, TX, USA David B. Dahl Assistant Professor Texas A&M University College Station,
More informationST 740: Markov Chain Monte Carlo
ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:
More informationMetropolis-Hastings Algorithm
Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to
More informationMALA versus Random Walk Metropolis Dootika Vats June 4, 2017
MALA versus Random Walk Metropolis Dootika Vats June 4, 2017 Introduction My research thus far has predominantly been on output analysis for Markov chain Monte Carlo. The examples on which I have implemented
More informationBayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014
Bayesian Prediction of Code Output ASA Albuquerque Chapter Short Course October 2014 Abstract This presentation summarizes Bayesian prediction methodology for the Gaussian process (GP) surrogate representation
More informationPIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS
PIER HLM Course July 30, 2011 Howard Seltman Discussion Guide for Bayes and BUGS 1. Classical Statistics is based on parameters as fixed unknown values. a. The standard approach is to try to discover,
More informationBayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017
Chalmers April 6, 2017 Bayesian philosophy Bayesian philosophy Bayesian statistics versus classical statistics: War or co-existence? Classical statistics: Models have variables and parameters; these are
More informationHastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model
UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced
More informationLatent Class Analysis for Models with Error of Measurement Using Log-Linear Models and An Application to Women s Liberation Data
Journal of Data Science 9(2011), 43-54 Latent Class Analysis for Models with Error of Measurement Using Log-Linear Models and An Application to Women s Liberation Data Haydar Demirhan Hacettepe University
More informationMarkov Chain Monte Carlo and Applied Bayesian Statistics
Markov Chain Monte Carlo and Applied Bayesian Statistics Trinity Term 2005 Prof. Gesine Reinert Markov chain Monte Carlo is a stochastic simulation technique that is very useful for computing inferential
More informationBayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference
1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE
More informationMonte Carlo integration
Monte Carlo integration Eample of a Monte Carlo sampler in D: imagine a circle radius L/ within a square of LL. If points are randoml generated over the square, what s the probabilit to hit within circle?
More informationBAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA
BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci
More informationDoing Bayesian Integrals
ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to
More informationSummary of Talk Background to Multilevel modelling project. What is complex level 1 variation? Tutorial dataset. Method 1 : Inverse Wishart proposals.
Modelling the Variance : MCMC methods for tting multilevel models with complex level 1 variation and extensions to constrained variance matrices By Dr William Browne Centre for Multilevel Modelling Institute
More informationMCMC Review. MCMC Review. Gibbs Sampling. MCMC Review
MCMC Review http://jackman.stanford.edu/mcmc/icpsr99.pdf http://students.washington.edu/fkrogsta/bayes/stat538.pdf http://www.stat.berkeley.edu/users/terry/classes/s260.1998 /Week9a/week9a/week9a.html
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More information10. Exchangeability and hierarchical models Objective. Recommended reading
10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.
More informationMultivariate Normal & Wishart
Multivariate Normal & Wishart Hoff Chapter 7 October 21, 2010 Reading Comprehesion Example Twenty-two children are given a reading comprehsion test before and after receiving a particular instruction method.
More informationBayesian Phylogenetics:
Bayesian Phylogenetics: an introduction Marc A. Suchard msuchard@ucla.edu UCLA Who is this man? How sure are you? The one true tree? Methods we ve learned so far try to find a single tree that best describes
More informationProbabilistic Machine Learning
Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent
More informationBUGS Bayesian inference Using Gibbs Sampling
BUGS Bayesian inference Using Gibbs Sampling Glen DePalma Department of Statistics May 30, 2013 www.stat.purdue.edu/~gdepalma 1 / 20 Bayesian Philosophy I [Pearl] turned Bayesian in 1971, as soon as I
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationCluster investigations using Disease mapping methods International workshop on Risk Factors for Childhood Leukemia Berlin May
Cluster investigations using Disease mapping methods International workshop on Risk Factors for Childhood Leukemia Berlin May 5-7 2008 Peter Schlattmann Institut für Biometrie und Klinische Epidemiologie
More informationMarkov Networks.
Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function
More informationMarkov Chain Monte Carlo, Numerical Integration
Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical
More informationMonetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results
Monetary and Exchange Rate Policy Under Remittance Fluctuations Technical Appendix and Additional Results Federico Mandelman February In this appendix, I provide technical details on the Bayesian estimation.
More informationSTAT 425: Introduction to Bayesian Analysis
STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 2) Fall 2017 1 / 19 Part 2: Markov chain Monte
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationSpatial Analysis of Incidence Rates: A Bayesian Approach
Spatial Analysis of Incidence Rates: A Bayesian Approach Silvio A. da Silva, Luiz L.M. Melo and Ricardo Ehlers July 2004 Abstract Spatial models have been used in many fields of science where the data
More informationUnfolding techniques for neutron spectrometry
Uncertainty Assessment in Computational Dosimetry: A comparison of Approaches Unfolding techniques for neutron spectrometry Physikalisch-Technische Bundesanstalt Braunschweig, Germany Contents of the talk
More informationStatistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS
Statistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS 1. (a) The posterior mean estimate of α is 14.27, and the posterior mean for the standard deviation of
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Karl Oskar Ekvall Galin L. Jones University of Minnesota March 12, 2019 Abstract Practically relevant statistical models often give rise to probability distributions that are analytically
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More informationA Lightweight Guide on Gibbs Sampling and JAGS
A Lightweight Guide on Gibbs Sampling and JAGS Gianpaolo Coro Istituto di Scienza e Tecnologie dell Informazione A. Faedo, Pisa, Italy 1 Summary In this document we give some insight about how Gibbs Sampling
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationThe Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations
The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture
More informationDownloaded from:
Camacho, A; Kucharski, AJ; Funk, S; Breman, J; Piot, P; Edmunds, WJ (2014) Potential for large outbreaks of Ebola virus disease. Epidemics, 9. pp. 70-8. ISSN 1755-4365 DOI: https://doi.org/10.1016/j.epidem.2014.09.003
More informationA Review of Pseudo-Marginal Markov Chain Monte Carlo
A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the
More informationUnivariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation
Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation PRE 905: Multivariate Analysis Spring 2014 Lecture 4 Today s Class The building blocks: The basics of mathematical
More informationKobe University Repository : Kernel
Kobe University Repository : Kernel タイトル Title 著者 Author(s) 掲載誌 巻号 ページ Citation 刊行日 Issue date 資源タイプ Resource Type 版区分 Resource Version 権利 Rights DOI URL Note on the Sampling Distribution for the Metropolis-
More informationBayesian phylogenetics. the one true tree? Bayesian phylogenetics
Bayesian phylogenetics the one true tree? the methods we ve learned so far try to get a single tree that best describes the data however, they admit that they don t search everywhere, and that it is difficult
More informationBayesian hypothesis testing for the distribution of insurance claim counts using the Gibbs sampler
Journal of Computational Methods in Sciences and Engineering 5 (2005) 201 214 201 IOS Press Bayesian hypothesis testing for the distribution of insurance claim counts using the Gibbs sampler Athanassios
More informationComputationalToolsforComparing AsymmetricGARCHModelsviaBayes Factors. RicardoS.Ehlers
ComputationalToolsforComparing AsymmetricGARCHModelsviaBayes Factors RicardoS.Ehlers Laboratório de Estatística e Geoinformação- UFPR http://leg.ufpr.br/ ehlers ehlers@leg.ufpr.br II Workshop on Statistical
More informationSawtooth Software. CVA/HB Technical Paper TECHNICAL PAPER SERIES
Sawtooth Software TECHNICAL PAPER SERIES CVA/HB Technical Paper Copyright 2002, Sawtooth Software, Inc. 530 W. Fir St. Sequim, WA 98382 (360) 681-2300 www.sawtoothsoftware.com The CVA/HB Technical Paper
More informationBayes Estimation in Meta-analysis using a linear model theorem
University of Wollongong Research Online Applied Statistics Education and Research Collaboration (ASEARC) - Conference Papers Faculty of Engineering and Information Sciences 2012 Bayes Estimation in Meta-analysis
More informationDavid Giles Bayesian Econometrics
David Giles Bayesian Econometrics 5. Bayesian Computation Historically, the computational "cost" of Bayesian methods greatly limited their application. For instance, by Bayes' Theorem: p(θ y) = p(θ)p(y
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationA Bayesian mixture model for term re-occurrence and burstiness
A Bayesian mixture model for term re-occurrence and burstiness Avik Sarkar 1, Paul H Garthwaite 2, Anne De Roeck 1 1 Department of Computing, 2 Department of Statistics The Open University Milton Keynes,
More informationMarkov chain Monte Carlo
1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop
More informationU.P., India Dr.Vijay Kumar, Govt. College for Women, M.A. Road Srinagar (J & K), India ABSTRACT
BAYESIAN ESTIMATION OF THE PARAMETER OF GENERALIZED EXPONENTIAL DISTRIBUTION USING MARKOV CHAIN MONTE CARLO METHOD IN OPEN BUGS FOR INFORMATIVE SET OF PRIORS Dr. Mahmood Alam Khan, Dr. Aijaz Ahmed Hakkak,
More informationBayesian GLMs and Metropolis-Hastings Algorithm
Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,
More informationA quick introduction to Markov chains and Markov chain Monte Carlo (revised version)
A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to
More informationDarwin Uy Math 538 Quiz 4 Dr. Behseta
Darwin Uy Math 538 Quiz 4 Dr. Behseta 1) Section 16.1.4 talks about how when sample size gets large, posterior distributions become approximately normal and centered at the MLE. This is largely due to
More informationBayesian data analysis in practice: Three simple examples
Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to
More informationBayesian time series classification
Bayesian time series classification Peter Sykacek Department of Engineering Science University of Oxford Oxford, OX 3PJ, UK psyk@robots.ox.ac.uk Stephen Roberts Department of Engineering Science University
More informationBayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material
Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material Quentin Frederik Gronau 1, Monique Duizer 1, Marjan Bakker
More informationThe lmm Package. May 9, Description Some improved procedures for linear mixed models
The lmm Package May 9, 2005 Version 0.3-4 Date 2005-5-9 Title Linear mixed models Author Original by Joseph L. Schafer . Maintainer Jing hua Zhao Description Some improved
More informationBivariate Degradation Modeling Based on Gamma Process
Bivariate Degradation Modeling Based on Gamma Process Jinglun Zhou Zhengqiang Pan Member IAENG and Quan Sun Abstract Many highly reliable products have two or more performance characteristics (PCs). The
More informationA Bayesian Approach for Sample Size Determination in Method Comparison Studies
A Bayesian Approach for Sample Size Determination in Method Comparison Studies Kunshan Yin a, Pankaj K. Choudhary a,1, Diana Varghese b and Steven R. Goodman b a Department of Mathematical Sciences b Department
More informationSpatial Statistics Chapter 4 Basics of Bayesian Inference and Computation
Spatial Statistics Chapter 4 Basics of Bayesian Inference and Computation So far we have discussed types of spatial data, some basic modeling frameworks and exploratory techniques. We have not discussed
More informationAppendix A Probability Distributions
Appendix A Probability Distributions A.1 Discrete Distributions A.1.1 Binomial Distribution The binomial distribution (Bernoulli model) describes the number of failures, X, in n independent Bernoulli trials.
More informationForward Problems and their Inverse Solutions
Forward Problems and their Inverse Solutions Sarah Zedler 1,2 1 King Abdullah University of Science and Technology 2 University of Texas at Austin February, 2013 Outline 1 Forward Problem Example Weather
More informationReport and Opinion 2016;8(6) Analysis of bivariate correlated data under the Poisson-gamma model
Analysis of bivariate correlated data under the Poisson-gamma model Narges Ramooz, Farzad Eskandari 2. MSc of Statistics, Allameh Tabatabai University, Tehran, Iran 2. Associate professor of Statistics,
More information