Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
|
|
- Eugenia Francis
- 5 years ago
- Views:
Transcription
1 Hierarchical models Dr. Jarad Niemi Iowa State University August 31, 2017 Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
2 Normal hierarchical model Let Y ig N(θ g, σ 2 ) for i = 1,..., n g, g = 1,..., G, and G g=1 n g = n. Now consider the following model assumptions: θ g N(µ, τ 2 ) θ g La(µ, τ) θ g t v (µ, τ 2 ) θ g πδ 0 + (1 π)n(µ, τ 2 ) θ g πδ 0 + (1 π)t v (µ, τ 2 ) To perform a Bayesian analysis, we need a prior on µ, τ 2, and (in the case of the discrete mixture) π. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
3 Gibbs sampling Normal hierarchical model Consider the model Y ig θ g N(θ g, σ 2 ) N(µ, τ 2 ) where i = 1,..., n g, g = 1,..., G, and n = G g=1 n g with prior distribution p(µ, σ 2, τ 2 ) = p(µ)p(σ 2 )p(τ 2 ) 1 σ 2 Ca + (τ; 0, C). For background on why we are using these priors for the variances, see Gelman (2006) Prior distributions for variance parameters in hierarchical models (comment on article by Browne and Draper). Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
4 Gibbs sampling Multi-step Gibbs sampler for normal hierarchical model Here is a possible Gibbs sampler for this model: For g = 1,..., G, sample θ g p(θ g ). Sample σ 2 p(σ 2 ). Sample µ p(µ ). Sample τ 2 p(τ 2 ). How many steps exist in this Gibbs sampler? G+3? 4? Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
5 Gibbs sampling 2-Step 2-Step Gibbs sampler for normal hierarchical model Here is a 2-step Gibbs sampler: 1. Sample θ = (θ 1,..., G) p(θ ). 2. Sample µ, σ 2, τ 2 p(µ, σ 2, τ 2 ). There is stronger theoretical support for 2-step Gibbs sampler, thus, if we can, it is prudent to construct a 2-step Gibbs sampler. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
6 Gibbs sampling Sampling θ Sampling θ The full conditional for θ is p(θ ) p(θ, µ, σ 2, τ 2 y) p(y θ, σ 2 )p(θ µ, τ 2 )p(µ, σ 2, τ 2 ) p(y θ, σ 2 )p(θ µ, τ 2 ) = G g=1 p(y g θ g, σ 2 )p(θ g µ, τ 2 ) where y g = (y 1g,..., y ng g ). We now know that the θ g are conditionally ependent of each other. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
7 Gibbs sampling Sampling θ g Sampling θ g The full conditional for θ g is p(θ g ) p(y g θ g, σ 2 )p(θ g µ, τ 2 ) = n g i=1 N(y ig ; θ g, σ 2 )N(θ g ; µ, τ 2 ) Notice that this does not include θ g for any g g. This is an alternative way to conclude that the θ g are conditionally ependent of each other. Thus where θ g N(µ g, τ 2 g ) τg 2 = [τ 2 + n g σ 2 ] 1 µ g y g = τg 2 [µτ 2 + y g n g σ 2 ] = 1 ng n g i=1 y ig. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
8 Gibbs sampling Sampling µ, σ 2, τ 2 Sampling µ, σ 2, τ 2 The full conditional for µ, σ 2, τ 2 is p(µ, σ 2, τ 2 ) p(y θ, σ 2 )p(θ µ, τ 2 )p(µ)p(σ 2 )p(τ 2 ) = p(y θ, σ 2 )p(σ 2 )p(θ µ, τ 2 )p(µ)p(τ 2 ) So we know that σ 2 is ependent of µ and τ 2. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
9 Gibbs sampling Sampling σ 2 Sampling σ 2 Recall that y ig N(θ g, σ 2 ) and p(σ 2 ) 1/σ 2. Thus, we are in the scenario of normal data with a known mean and unknown variance and the unknown variance has our default prior. Thus, we should know ( the full conditional is σ 2 IG n 2, 1 G ng 2 g=1 i=1 (y ig θ g ) ). 2 To derive the full conditional, use p(σ 2 ) G ng g=1 i=1 (σ2 ) 1/2 exp ( 1 (y 2σ 2 ig θ g ) 2) 1 ( σ 2 = (σ 2 ) n/2 1 exp 1 G / ) ng 2 g=1 i=1 (y ig θ g ) 2 σ 2 ( which is the kernel of a IG n 2, 1 G ng 2 g=1 i=1 (y ig θ g ) ). 2 Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
10 Sampling µ, τ 2 Sampling µ, τ 2 Recall that θ g N(µ, τ 2 ) and p(µ, τ 2 ) Ca + (τ; 0, C). This is a non-standard distribution, but is extremely close a normal model with unknown mean and variance with the standard non-informative prior p(µ, τ 2 ) 1/τ 2 or the conjugate normal-inverse-gamma prior. Here are some options for sampling from this distribution: random-walk Metropolis (in 2 dimensions), ependent Metropolis-Hastings using posterior from standard non-informative prior as the proposal, or rejection sampling using posterior from standard non-informative prior as the proposal The posterior under the standard non-informative prior is τ 2 Inv-χ 2 (G 1, s 2 θ) and µ τ 2,... N(θ, τ 2 /G) where θ = 1 G G g=1 θ g and s 2 θ = 1 G 1 (θ g θ) 2. What is the MH ratio? Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
11 Sampling µ, τ 2 Summary Markov chain Monte Carlo for normal hierarchical model 1. Sample θ p(θ ): a. For g = 1,..., G, sample θ g N(µ g, τ 2 g ). 2. Sample µ, σ 2, τ 2 : a. Sample σ 2 IG(n/2, SSE/2). b. Sample µ, τ 2 using ependent Metropolis-Hastings using posterior from standard non-informative prior as the proposal. What happens if θ g La(µ, τ) or θ g t v (µ, τ 2 )? Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
12 Scale mixtures of normals Scale mixtures of normals Recall that if then This is called a location mixture. Now, if θ φ N(φ, V ) and φ N(m, C) θ N(m, V + C). θ φ N(m, Cφ) and we assume a mixing distribution for φ, we have a scale mixture. Since the top level distributional assumption is normal, we refer to this as a scale mixture of normals. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
13 Scale mixtures of normals t distribution t distribution Let then p(θ) θ φ N(m, φc) and φ IG(a, b) = p(θ φ)p(φ)dφ = (2π C) 1/2 ba Γ(a) φ 1/2 e (θ m)2 /2φC φ (a+1) e b/φ dφ = (2πC) 1/2 ba φ (a+1/2+1) e [b+(θ m)2 /2C]/φ dφ = (2πC) = Γ(a) 1/2 ba Γ(a+1/2) Γ(a) [b+(θ m) 2 /2C] [ a+1/2 ] (θ m) 2 [2a+1]/2 2a bc/a Γ([2a+1]/2) Γ(2a/2) 2aπbC/a Thus θ t 2a (m, bc/a) i.e. θ has a t distribution with 2a degrees of freedom, location m, scale bc/a, and variance bc a 1. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
14 Scale mixtures of normals t distribution Hierarchical t distribution Let m = µ, C = 1, a = ν/2, and b = ντ 2 /2, i.e. θ φ N(µ, φ) and φ IG(ν/2, ντ 2 /2). Then, we have θ t ν (µ, τ 2 ), i.e. a t distribution with ν degrees of freedom, location µ, and scale τ 2. Notice that the parameterization has a redundancy between C and a/b, i.e. we could have chosen C = τ 2, a = ν/2, and b = ν/2 and we would have obtained the same marginal distribution for θ. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
15 Scale mixtures of normals t distribution Laplace distribution Let θ φ N(m, φc 2 ) and φ Exp(1/2b 2 ) where E[φ] = 2b 2 and Var[φ] = 4b 4. Then, by an extension of equation (4) in Park and Casella (2008), we have p(θ) = 1 θ m 2Cb e Cb. This is the pdf for a Laplace (double exponential) distribution with location m and scale Cb which we write θ La(m, Cb). and say θ has a Laplace distribution with location m and scale Cb and E[θ] = m and Var[θ] = 2[Cb] 2 = 2C 2 b 2. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
16 Scale mixtures of normals t distribution Hierarchical Laplace distribution Let m = µ, C = 1, and b = τ i.e. θ φ N(µ, φ) and φ Exp(1/2τ 2 ). Then, we have θ La(µ, τ), i.e. a Laplace distribution with location µ and scale τ. Notice that the parameterization has a redundancy between C and b, i.e. we could have chosen C = τ and b = 1 and we would have obtained the same marginal distribution for θ. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
17 Normal hierarchical model Normal hierarchical model Recall our hierarchical model Y ig N(θ g, σ 2 ) for g = 1,..., G and i = 1,..., n g. Now consider the following model assumptions: θ g φ g N(µ, φ g ), φ g = τ 2 = θ g N(µ, τ 2 ) θ g φ g N(µ, φ g ), θ g φ g N(µ, φ g ), φ g Exp(1/2τ 2 ) = θ g La(µ, τ) φ g IG(v/2, vτ 2 /2) = θ g t v (µ, τ 2 ) For simplicity, let s assume σ 2 IG(a, b), µ N(m, C), and τ Ca + (0, c) and that σ 2, µ, and τ are a priori ependent. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
18 MCMC Gibbs sampling The following Gibbs sampler will converge to the posterior p(θ, σ, µ, φ, τ y): 1. Independently, sample θ g p(θ g ). 2. Sample µ, σ, τ p(µ, σ, τ ): a. Sample µ p(µ ). b. Sample σ p(σ ). c. Sample τ p(τ ). 3. Independently, sample φ g p(φ g ). Steps 1, 2a, and 2b will be the same for all models, but steps 2c and 3 will be different. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
19 MCMC θ Sample θ Y ig N(θ g, σ 2 ) and θ g N(µ, φ g ) p(θ ) [ G ng g=1 G g=1 i=1 e (y ig θ g ) 2 /2σ 2] [ ] G g=1 e (θg µ)2 /2φ g [ ng ] i=1 e (y ig θ g ) 2 /2σ 2 e (θg µ)2 /2φ g Thus θ g are conditionally ependent given everything else. It should be obvious that ( [ 1 θ g N µ + n ] [ g 1 φ g σ 2 y g, + n ] ) 1 g φ g σ 2 where y g = n g i=1 y ig /n g. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
20 MCMC µ Sample µ θ g N(µ, φ g ) and µ N(m, C) Immediately, we should know that µ N(m, C ) with ( C = 1 C + ) 1 G g=1 1 ( φ g m = C 1 C m + ) G g=1 1 φ g θ g Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
21 MCMC σ Sample σ 2 Y ig N(θ g, σ 2 ) and σ 2 IG(a, b) This is just a normal data model with an unknown variance that has the conjugate prior. The only difficulty is that we have several groups here. But very quickly you should be able to determine that σ 2 IG(a, b ) where a = a + 1 G 2 g=1 n g = a + n 2 b = b + 1 G ng 2 g=1 i=1 (y ig θ g ) 2. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
22 MCMC Distributional assumption for θ g Distributional assumption for θ g Y ig N(θ g, σ 2 ) and θ g N(µ, φ g ) φ g φ g φ g = τ Exp(1/2τ 2 ) IG(v/2, vτ 2 /2) The steps that are left are 1) sample φ and 2) sample τ 2, Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
23 MCMC φ Sample φ for normal model For normal model, φ g = τ, so we will address this when we sample τ. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
24 MCMC φ Sample φ for Laplace model For Laplace model, θ g N(µ, φ g ) and φ g Exp(1/2τ 2 ), so the full conditional is [ G ] p(φ ) N(θ g ; µ, φ g )Exp(φ g ; 1/2τ 2 ). g=1 So the ividual φ g are conditionally ependent with p(φ g ) N(θ g ; µ, φ g )Exp(φ g ; 1/2τ 2 ) φ 1/2 g e (θg µ)2 2 /2φ g φg /2τ e If we perform the transformation η g = 1/φ g, we have p(η g ) η 3/2 g e (θg µ) 2 2 η g 1 2τ 2 ηg which is the kernel of an inverse Gaussian distribution with mean 1/τ 2 (θ g µ) 2 and scale 1/τ 2 where the parameterization is such that the variance is µ 3 /λ (different from the mgcv::rig parameterization). Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
25 MCMC φ Sample φ for t model For the t model, so we have θ g N(µ, φ g ) and φ g IG(v/2, vτ 2 /2), ( φ g v + 1 IG, vτ2 + (θ g µ) 2 ). 2 2 Since this is just G ependent normal data models with a known mean and ependent conjugate inverse gamma priors on the variance. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
26 MCMC Then we calculate τ = η. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31 τ Sample τ for normal model Let so the full conditional is θ g N(µ, τ 2 ) and τ Ca + (0, c). p(η ) η G/2 e G g=1 (θg µ)2 /2η ( 1 + η/c 2) 1 η 1/2 where we performed the transformation η = τ 2 on the prior. Let s use Metropolis-Hastings with proposal distribution ( ) η G 1 G (θ g µ) 2 IG, 2 2 g=1 and acceptance probability min{1, ρ} where ( 1 + η /c 2) 1 ρ = ( 1 + η (i) /c 2) 1 = 1 + η(i) /c η /c 2 where η (i) and η are the current and proposed value respective.
27 MCMC τ Sample τ for Laplace model Let so the full conditional is φ g Exp(1/2τ 2 ) and τ Ca + (0, c) p(η ) η G e G g=1 φg /2η ( 1 + η/c 2) 1 η 1/2. Let s use Metropolis-Hastings with proposal distribution η IG G 1 G 2, φ g 2 g=1 and acceptance probability min{1, ρ} where again Then we calculate τ = η. ρ = 1 + η(i) /c η /c 2. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
28 MCMC τ Sample τ for t model Let so the full conditional is φ g IG(v/2, vτ 2 /2) and τ Ca + (0, c) p(η ) η Gv/2 e η 2 G g=1 1 φg ( 1 + η/c 2 ) 1 η 1/2. Let s use Metropolis-Hastings with proposal distribution η Ga Gv + 1, 1 G φ g=1 g and acceptance probability min{1, ρ} where again Then we calculate τ = η. ρ = 1 + η(i) /c η /c 2. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
29 MCMC Point-mass distributions Dealing with point-mass distributions We would also like to consider models with θ g πδ 0 + (1 π)n(µ, φ g ) where φ g = τ 2 corresponds to a normal and φ g IG(v/2, vτ 2 /2) corresponds to a t distribution for the non-zero θ g. Similar to the previous, the θ g are conditionally ependent. To sample θ g, we calculate π = π ng i=1 N(y ig ;0,σ 2 ) π ng i=1 N(y ig ;0,σ 2 )+(1 π) ng. i=1 N(y ig ;µ,φ g +σ 2 ) ( φ g = 1 φ g + ng ( ) µ g = φ µ g + ng y φg σ 2 g σ 2 ) 1 Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
30 MCMC Point-mass distributions Dealing with point-mass distributions (cont.) Let and π Beta(s, f ). θ g πδ 0 + (1 π)f (θ g ) The full conditional for π is G G π Beta s + I(θ g = 0), f + I(θ g 0) g=1 g=1 and µ and φ g get updated using only those θ g that are non-zero. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
31 MCMC Point-mass distributions Dealing with point-mass distributions (cont.) Updating µ, φ g, and τ, will be exactly the same as it was before EXCEPT, you will only use g such that θ g 0. So µ N(m, C ) with ( C = 1 C + ) 1 g:θ g 0 1 ( φ g m = C 1 C m + ) g:θ g 0 1 φ g θ g, τ will still require a MH step, but the proposal should only depend on g : θ g 0, and φ g will be exactly as before when θ g 0, but when θ g = 0, then it actually doesn t matter what distribution you are sampling from (Carlin and Chib 1995). Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
CS281A/Stat241A Lecture 22
CS281A/Stat241A Lecture 22 p. 1/4 CS281A/Stat241A Lecture 22 Monte Carlo Methods Peter Bartlett CS281A/Stat241A Lecture 22 p. 2/4 Key ideas of this lecture Sampling in Bayesian methods: Predictive distribution
More informationMultiparameter models (cont.)
Multiparameter models (cont.) Dr. Jarad Niemi STAT 544 - Iowa State University February 1, 2018 Jarad Niemi (STAT544@ISU) Multiparameter models (cont.) February 1, 2018 1 / 20 Outline Multinomial Multivariate
More informationThe binomial model. Assume a uniform prior distribution on p(θ). Write the pdf for this distribution.
The binomial model Example. After suspicious performance in the weekly soccer match, 37 mathematical sciences students, staff, and faculty were tested for the use of performance enhancing analytics. Let
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationBayesian Linear Regression
Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective
More informationLecture 4: Dynamic models
linear s Lecture 4: s Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu
More informationSpring 2006: Introduction to Markov Chain Monte Carlo (MCMC)
36-724 Spring 2006: Introduction to Marov Chain Monte Carlo (MCMC) Brian Juner February 16, 2006 Hierarchical Normal Model Direct Simulation An Alternative Approach: MCMC Complete Conditionals for Hierarchical
More informationBayesian data analysis in practice: Three simple examples
Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to
More informationDavid Giles Bayesian Econometrics
David Giles Bayesian Econometrics 5. Bayesian Computation Historically, the computational "cost" of Bayesian methods greatly limited their application. For instance, by Bayes' Theorem: p(θ y) = p(θ)p(y
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationMetropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9
Metropolis Hastings Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 9 1 The Metropolis-Hastings algorithm is a general term for a family of Markov chain simulation methods
More informationThe Metropolis-Hastings Algorithm. June 8, 2012
The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings
More informationMultivariate Normal & Wishart
Multivariate Normal & Wishart Hoff Chapter 7 October 21, 2010 Reading Comprehesion Example Twenty-two children are given a reading comprehsion test before and after receiving a particular instruction method.
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationBayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference
1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE
More informationMarkov chain Monte Carlo
1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop
More informationMarkov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation. Luke Tierney Department of Statistics & Actuarial Science University of Iowa
Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation Luke Tierney Department of Statistics & Actuarial Science University of Iowa Basic Ratio of Uniforms Method Introduced by Kinderman and
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More informationMetropolis-Hastings sampling
Metropolis-Hastings sampling Gibbs sampling requires that a sample from each full conditional distribution. In all the cases we have looked at so far the conditional distributions were conjugate so sampling
More informationF denotes cumulative density. denotes probability density function; (.)
BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee September 03 05, 2017 Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles Linear Regression Linear regression is,
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public
More informationHastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model
UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced
More informationINTRODUCTION TO BAYESIAN STATISTICS
INTRODUCTION TO BAYESIAN STATISTICS Sarat C. Dass Department of Statistics & Probability Department of Computer Science & Engineering Michigan State University TOPICS The Bayesian Framework Different Types
More informationComputer Practical: Metropolis-Hastings-based MCMC
Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov
More informationComputationalToolsforComparing AsymmetricGARCHModelsviaBayes Factors. RicardoS.Ehlers
ComputationalToolsforComparing AsymmetricGARCHModelsviaBayes Factors RicardoS.Ehlers Laboratório de Estatística e Geoinformação- UFPR http://leg.ufpr.br/ ehlers ehlers@leg.ufpr.br II Workshop on Statistical
More informationA Review of Pseudo-Marginal Markov Chain Monte Carlo
A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the
More informationBayesian Defect Signal Analysis
Electrical and Computer Engineering Publications Electrical and Computer Engineering 26 Bayesian Defect Signal Analysis Aleksandar Dogandžić Iowa State University, ald@iastate.edu Benhong Zhang Iowa State
More informationMarkov Chain Monte Carlo Methods
Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)
More informationGeneral Bayesian Inference I
General Bayesian Inference I Outline: Basic concepts, One-parameter models, Noninformative priors. Reading: Chapters 10 and 11 in Kay-I. (Occasional) Simplified Notation. When there is no potential for
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationIntroduction to Machine Learning
Introduction to Machine Learning Generative Models Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1
More informationMarkov Chain Monte Carlo, Numerical Integration
Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical
More informationControl Variates for Markov Chain Monte Carlo
Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability
More informationLecture 3. Univariate Bayesian inference: conjugate analysis
Summary Lecture 3. Univariate Bayesian inference: conjugate analysis 1. Posterior predictive distributions 2. Conjugate analysis for proportions 3. Posterior predictions for proportions 4. Conjugate analysis
More informationStat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC
Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationLecture 8: Bayesian Estimation of Parameters in State Space Models
in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space
More informationeqr094: Hierarchical MCMC for Bayesian System Reliability
eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167
More informationBayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014
Bayesian Prediction of Code Output ASA Albuquerque Chapter Short Course October 2014 Abstract This presentation summarizes Bayesian prediction methodology for the Gaussian process (GP) surrogate representation
More informationLikelihood-free MCMC
Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll
More informationLinear Models A linear model is defined by the expression
Linear Models A linear model is defined by the expression x = F β + ɛ. where x = (x 1, x 2,..., x n ) is vector of size n usually known as the response vector. β = (β 1, β 2,..., β p ) is the transpose
More informationSupplementary Notes: Segment Parameter Labelling in MCMC Change Detection
Supplementary Notes: Segment Parameter Labelling in MCMC Change Detection Alireza Ahrabian 1 arxiv:1901.0545v1 [eess.sp] 16 Jan 019 Abstract This work addresses the problem of segmentation in time series
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationMotivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University
Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined
More informationIntroduction to Bayesian Computation
Introduction to Bayesian Computation Dr. Jarad Niemi STAT 544 - Iowa State University March 20, 2018 Jarad Niemi (STAT544@ISU) Introduction to Bayesian Computation March 20, 2018 1 / 30 Bayesian computation
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationBayesian Gaussian Process Regression
Bayesian Gaussian Process Regression STAT8810, Fall 2017 M.T. Pratola October 7, 2017 Today Bayesian Gaussian Process Regression Bayesian GP Regression Recall we had observations from our expensive simulator,
More informationBAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA
BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci
More informationθ 1 θ 2 θ n y i1 y i2 y in Hierarchical models (chapter 5) Hierarchical model Introduction to hierarchical models - sometimes called multilevel model
Hierarchical models (chapter 5) Introduction to hierarchical models - sometimes called multilevel model Exchangeability Slide 1 Hierarchical model Example: heart surgery in hospitals - in hospital j survival
More informationDavid Giles Bayesian Econometrics
David Giles Bayesian Econometrics 1. General Background 2. Constructing Prior Distributions 3. Properties of Bayes Estimators and Tests 4. Bayesian Analysis of the Multiple Regression Model 5. Bayesian
More informationMonte Carlo Methods in Bayesian Inference: Theory, Methods and Applications
University of Arkansas, Fayetteville ScholarWorks@UARK Theses and Dissertations 1-016 Monte Carlo Methods in Bayesian Inference: Theory, Methods and Applications Huarui Zhang University of Arkansas, Fayetteville
More informationClassical and Bayesian inference
Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9 Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli
More informationLearning the hyper-parameters. Luca Martino
Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth
More informationHierarchical Modeling for Spatial Data
Bayesian Spatial Modelling Spatial model specifications: P(y X, θ). Prior specifications: P(θ). Posterior inference of model parameters: P(θ y). Predictions at new locations: P(y 0 y). Model comparisons.
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationBayesian Estimation with Sparse Grids
Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse
More informationMetropolis-Hastings Algorithm
Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationThe exponential family: Conjugate priors
Chapter 9 The exponential family: Conjugate priors Within the Bayesian framework the parameter θ is treated as a random quantity. This requires us to specify a prior distribution p(θ), from which we can
More information6.1 Approximating the posterior by calculating on a grid
Chapter 6 Approximations 6.1 Approximating the posterior by calculating on a grid When one is confronted with a low-dimensional problem with a continuous parameter, then it is usually feasible to approximate
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.
Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationBayesian Inference. Chapter 9. Linear models and regression
Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering
More informationAdvanced Statistical Modelling
Markov chain Monte Carlo (MCMC) Methods and Their Applications in Bayesian Statistics School of Technology and Business Studies/Statistics Dalarna University Borlänge, Sweden. Feb. 05, 2014. Outlines 1
More informationBayesian Inference. Chapter 2: Conjugate models
Bayesian Inference Chapter 2: Conjugate models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in
More informationBayesian GLMs and Metropolis-Hastings Algorithm
Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,
More informationIntroduction to Bayesian Inference
University of Pennsylvania EABCN Training School May 10, 2016 Bayesian Inference Ingredients of Bayesian Analysis: Likelihood function p(y φ) Prior density p(φ) Marginal data density p(y ) = p(y φ)p(φ)dφ
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More informationMCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham
MCMC 2: Lecture 2 Coding and output Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. General (Markov) epidemic model 2. Non-Markov epidemic model 3. Debugging
More informationDeblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.
Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical
More information10. Exchangeability and hierarchical models Objective. Recommended reading
10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.
More informationBayesian model selection: methodology, computation and applications
Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program
More informationMCMC for Cut Models or Chasing a Moving Target with MCMC
MCMC for Cut Models or Chasing a Moving Target with MCMC Martyn Plummer International Agency for Research on Cancer MCMSki Chamonix, 6 Jan 2014 Cut models What do we want to do? 1. Generate some random
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 4 Problem: Density Estimation We have observed data, y 1,..., y n, drawn independently from some unknown
More informationCross-sectional space-time modeling using ARNN(p, n) processes
Cross-sectional space-time modeling using ARNN(p, n) processes W. Polasek K. Kakamu September, 006 Abstract We suggest a new class of cross-sectional space-time models based on local AR models and nearest
More informationMarkov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017
Markov Chain Monte Carlo (MCMC) and Model Evaluation August 15, 2017 Frequentist Linking Frequentist and Bayesian Statistics How can we estimate model parameters and what does it imply? Want to find the
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationMCMC 2: Lecture 3 SIR models - more topics. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham
MCMC 2: Lecture 3 SIR models - more topics Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. What can be estimated? 2. Reparameterisation 3. Marginalisation
More informationMH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution
MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous
More informationSpring 2006: Examples: Laplace s Method; Hierarchical Models
36-724 Spring 2006: Examples: Laplace s Method; Hierarchical Models Brian Junker February 13, 2006 Second-order Laplace Approximation for E[g(θ) y] An Analytic Example Hierarchical Models Example of Hierarchical
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationTheory of Stochastic Processes 8. Markov chain Monte Carlo
Theory of Stochastic Processes 8. Markov chain Monte Carlo Tomonari Sei sei@mist.i.u-tokyo.ac.jp Department of Mathematical Informatics, University of Tokyo June 8, 2017 http://www.stat.t.u-tokyo.ac.jp/~sei/lec.html
More informationIntroduction to Bayesian Methods. Introduction to Bayesian Methods p.1/??
to Bayesian Methods Introduction to Bayesian Methods p.1/?? We develop the Bayesian paradigm for parametric inference. To this end, suppose we conduct (or wish to design) a study, in which the parameter
More information7. Estimation and hypothesis testing. Objective. Recommended reading
7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing
More informationLecture 7 and 8: Markov Chain Monte Carlo
Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani
More informationBAYESIAN MODEL CRITICISM
Monte via Chib s BAYESIAN MODEL CRITICM Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationHierarchical Bayesian approaches for robust inference in ARX models
Hierarchical Bayesian approaches for robust inference in ARX models Johan Dahlin, Fredrik Lindsten, Thomas Bo Schön and Adrian George Wills Linköping University Post Print N.B.: When citing this work,
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationMCMC and Gibbs Sampling. Sargur Srihari
MCMC and Gibbs Sampling Sargur srihari@cedar.buffalo.edu 1 Topics 1. Markov Chain Monte Carlo 2. Markov Chains 3. Gibbs Sampling 4. Basic Metropolis Algorithm 5. Metropolis-Hastings Algorithm 6. Slice
More informationIntroduction to Markov Chain Monte Carlo & Gibbs Sampling
Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu
More information